This is the implementation of multi head attention model on the cornell movie dataset. This is an implementation of Transformer models introduced in 2017 and is used in NLP in various task such as Translation (you can visit my other repository Neural Style Translation).
- first clone the repository
$ git clone https://github.com/sarthak7509/ConversationalAi.git
- Download the pretrained weights from my drive and place it in the model weights section
- Run conversationAi.py for demo or run app.py to start a local server that return the chatbot value for the POST methon
- The training files are included under training folder use it to train custom model
- As of 1st July 2021 I have used wkiiqna dataset provided by microsoft and used it to train on my existing model architecture with current accuracy of 25% which is not bad for 20 epochs
- Since the model used is same we just need to download latest weights and provide its path to the model my drive as on now the latest weight is model2.h5 and place it in folder model_weight_WkiiQna
- Training notebook is provided under training folder
- Just need to specify the model path in the code where ever it is mentioned
- Tokenizer is preincluded in the github file
Note wikki qna bot model weights are preincluded in the git directory name model_weight_WkiiQna. So no need to download current latest weight is model3.h5
$ python talkingdemo.py -b WEIGHT_PATH -t TOKENIZER_PATH -n BOT_NAME(optional)