Skip to content
Related Articles

Related Articles

Corona HelpBot

View Discussion
Improve Article
Save Article
  • Difficulty Level : Hard
  • Last Updated : 14 Jul, 2022
View Discussion
Improve Article
Save Article

This is a chatbot that will give answers to most of your corona-related questions/FAQ. The chatbot will give you answers from the data given by WHO( This will help those who need information or help to know more about this virus.

It uses a neural network with two hidden layers(enough for these QnA) that predicts which pattern matches the user’s question and sends it towards that node. More patterns can be added to the user’s questions to train it for more improved results and add more info about coronavirus in the JSON file. The more you train this chatbot the more it gets precise. The advantage of using deep learning is that you don’t have to ask the same question as written in the JSON file cause stemmed words from the pattern are matched with the user question


Python 3
TensorFlow v.1.15 (no GPU required)

Training Data: 
To feed the data to the chatbot I have used JSON with possible question patterns and our desired answers. 
The JSON file used for the project is WHO 
For this project, I have named my JSON file WHO.json 
The JSON file tag is the category in which all those responses fall. 
patterns are used for listing all possible question patterns. 
responses contain all the responses with respect to the patterned questions 


import nltk
import numpy
import tflearn
import tensorflow
import pickle
import random
import json'punkt')
from import LancasterStemmer
stemmer = LancasterStemmer()
 #loading the json data
with open("WHO.json") as file:                 
    data = json.load(file)
    with open("data.pickle", "rb") as f:
        words, l, training, output = pickle.load(f)
    #  Extracting Data
    words = []
    l = []
    docs_x = []
    docs_y = []
   # converting each pattern into list of words using nltk.word_tokenizer
    for i in data["intents"]:  
        for p in i["patterns"]:
            wrds = nltk.word_tokenize(p)
            if i["tag"] not in l:
    # Word Stemming           
    words = [stemmer.stem(w.lower()) for w in words if w != "?"]        
    words = sorted(list(set(words)))
    l = sorted(l)                                     
    # This code will simply create a unique list of stemmed
    # words to use in the next step of our data preprocessing
    training = []
    output = []
    out_empty = [0 for _ in range(len(l))]
    for x, doc in enumerate(docs_x):
        bag = []
        wrds = [stemmer.stem(w) for w in doc]
        for w in words:
            if w in wrds:
        output_row = out_empty[:]
        output_row[l.index(docs_y[x])] = 1
    # Finally we will convert our training data and output to numpy arrays   
    training = numpy.array(training)       
    output = numpy.array(output)
    with open("data.pickle", "wb") as f:
        pickle.dump((words, l, training, output), f)
# Developing a Model       
net = tflearn.input_data(shape=[None, len(training[0])])
net = tflearn.fully_connected(net, 8)
net = tflearn.fully_connected(net, 8)
net = tflearn.fully_connected(net, len(output[0]), activation="softmax")
net = tflearn.regression(net)
# remove comment to not train model after you satisfied with the accuracy
model = tflearn.DNN(net)
# Training & Saving the Model, output, n_epoch=1000, batch_size=8, show_metric=True)"model.tflearn")
# making predictions
def bag_of_words(s, words):                               
    bag = [0 for _ in range(len(words))]
    s_words = nltk.word_tokenize(s)
    s_words = [stemmer.stem(word.lower()) for word in s_words]
    for se in s_words:
        for i, w in enumerate(words):
            if w == se:
                bag[i] = 1
    return numpy.array(bag)
def chat():
    print("""Start talking with the bot and ask your
    queries about Corona-virus(type quit to stop)!""")
    while True:
        inp = input("You: ")
        if inp.lower() == "quit":
        results = model.predict([bag_of_words(inp, words)])[0]
        results_index = numpy.argmax(results)
        tag = l[results_index]
        if results[results_index] > 0.7:
            for tg in data["intents"]:
                if tg['tag'] == tag:
                    responses = tg['responses']
            print("I am sorry but I can't understand")

Bag of Words: 
As we know neural networks and machine learning algorithms require numerical input. So our list of strings won’t cut it. We need some way to represent our sentences with numbers and this is where a bag of words comes in. What we are going to do is represent each sentence with a list of the length of the number of words in our model vocabulary. Each position in the list will represent a word from our vocabulary the position in the list is a 1 then that will mean that the word exists in our sentence, if it is a 0 then the word is not present.

Developing a Model:

model = tflearn.DNN(net), output, n_epoch=1000, batch_size=8, show_metric=True)"model.tflearn")

Our chatbot will predict the answers based on the model we train. Here we will use the neural networks to build the model. The goal of our network will be to look at a bag of words and give a class that they belong too (one of our tags from the JSON file).

Input: What is coronavirus?
Output: COVID-19 is an infectious disease caused by the most recently discovered coronavirus. This new virus and disease were unknown before the outbreak began in Wuhan, China, in December 2019.

Input: What are the symptoms of COVID 19?
Output: The most common symptoms of COVID-19 are fever, tiredness, and dry cough. Some patients may have aches and pains, nasal congestion, runny nose, sore throat, or diarrhoea. These symptoms are usually mild and begin gradually. Some people become infected but don’t develop any symptoms and don't feel unwell. Most people (about 80%) recover from the disease without needing special treatment.


My Personal Notes arrow_drop_up
Recommended Articles
Page :

Start Your Coding Journey Now!