EbubeJohnEnyi commited on
Commit
5524045
1 Parent(s): bfea49d

Upload ChatBot.py

Browse files
Files changed (1) hide show
  1. ChatBot.py +81 -0
ChatBot.py ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from flask import Flask, render_template, request
2
+ from transformers import GPT2LMHeadModel, GPT2Tokenizer
3
+ from sklearn.feature_extraction.text import CountVectorizer
4
+ from sklearn.metrics.pairwise import cosine_similarity
5
+ import json
6
+
7
+ tokenizer = GPT2Tokenizer.from_pretrained('gpt2')
8
+ model = GPT2LMHeadModel.from_pretrained('gpt2')
9
+
10
+ json_file_path = 'interview-workspace\src\data\short_let_room.json'
11
+
12
+ app = Flask(__name__)
13
+
14
+
15
+ def generate_gpt2_response(question):
16
+ input_ids = tokenizer.encode(question, return_tensors='pt')
17
+ generated_output = model.generate(input_ids, max_length=len(input_ids[0]) + 100,
18
+ num_beams=5,
19
+ no_repeat_ngram_size=2,
20
+ top_k=10,
21
+ top_p=1,
22
+ temperature=0.9,
23
+ pad_token_id=model.config.eos_token_id)
24
+ generated_response = tokenizer.decode(generated_output[0], skip_special_tokens=True)
25
+ return generated_response
26
+
27
+ def find_question_and_answer(json_file, question, input_ids):
28
+ with open(json_file, "r") as json_file:
29
+ data = json.load(json_file)
30
+
31
+ quest = question.lower()
32
+
33
+ all_questions = data.get("questions", [])
34
+ all_response = data.get("responses", [])
35
+ for dataset_question in all_questions:
36
+ if 'question' in dataset_question and dataset_question['question'].lower() == quest:
37
+ for dataset_response in all_response:
38
+ if dataset_question['response_id'] == dataset_response['id']:
39
+ response = {
40
+ "Chat Bot ": dataset_response["response_message"],
41
+ "Apartment ": dataset_response["response_message1"],
42
+ "Address ": dataset_response["shortlet_Address"],
43
+ "Price ": dataset_response["shortlet_Price"],
44
+ "URL ": dataset_response.get("shortlet_url")
45
+ }
46
+ return response
47
+
48
+ input_ids = tokenizer.encode(question, return_tensors='pt')
49
+ generated_output = model.generate(input_ids, max_length=len(input_ids[0]) + 100,
50
+ num_beams=5,
51
+ no_repeat_ngram_size=2,
52
+ top_k=10,
53
+ top_p=1,
54
+ temperature=0.9,
55
+ pad_token_id=model.config.eos_token_id)
56
+ generated_response = tokenizer.decode(generated_output[0], skip_special_tokens=True)
57
+ return generated_response
58
+
59
+ @app.route("/")
60
+ def index():
61
+ return render_template('index.html')
62
+
63
+ @app.route('/ask', methods=['POST'])
64
+ def ask():
65
+ user_input = request.form['user_input']
66
+ if not user_input:
67
+ response = ""
68
+ else:
69
+ input_ids = tokenizer.encode(user_input, return_tensors='pt')
70
+ response = find_question_and_answer(json_file_path, user_input, input_ids)
71
+
72
+ if isinstance(response, dict):
73
+ response_type = "mapping"
74
+ else:
75
+ response_type = "string"
76
+
77
+ return render_template('index.html', user_input=user_input, response=response, response_type=response_type)
78
+
79
+
80
+ if __name__ == '__main__':
81
+ app.run(debug=True)