KonradSzafer commited on
Commit
7f14c95
1 Parent(s): 9e86d4e

llama prompt and debug update

Browse files
config/prompt_templates/llama2.txt CHANGED
@@ -1,6 +1,7 @@
1
- <<SYS>>
2
- **CONTEXT FOR THE QUESTION**:
3
- {context}
 
4
  <</SYS>>
5
 
6
- [INST] Respond for the user question as factualy as possible, using given context. [/INST] User: {question}
 
1
+ <<SYS>>Using the information contained in the context,
2
+ give a comprehensive answer to the question.
3
+ Respond only to the question asked, response should be concise and relevant to the question.
4
+ If the answer cannot be deduced from the context, do not give an answer.
5
  <</SYS>>
6
 
7
+ [INST] Context: {context} [/INST] User: {question}
qa_engine/qa_engine.py CHANGED
@@ -227,7 +227,7 @@ class QAEngine():
227
  self.knowledge_index = FAISS.load_local('./indexes/run/', embedding_model)
228
  self.reranker = CrossEncoder('cross-encoder/ms-marco-MiniLM-L-12-v2')
229
 
230
-
231
  @staticmethod
232
  def _preprocess_question(question: str) -> str:
233
  if question[-1] != '?':
@@ -300,8 +300,8 @@ class QAEngine():
300
  logger.info('Running LLM chain')
301
  question_processed = QAEngine._preprocess_question(question)
302
  answer = self.llm_chain.run(question=question_processed, context=context)
303
- answer = QAEngine._postprocess_answer(answer)
304
- response.set_answer(answer)
305
  logger.info('Received answer')
306
 
307
  if self.debug:
@@ -310,7 +310,8 @@ class QAEngine():
310
  logger.info(f'question len: {len(question)} {sep}')
311
  logger.info(f'question: {question} {sep}')
312
  logger.info(f'answer len: {len(response.get_answer())} {sep}')
313
- logger.info(f'answer: {response.get_answer()} {sep}')
 
314
  logger.info(f'{response.get_sources_as_text()} {sep}')
315
  logger.info(f'messages_contex: {messages_context} {sep}')
316
  logger.info(f'relevant_docs: {relevant_docs} {sep}')
 
227
  self.knowledge_index = FAISS.load_local('./indexes/run/', embedding_model)
228
  self.reranker = CrossEncoder('cross-encoder/ms-marco-MiniLM-L-12-v2')
229
 
230
+
231
  @staticmethod
232
  def _preprocess_question(question: str) -> str:
233
  if question[-1] != '?':
 
300
  logger.info('Running LLM chain')
301
  question_processed = QAEngine._preprocess_question(question)
302
  answer = self.llm_chain.run(question=question_processed, context=context)
303
+ answer_postprocessed = QAEngine._postprocess_answer(answer)
304
+ response.set_answer(answer_postprocessed)
305
  logger.info('Received answer')
306
 
307
  if self.debug:
 
310
  logger.info(f'question len: {len(question)} {sep}')
311
  logger.info(f'question: {question} {sep}')
312
  logger.info(f'answer len: {len(response.get_answer())} {sep}')
313
+ logger.info(f'answer original: {answer} {sep}')
314
+ logger.info(f'answer postprocessed: {response.get_answer()} {sep}')
315
  logger.info(f'{response.get_sources_as_text()} {sep}')
316
  logger.info(f'messages_contex: {messages_context} {sep}')
317
  logger.info(f'relevant_docs: {relevant_docs} {sep}')