jonathanjordan21 commited on
Commit
cdd670e
1 Parent(s): 1d7dab1

Create custom_llm.py

Browse files
Files changed (1) hide show
  1. custom_llm.py +128 -0
custom_llm.py ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Any, List, Mapping, Optional
2
+
3
+ from langchain_core.callbacks.manager import CallbackManagerForLLMRun
4
+ from langchain_core.language_models.llms import LLM
5
+ from typing import Literal
6
+ import requests
7
+ from langchain.prompts import PromptTemplate, ChatPromptTemplate
8
+ from operator import itemgetter
9
+
10
+ from langchain.memory import ChatMessageHistory, ConversationBufferMemory
11
+ from langchain.prompts import ChatPromptTemplate, MessagesPlaceholder
12
+ from langchain_community.chat_models import ChatOpenAI
13
+ from langchain_core.runnables import RunnableLambda, RunnablePassthrough
14
+ from langchain_core.messages import AIMessage, HumanMessage
15
+
16
+ from langchain_community.document_loaders import DirectoryLoader
17
+ from langchain.text_splitter import RecursiveCharacterTextSplitter
18
+ from langchain_community.document_loaders import PyMuPDFLoader
19
+ import os
20
+ from langchain.embeddings import HuggingFaceEmbeddings
21
+ from langchain.vectorstores import FAISS
22
+
23
+
24
+ def create_vectorstore():
25
+ loader = os.getenv('knowledge_base')
26
+
27
+ splitter = RecursiveCharacterTextSplitter(chunk_size=512, chunk_overlap=20)
28
+
29
+ docs = splitter.create_documents([loader])
30
+
31
+ emb_model = HuggingFaceEmbeddings(model_name='sentence-transformers/paraphrase-multilingual-mpnet-base-v2', encode_kwargs={'normalize_embeddings': True})
32
+ db = FAISS.from_documents(docs, emb_model)
33
+ return db
34
+
35
+ def custom_chain_with_history(llm, memory):
36
+
37
+ prompt = PromptTemplate.from_template("""<s><INST><|system|>
38
+ You are a college student at Gunadarma University, majoring in Informatics / Computer Science. Your name is Jonathan Jordan, but people call you Jojo. You have expertise in Artificial Intelligence and Data Science. You are also good at Data Engineering with AirFlow and Apache Spark. You currently live in Jakarta, Indonesia. You are 21 years old. Your birthday is Friday, 21 June 2002.
39
+
40
+ Below are the detailed personal information of you:
41
+ {context}
42
+
43
+ Respond to the user with the following chat history between you and the user:
44
+ {chat_history}
45
+ <|interviewer|>
46
+ {question}
47
+ <|you|>
48
+ """)
49
+
50
+ def prompt_memory(_):
51
+ t = ""
52
+ for x in memory.chat_memory.messages:
53
+ t += f"<|assistant|>\n<s>{x.content}</s>\n" if type(x) is AIMessage else f"<|user|>\n{x.content}\n"
54
+ return "" if len(t) == 0 else t
55
+
56
+ def format_docs(docs):
57
+ print(len(docs))
58
+ return "\n".join([f"{i+1}. {d.page_content}" for i,d in enumerate(docs)])
59
+
60
+ # prompt = ChatPromptTemplate.from_messages(
61
+ # [
62
+ # ("system", "You are a helpful chatbot"),
63
+ # MessagesPlaceholder(variable_name="history"),
64
+ # ("human", "{input}"),
65
+ # ]
66
+ # )
67
+
68
+ return {"chat_history":prompt_memory, "context":create_vectorstore().as_retriever(search_type="similarity", search_kwargs={"k": 8}) | format_docs, "question": RunnablePassthrough()} | prompt | llm
69
+
70
+ class CustomLLM(LLM):
71
+ repo_id : str
72
+ api_token : str
73
+ model_type: Literal["text2text-generation", "text-generation"]
74
+ max_new_tokens: int = None
75
+ temperature: float = 0.001
76
+ timeout: float = None
77
+ top_p: float = None
78
+ top_k : int = None
79
+ repetition_penalty : float = None
80
+ stop : List[str] = []
81
+
82
+
83
+ @property
84
+ def _llm_type(self) -> str:
85
+ return "custom"
86
+
87
+ def _call(
88
+ self,
89
+ prompt: str,
90
+ stop: Optional[List[str]] = None,
91
+ run_manager: Optional[CallbackManagerForLLMRun] = None,
92
+ **kwargs: Any,
93
+ ) -> str:
94
+
95
+ headers = {"Authorization": f"Bearer {self.api_token}"}
96
+ API_URL = f"https://api-inference.huggingface.co/models/{self.repo_id}"
97
+
98
+ parameters_dict = {
99
+ 'max_new_tokens': self.max_new_tokens,
100
+ 'temperature': self.temperature,
101
+ 'timeout': self.timeout,
102
+ 'top_p': self.top_p,
103
+ 'top_k': self.top_k,
104
+ 'repetition_penalty': self.repetition_penalty,
105
+ 'stop':self.stop
106
+ }
107
+
108
+ if self.model_type == 'text-generation':
109
+ parameters_dict["return_full_text"]=False
110
+
111
+ data = {"inputs": prompt, "parameters":parameters_dict, "options":{"wait_for_model":True}}
112
+ data = requests.post(API_URL, headers=headers, json=data).json()
113
+ return data[0]['generated_text']
114
+
115
+ @property
116
+ def _identifying_params(self) -> Mapping[str, Any]:
117
+ """Get the identifying parameters."""
118
+ return {
119
+ 'repo_id': self.repo_id,
120
+ 'model_type':self.model_type,
121
+ 'stop_sequences':self.stop,
122
+ 'max_new_tokens': self.max_new_tokens,
123
+ 'temperature': self.temperature,
124
+ 'timeout': self.timeout,
125
+ 'top_p': self.top_p,
126
+ 'top_k': self.top_k,
127
+ 'repetition_penalty': self.repetition_penalty
128
+ }