Paridhim commited on
Commit
36b15e0
·
1 Parent(s): 42eb242

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +14 -26
main.py CHANGED
@@ -1,38 +1,26 @@
 
 
 
1
  import uvicorn
2
  from fastapi import FastAPI, Request
3
- from langchain_community.llms.ctransformers import CTransformers
4
-
5
-
6
-
7
- MODEL_TYPE = 'mistral'
8
- MODEL_BIN_PATH = "mistral-7b-instruct-v0.1.Q3_K_S.gguf"
9
- MAX_NEW_TOKEN = 600
10
- TEMPRATURE = 0.01
11
- CONTEXT_LENGTH = 6000
12
-
13
 
14
- llm = CTransformers(
15
- model=MODEL_BIN_PATH,
16
- config={
17
- 'max_new_tokens': MAX_NEW_TOKEN,
18
- 'temperature': TEMPRATURE,
19
- 'context_length': CONTEXT_LENGTH
20
- },
21
- model_type=MODEL_TYPE
22
- )
23
 
24
  app = FastAPI()
 
25
 
26
 
27
- @app.post("//llm_on_cpu")
 
28
  async def generate_text(request: Request):
29
  raw_data = await request.body() # Get the raw body data from the request
30
  prompt = raw_data.decode('utf-8')
31
  if not prompt:
32
  return {'error': 'Prompt is required'}, 400
33
- try:
34
- generated_text = llm(prompt)
35
- print(generated_text)
36
- return {'generated_text': generated_text}
37
- except:
38
- return {"generated_text": "Error!"}
 
 
 
1
+
2
+ from typing import Union
3
+ from wrapper import LLMWrapper
4
  import uvicorn
5
  from fastapi import FastAPI, Request
 
 
 
 
 
 
 
 
 
 
6
 
 
 
 
 
 
 
 
 
 
7
 
8
  app = FastAPI()
9
+ llm_wrapper = LLMWrapper()
10
 
11
 
12
+
13
+ @app.post("/")
14
  async def generate_text(request: Request):
15
  raw_data = await request.body() # Get the raw body data from the request
16
  prompt = raw_data.decode('utf-8')
17
  if not prompt:
18
  return {'error': 'Prompt is required'}, 400
19
+
20
+ generated_text = llm_wrapper.generate_text(prompt)
21
+ print(generated_text)
22
+ return {'generated_text': generated_text}
23
+
24
+
25
+ if __name__ == '__main__':
26
+ uvicorn.run(app, host='127.0.0.1', port=8001)