Spaces:
Sleeping
Sleeping
Update main.py
Browse files
main.py
CHANGED
@@ -1,38 +1,26 @@
|
|
|
|
|
|
|
|
1 |
import uvicorn
|
2 |
from fastapi import FastAPI, Request
|
3 |
-
from langchain_community.llms.ctransformers import CTransformers
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
MODEL_TYPE = 'mistral'
|
8 |
-
MODEL_BIN_PATH = "mistral-7b-instruct-v0.1.Q3_K_S.gguf"
|
9 |
-
MAX_NEW_TOKEN = 600
|
10 |
-
TEMPRATURE = 0.01
|
11 |
-
CONTEXT_LENGTH = 6000
|
12 |
-
|
13 |
|
14 |
-
llm = CTransformers(
|
15 |
-
model=MODEL_BIN_PATH,
|
16 |
-
config={
|
17 |
-
'max_new_tokens': MAX_NEW_TOKEN,
|
18 |
-
'temperature': TEMPRATURE,
|
19 |
-
'context_length': CONTEXT_LENGTH
|
20 |
-
},
|
21 |
-
model_type=MODEL_TYPE
|
22 |
-
)
|
23 |
|
24 |
app = FastAPI()
|
|
|
25 |
|
26 |
|
27 |
-
|
|
|
28 |
async def generate_text(request: Request):
|
29 |
raw_data = await request.body() # Get the raw body data from the request
|
30 |
prompt = raw_data.decode('utf-8')
|
31 |
if not prompt:
|
32 |
return {'error': 'Prompt is required'}, 400
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
|
|
|
|
|
1 |
+
|
2 |
+
from typing import Union
|
3 |
+
from wrapper import LLMWrapper
|
4 |
import uvicorn
|
5 |
from fastapi import FastAPI, Request
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
|
8 |
app = FastAPI()
|
9 |
+
llm_wrapper = LLMWrapper()
|
10 |
|
11 |
|
12 |
+
|
13 |
+
@app.post("/")
|
14 |
async def generate_text(request: Request):
|
15 |
raw_data = await request.body() # Get the raw body data from the request
|
16 |
prompt = raw_data.decode('utf-8')
|
17 |
if not prompt:
|
18 |
return {'error': 'Prompt is required'}, 400
|
19 |
+
|
20 |
+
generated_text = llm_wrapper.generate_text(prompt)
|
21 |
+
print(generated_text)
|
22 |
+
return {'generated_text': generated_text}
|
23 |
+
|
24 |
+
|
25 |
+
if __name__ == '__main__':
|
26 |
+
uvicorn.run(app, host='127.0.0.1', port=8001)
|