Hristo ZHANG 张鹤立 commited on
Commit
cc4e355
1 Parent(s): a023bcb
Files changed (2) hide show
  1. app.py +6 -3
  2. requirements.txt +1 -1
app.py CHANGED
@@ -20,8 +20,8 @@ parser.add_argument("-t", "--threads", default=0, type=int, help="number of thre
20
  parser.add_argument("--plain", action="store_true", help="display in plain text without markdown support")
21
  args = parser.parse_args()
22
 
23
- from pyllamacpp.model import Model
24
- model = Model(model_path=model_file)
25
 
26
 
27
 
@@ -41,7 +41,10 @@ def predict(input, system_prompt, chatbot, max_length, ctx_length, top_p, temper
41
  num_threads=0,
42
  stream=True,
43
  )
44
- for response_piece in model.generate(input):
 
 
 
45
  response += response_piece
46
  chatbot[-1] = (chatbot[-1][0], response)
47
 
 
20
  parser.add_argument("--plain", action="store_true", help="display in plain text without markdown support")
21
  args = parser.parse_args()
22
 
23
+ from llama_cpp import Llama
24
+ llm = Llama(model_path=model_file)
25
 
26
 
27
 
 
41
  num_threads=0,
42
  stream=True,
43
  )
44
+ output = llm(input)
45
+ response = output['choices'][0]['text']
46
+
47
+ for response_piece in response:
48
  response += response_piece
49
  chatbot[-1] = (chatbot[-1][0], response)
50
 
requirements.txt CHANGED
@@ -1,2 +1,2 @@
1
  gradio
2
- git+https://github.com/zhangheli/pyllamacpp.git@main
 
1
  gradio
2
+ llama_cpp_python==0.2.14