ffreemt commited on
Commit
0e7c90d
1 Parent(s): 1132eaf

Update block.load to update buff Textbox

Browse files
Files changed (1) hide show
  1. app.py +8 -2
app.py CHANGED
@@ -219,6 +219,12 @@ logger.info("start dl")
219
  _ = """full url: https://huggingface.co/TheBloke/mpt-30B-chat-GGML/blob/main/mpt-30b-chat.ggmlv0.q4_1.bin"""
220
 
221
  repo_id = "TheBloke/mpt-30B-chat-GGML"
 
 
 
 
 
 
222
  model_filename = "mpt-30b-chat.ggmlv0.q4_1.bin"
223
  destination_folder = "models"
224
 
@@ -431,8 +437,8 @@ with gr.Blocks(
431
  show_progress="full",
432
  )
433
 
434
- # update buff Textbox
435
- block.load(lambda: ns.response, [], [buff])
436
 
437
  # concurrency_count=5, max_size=20
438
  # max_size=36, concurrency_count=14
 
219
  _ = """full url: https://huggingface.co/TheBloke/mpt-30B-chat-GGML/blob/main/mpt-30b-chat.ggmlv0.q4_1.bin"""
220
 
221
  repo_id = "TheBloke/mpt-30B-chat-GGML"
222
+
223
+ # https://huggingface.co/TheBloke/mpt-30B-chat-GGML
224
+ _ = """
225
+ mpt-30b-chat.ggmlv0.q4_0.bin q4_0 4 16.85 GB 19.35 GB 4-bit.
226
+ mpt-30b-chat.ggmlv0.q4_1.bin q4_1 4 18.73 GB 21.23 GB 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.
227
+ """
228
  model_filename = "mpt-30b-chat.ggmlv0.q4_1.bin"
229
  destination_folder = "models"
230
 
 
437
  show_progress="full",
438
  )
439
 
440
+ # update buff Textbox, every: units in seconds)
441
+ block.run_forever(lambda: ns.response, None, [buff], every=1)
442
 
443
  # concurrency_count=5, max_size=20
444
  # max_size=36, concurrency_count=14