winglian commited on
Commit
9999ab8
1 Parent(s): 7c60b85

llama adapter doesn't support use_cache yet

Browse files
Files changed (1) hide show
  1. llama_lora/ui/inference_ui.py +1 -0
llama_lora/ui/inference_ui.py CHANGED
@@ -111,6 +111,7 @@ def do_inference(
111
  top_k=top_k,
112
  repetition_penalty=repetition_penalty,
113
  num_beams=num_beams,
 
114
  )
115
 
116
  generate_params = {
 
111
  top_k=top_k,
112
  repetition_penalty=repetition_penalty,
113
  num_beams=num_beams,
114
+ use_cache=False,
115
  )
116
 
117
  generate_params = {