BeTaLabs commited on
Commit
2d2cdc1
·
verified ·
1 Parent(s): c886053

Update llm_handler.py

Browse files
Files changed (1) hide show
  1. llm_handler.py +2 -2
llm_handler.py CHANGED
@@ -4,7 +4,7 @@ from llama_cpp_agent import MessagesFormatterType
4
  from llama_cpp_agent.providers import LlamaCppPythonProvider
5
 
6
  # Initialize the Llama model
7
- llama_model = Llama("Qwen2-1.5B-Instruct-Abliterated-iMat-GGUF/Qwen2-1.5B-Instruct-Abliterated-Q8_0.gguf", n_batch=1024, n_threads=24, n_gpu_layers=33, n_ctx=4098, verbose=False)
8
 
9
  # Create the provider
10
  provider = LlamaCppPythonProvider(llama_model)
@@ -19,7 +19,7 @@ agent = LlamaCppAgent(
19
 
20
  # Set provider settings
21
  settings = provider.get_provider_default_settings()
22
- settings.max_tokens = 2000
23
  settings.stream = True
24
 
25
  def send_to_llm(provider, msg_list):
 
4
  from llama_cpp_agent.providers import LlamaCppPythonProvider
5
 
6
  # Initialize the Llama model
7
+ llama_model = Llama("Qwen2-500M-Instruct-GGUF/Qwen2-500M-Instruct-f32.gguf", n_batch=1024, n_threads=24, n_gpu_layers=33, n_ctx=4098, verbose=False)
8
 
9
  # Create the provider
10
  provider = LlamaCppPythonProvider(llama_model)
 
19
 
20
  # Set provider settings
21
  settings = provider.get_provider_default_settings()
22
+ settings.max_tokens = 4098
23
  settings.stream = True
24
 
25
  def send_to_llm(provider, msg_list):