Update llm_handler.py
Browse files- llm_handler.py +2 -2
llm_handler.py
CHANGED
@@ -4,7 +4,7 @@ from llama_cpp_agent import MessagesFormatterType
|
|
4 |
from llama_cpp_agent.providers import LlamaCppPythonProvider
|
5 |
|
6 |
# Initialize the Llama model
|
7 |
-
llama_model = Llama("Qwen2-
|
8 |
|
9 |
# Create the provider
|
10 |
provider = LlamaCppPythonProvider(llama_model)
|
@@ -19,7 +19,7 @@ agent = LlamaCppAgent(
|
|
19 |
|
20 |
# Set provider settings
|
21 |
settings = provider.get_provider_default_settings()
|
22 |
-
settings.max_tokens =
|
23 |
settings.stream = True
|
24 |
|
25 |
def send_to_llm(provider, msg_list):
|
|
|
4 |
from llama_cpp_agent.providers import LlamaCppPythonProvider
|
5 |
|
6 |
# Initialize the Llama model
|
7 |
+
llama_model = Llama("Qwen2-500M-Instruct-GGUF/Qwen2-500M-Instruct-f32.gguf", n_batch=1024, n_threads=24, n_gpu_layers=33, n_ctx=4098, verbose=False)
|
8 |
|
9 |
# Create the provider
|
10 |
provider = LlamaCppPythonProvider(llama_model)
|
|
|
19 |
|
20 |
# Set provider settings
|
21 |
settings = provider.get_provider_default_settings()
|
22 |
+
settings.max_tokens = 4098
|
23 |
settings.stream = True
|
24 |
|
25 |
def send_to_llm(provider, msg_list):
|