Update main.py
Browse files
main.py
CHANGED
@@ -858,8 +858,7 @@ async def connexion_vector_database():
|
|
858 |
@cl.step(type="llm")
|
859 |
async def IA():
|
860 |
os.environ['HUGGINGFACEHUB_API_TOKEN'] = os.environ['HUGGINGFACEHUB_API_TOKEN']
|
861 |
-
|
862 |
-
repo_id = "meta-llama/Llama-3.2-3B-Instruct"
|
863 |
llm = HuggingFaceEndpoint(
|
864 |
repo_id=repo_id, max_new_tokens=5000, temperature=1.0, task="text2text-generation", streaming=True
|
865 |
)
|
@@ -898,8 +897,7 @@ async def setup_agent(settings):
|
|
898 |
contextChat = "Il n'y a pas de contexte."
|
899 |
|
900 |
os.environ['HUGGINGFACEHUB_API_TOKEN'] = os.environ['HUGGINGFACEHUB_API_TOKEN']
|
901 |
-
|
902 |
-
repo_id = "meta-llama/Llama-3.2-3B-Instruct"
|
903 |
model = HuggingFaceEndpoint(
|
904 |
repo_id=repo_id,
|
905 |
max_new_tokens=3600,
|
|
|
858 |
@cl.step(type="llm")
|
859 |
async def IA():
|
860 |
os.environ['HUGGINGFACEHUB_API_TOKEN'] = os.environ['HUGGINGFACEHUB_API_TOKEN']
|
861 |
+
repo_id = "mistralai/Mistral-7B-Instruct-v0.3"
|
|
|
862 |
llm = HuggingFaceEndpoint(
|
863 |
repo_id=repo_id, max_new_tokens=5000, temperature=1.0, task="text2text-generation", streaming=True
|
864 |
)
|
|
|
897 |
contextChat = "Il n'y a pas de contexte."
|
898 |
|
899 |
os.environ['HUGGINGFACEHUB_API_TOKEN'] = os.environ['HUGGINGFACEHUB_API_TOKEN']
|
900 |
+
repo_id = "mistralai/Mistral-7B-Instruct-v0.3"
|
|
|
901 |
model = HuggingFaceEndpoint(
|
902 |
repo_id=repo_id,
|
903 |
max_new_tokens=3600,
|