alexkueck commited on
Commit
c10830c
1 Parent(s): aa47f31

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -85,7 +85,7 @@ MODEL_NAME = "gpt-3.5-turbo-16k"
85
 
86
  #HuggingFace--------------------------------
87
  #repo_id = "meta-llama/Llama-2-13b-chat-hf"
88
- repo_id = "HuggingFaceH4/zephyr-7b-alpha"
89
  #repo_id = "meta-llama/Llama-2-70b-chat-hf"
90
  #repo_id = "tiiuae/falcon-40b"
91
  #repo_id = "Vicuna-33b"
@@ -304,9 +304,9 @@ def invoke (prompt, history, rag_option, openai_api_key, temperature=0.9, max_n
304
  #LLM auswählen (OpenAI oder HF)
305
  ###########################
306
  #Anfrage an OpenAI ----------------------------
307
- #llm = ChatOpenAI(model_name = MODEL_NAME, openai_api_key = openai_api_key, temperature=temperature)#, top_p = top_p)
308
  #oder an Hugging Face --------------------------
309
- llm = HuggingFaceHub(repo_id=repo_id, model_kwargs={"temperature": 0.5, "max_length": 64})
310
  #llm = HuggingFaceHub(url_??? = "https://wdgsjd6zf201mufn.us-east-1.aws.endpoints.huggingface.cloud", model_kwargs={"temperature": 0.5, "max_length": 64})
311
  #llm = HuggingFaceTextGenInference( inference_server_url="http://localhost:8010/", max_new_tokens=max_new_tokens,top_k=10,top_p=top_p,typical_p=0.95,temperature=temperature,repetition_penalty=repetition_penalty,)
312
 
 
85
 
86
  #HuggingFace--------------------------------
87
  #repo_id = "meta-llama/Llama-2-13b-chat-hf"
88
+ repo_id = "HuggingFaceH4/zephyr-7b-alpha" #das Modell ist echt gut!!!
89
  #repo_id = "meta-llama/Llama-2-70b-chat-hf"
90
  #repo_id = "tiiuae/falcon-40b"
91
  #repo_id = "Vicuna-33b"
 
304
  #LLM auswählen (OpenAI oder HF)
305
  ###########################
306
  #Anfrage an OpenAI ----------------------------
307
+ llm = ChatOpenAI(model_name = MODEL_NAME, openai_api_key = openai_api_key, temperature=temperature)#, top_p = top_p)
308
  #oder an Hugging Face --------------------------
309
+ #llm = HuggingFaceHub(repo_id=repo_id, model_kwargs={"temperature": 0.5, "max_length": 64})
310
  #llm = HuggingFaceHub(url_??? = "https://wdgsjd6zf201mufn.us-east-1.aws.endpoints.huggingface.cloud", model_kwargs={"temperature": 0.5, "max_length": 64})
311
  #llm = HuggingFaceTextGenInference( inference_server_url="http://localhost:8010/", max_new_tokens=max_new_tokens,top_k=10,top_p=top_p,typical_p=0.95,temperature=temperature,repetition_penalty=repetition_penalty,)
312