jaelin215 commited on
Commit
b28c76d
1 Parent(s): bfa5372

Changed default temperature to 0.5 and token length 128 * 4

Browse files
Files changed (1) hide show
  1. llm_response_generator.py +4 -4
llm_response_generator.py CHANGED
@@ -36,8 +36,8 @@ class LLLResponseGenerator():
36
  openai_model_name: str = "",
37
  # hf_repo_id: str = "tiiuae/falcon-7b-instruct",
38
  hf_repo_id: str = "mistralai/Mistral-7B-Instruct-v0.2",
39
- temperature: float = 0.1,
40
- max_length: int = 128,
41
  ) -> str:
42
  """Call HuggingFace/OpenAI model for inference
43
 
@@ -142,8 +142,8 @@ if __name__ == "__main__":
142
  Response;
143
  """
144
 
145
- temperature = 0.1
146
- max_length = 128
147
 
148
  model = LLLResponseGenerator()
149
 
 
36
  openai_model_name: str = "",
37
  # hf_repo_id: str = "tiiuae/falcon-7b-instruct",
38
  hf_repo_id: str = "mistralai/Mistral-7B-Instruct-v0.2",
39
+ temperature: float = 0.5,
40
+ max_length: int = 128 * 4,
41
  ) -> str:
42
  """Call HuggingFace/OpenAI model for inference
43
 
 
142
  Response;
143
  """
144
 
145
+ temperature = 0.5
146
+ max_length = 128 *4
147
 
148
  model = LLLResponseGenerator()
149