Romain-Cosentino commited on
Commit
6686abd
1 Parent(s): 2cc6dec

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -29,7 +29,7 @@ if not torch.cuda.is_available():
29
 
30
  if torch.cuda.is_available():
31
  model_id = "tenyx/Llama3-TenyxChat-70B"
32
- model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype=torch.float16)
33
  # model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_8bit=True)
34
  tokenizer = AutoTokenizer.from_pretrained(model_id)
35
  tokenizer.use_default_system_prompt = False
@@ -121,7 +121,7 @@ chat_interface = gr.ChatInterface(
121
  minimum=1.0,
122
  maximum=2.0,
123
  step=0.05,
124
- value=1.,
125
  ),
126
  ],
127
  stop_btn=None,
 
29
 
30
  if torch.cuda.is_available():
31
  model_id = "tenyx/Llama3-TenyxChat-70B"
32
+ model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype=torch.bfloat16)
33
  # model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_8bit=True)
34
  tokenizer = AutoTokenizer.from_pretrained(model_id)
35
  tokenizer.use_default_system_prompt = False
 
121
  minimum=1.0,
122
  maximum=2.0,
123
  step=0.05,
124
+ value=1.2,
125
  ),
126
  ],
127
  stop_btn=None,