Yersel commited on
Commit
1a908a5
·
1 Parent(s): 65adc46

adapt chatbot

Browse files
Files changed (1) hide show
  1. app.py +5 -7
app.py CHANGED
@@ -6,7 +6,7 @@ import gradio as gr
6
  model_id = "Qwen/Qwen3-1.7B"
7
  model = AutoModelForCausalLM.from_pretrained(
8
  model_id,
9
- torch_dtype = torch.bfloat16
10
  )
11
  tokenizer = AutoTokenizer.from_pretrained(model_id)
12
 
@@ -38,17 +38,15 @@ def respuesta(
38
 
39
  input_ids = tokenizer.apply_chat_template(
40
  messages,
 
41
  add_generation_prompt=True,
42
- return_tensors='pt'
43
  ).to(model.device)
44
 
45
- # terminators = [
46
- # tokenizer.eos_token_id,
47
- # tokenizer.convert_tokens_to_ids("<|eot_id|>")
48
- # ]
49
 
50
  outputs = model.generate(
51
- input_ids,
52
  max_new_tokens=max_tokens,
53
  do_sample=True,
54
  temperature=temperature,
 
6
  model_id = "Qwen/Qwen3-1.7B"
7
  model = AutoModelForCausalLM.from_pretrained(
8
  model_id,
9
+ torch_dtype = "auto"
10
  )
11
  tokenizer = AutoTokenizer.from_pretrained(model_id)
12
 
 
38
 
39
  input_ids = tokenizer.apply_chat_template(
40
  messages,
41
+ tokenize=False,
42
  add_generation_prompt=True,
43
+ enable_thinking=True
44
  ).to(model.device)
45
 
46
+ model_inputs = tokenizer([text], return_tensor='pt')
 
 
 
47
 
48
  outputs = model.generate(
49
+ **model_inputs,
50
  max_new_tokens=max_tokens,
51
  do_sample=True,
52
  temperature=temperature,