zqh11 commited on
Commit
c2a7935
1 Parent(s): 41ed9dd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -28,6 +28,9 @@ Here give some examples of how to use our model.
28
 
29
  **Chat Completion**
30
 
 
 
 
31
 
32
  ```python
33
  import torch
@@ -40,7 +43,7 @@ model.generation_config = GenerationConfig.from_pretrained(model_name)
40
  model.generation_config.pad_token_id = model.generation_config.eos_token_id
41
 
42
  messages = [
43
- {"role": "user", "content": "what is the integral of x^2 from 0 to 2?"}
44
  ]
45
  input_tensor = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
46
  outputs = model.generate(input_tensor.to(model.device), max_new_tokens=100)
 
28
 
29
  **Chat Completion**
30
 
31
+ ❗❗❗ Please use chain-of-thought prompt to test DeepSeekMath-Instruct and DeepSeekMath-RL:
32
+ * English questions: `{question}\nPlease reason step by step, and put your final answer within \\boxed{}.`
33
+ * Chinese questions: `{question}\n请通过逐步推理来解答问题,并把最终答案放置于\\boxed{}中。`
34
 
35
  ```python
36
  import torch
 
43
  model.generation_config.pad_token_id = model.generation_config.eos_token_id
44
 
45
  messages = [
46
+ {"role": "user", "content": "what is the integral of x^2 from 0 to 2?\nPlease reason step by step, and put your final answer within \\boxed{}."}
47
  ]
48
  input_tensor = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
49
  outputs = model.generate(input_tensor.to(model.device), max_new_tokens=100)