LC-Edward commited on
Commit
16514b0
β€’
1 Parent(s): 8de12f3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -28,6 +28,9 @@ license: llama2
28
  |Xwin-Math-70B-V1.0| 87.0 | 31.8 | πŸ€— <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.0" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
29
  |Xwin-Math-70B-V1.1| 90.6 | 51.9 | πŸ€— <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.1" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
30
 
 
 
 
31
  ## πŸš€ Benchmarks
32
  ### Xwin-Math performance on [MATH](https://github.com/hendrycks/math) and [GSM8K](https://github.com/openai/grade-school-math).
33
 
@@ -124,7 +127,7 @@ To generate the model's responses, you can use the `generate.py` script. Please
124
 
125
  For the generation process, we use the Vicuna-v1.1 system prompt with chain-of-thought and format instruction. We also employ a greedy decoding strategy and set the maximum sequence length to 2048.
126
  ```
127
- "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {instruction} Give your solution in detail. In the end, write your final answer in the format of 'The answer is: <ANSWER>.'. ASSISTANT: "
128
  ```
129
 
130
  Here is an simple example to generate using [vLLM](https://docs.vllm.ai/en/latest/).
 
28
  |Xwin-Math-70B-V1.0| 87.0 | 31.8 | πŸ€— <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.0" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
29
  |Xwin-Math-70B-V1.1| 90.6 | 51.9 | πŸ€— <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.1" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
30
 
31
+ * Xwin-Math-7B-V1.1 uses 1.92M GSM8K and 960K MATH synthetic data
32
+ * Xwin-Math-70B-V1.1 uses 960K GSM8K and 480KMATH synthetic data
33
+
34
  ## πŸš€ Benchmarks
35
  ### Xwin-Math performance on [MATH](https://github.com/hendrycks/math) and [GSM8K](https://github.com/openai/grade-school-math).
36
 
 
127
 
128
  For the generation process, we use the Vicuna-v1.1 system prompt with chain-of-thought and format instruction. We also employ a greedy decoding strategy and set the maximum sequence length to 2048.
129
  ```
130
+ "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {instruction} Give your solution in detail. In the end, write your final answer in the format of 'The answer is: <ANSWER>.'. ASSISTANT:"
131
  ```
132
 
133
  Here is an simple example to generate using [vLLM](https://docs.vllm.ai/en/latest/).