dball commited on
Commit
ffcebba
1 Parent(s): a46692a

Manually fix minor details in README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -13
README.md CHANGED
@@ -17,8 +17,8 @@ model-index:
17
 
18
  # zephyr-7b-sft-qlora
19
 
20
- This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mist\
21
- ral-7B-v0.1) on the HuggingFaceH4/ultrachat_200k dataset.
22
  It achieves the following results on the evaluation set:
23
  - Loss: 0.9523
24
 
@@ -27,23 +27,17 @@ It achieves the following results on the evaluation set:
27
  QLoRA SFT via
28
  ```
29
  # Step 1 - SFT
30
- ACCELERATE_LOG_LEVEL=info accelerate launch --config_file recipes/accelerate_configs/multi_gpu.yaml --n\
31
- um_processes=1 scripts/run_sft.py recipes/zephyr-7b-beta/sft/config_qlora.yaml --load_in_4bit=true
32
  ```
33
  see https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/README.md
34
 
35
  ## Intended uses & limitations
36
 
37
  ```
38
- chat_template: "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + me\
39
- ssage['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['c\
40
- ontent'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['co\
41
- ntent'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }\
42
- }\n{% endif %}\n{% endfor %}"
43
  ```
44
 
45
- see https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/sft/config_qlora\
46
- .yaml
47
 
48
  ## Training and evaluation data
49
 
@@ -55,8 +49,7 @@ dataset_splits:
55
  - test_sft
56
  ```
57
 
58
- see https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/sft/config_qlora\
59
- .yaml
60
 
61
  ## Training procedure
62
 
 
17
 
18
  # zephyr-7b-sft-qlora
19
 
20
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the HuggingFaceH4ultrachat_200k dataset.
21
+ It is the first step (Step 1 SFT, see below) of building Zephyr, i.e. before DPO.
22
  It achieves the following results on the evaluation set:
23
  - Loss: 0.9523
24
 
 
27
  QLoRA SFT via
28
  ```
29
  # Step 1 - SFT
30
+ ACCELERATE_LOG_LEVEL=info accelerate launch --config_file recipes/accelerate_configs/multi_gpu.yaml --num_processes=1 scripts/run_sft.py recipes/zephyr-7b-beta/sft/config_qlora.yaml --load_in_4bit=true
 
31
  ```
32
  see https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/README.md
33
 
34
  ## Intended uses & limitations
35
 
36
  ```
37
+ chat_template: "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}"
 
 
 
 
38
  ```
39
 
40
+ see https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/sft/config_qlora.yaml
 
41
 
42
  ## Training and evaluation data
43
 
 
49
  - test_sft
50
  ```
51
 
52
+ see https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/sft/config_qlora.yaml
 
53
 
54
  ## Training procedure
55