UphamProjects commited on
Commit
010d317
1 Parent(s): 77439e3

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -139
README.md DELETED
@@ -1,139 +0,0 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- tags:
6
- - sft
7
- pipeline_tag: text-generation
8
- widget:
9
- - text: <|prompter|>What is a meme, and what's the history behind this word?<|endoftext|><|assistant|>
10
- - text: <|prompter|>What's the Earth total population<|endoftext|><|assistant|>
11
- - text: <|prompter|>Write a story about future of AI development<|endoftext|><|assistant|>
12
- ---
13
-
14
- # Open-Assistant SFT-4 12B Model
15
-
16
-
17
- This is the 4th iteration English supervised-fine-tuning (SFT) model of
18
- the [Open-Assistant](https://github.com/LAION-AI/Open-Assistant) project.
19
- It is based on a Pythia 12B that was fine-tuned on human demonstrations
20
- of assistant conversations collected through the
21
- [https://open-assistant.io/](https://open-assistant.io/) human feedback web
22
- app before March 25, 2023.
23
-
24
- ## Model Details
25
-
26
- - **Developed by:** [Open-Assistant Contributors](https://open-assistant.io/)
27
- - **Model type:** Transformer-based Language Model
28
- - **Language:** English
29
- - **Finetuned from:** [EleutherAI / pythia-12b-deduped](https://huggingface.co/EleutherAI/pythia-12b-deduped)
30
- - **Code:** [Open-Assistant/model/model_training](https://github.com/LAION-AI/Open-Assistant/tree/main/model/model_training)
31
- - **Demo:** [Continuations for 250 random prompts](https://open-assistant.github.io/oasst-model-eval/?f=https%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Foasst-sft%2F2023-04-03_andreaskoepf_oasst-sft-4-pythia-12b-epoch-3_5_sampling_noprefix_lottery.json%0Ahttps%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Fchat-gpt%2F2023-04-11_gpt-3.5-turbo_lottery.json)
32
- - **License:** Apache 2.0
33
- - **Contact:** [Open-Assistant Discord](https://ykilcher.com/open-assistant-discord)
34
-
35
- ## Prompting
36
-
37
- Two special tokens are used to mark the beginning of user and assistant turns:
38
- `<|prompter|>` and `<|assistant|>`. Each turn ends with a `<|endoftext|>` token.
39
-
40
- Input prompt example:
41
- ```
42
- <|prompter|>What is a meme, and what's the history behind this word?<|endoftext|><|assistant|>
43
- ```
44
- The input ends with the `<|assistant|>` token to signal that the model should
45
- start generating the assistant reply.
46
-
47
-
48
- ## Dev Details
49
-
50
- - wandb: https://wandb.ai/open-assistant/supervised-finetuning/runs/770a0t41
51
- - base model: [andreaskoepf/pythia-12b-pre-2000](https://huggingface.co/andreaskoepf/pythia-12b-pre-2000)
52
- - checkpoint: 4000 steps
53
-
54
- command: `deepspeed trainer_sft.py --configs defaults reference-data reference-pythia-12b --cache_dir /home/ubuntu/data_cache --output_dir .saved/oasst-sft-3-pythia-12b-reference_2kpre --num_train_epochs 8 --residual_dropout 0.2 --deepspeed --use_flash_attention true --model_name andreaskoepf/pythia-12b-pre-2000`
55
-
56
- data:
57
- ```
58
- reference-data:
59
- datasets:
60
- - oasst_export:
61
- lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk"
62
- input_file_path: 2023-03-25_oasst_research_ready_synth_labels.jsonl.gz
63
- val_split: 0.05
64
- - alpaca
65
- sort_by_length: false
66
- use_custom_sampler: false
67
- ```
68
-
69
-
70
- pythia:
71
- ```
72
- reference-pythia-12b:
73
- dtype: fp16
74
- log_dir: "pythia_log_12b"
75
- learning_rate: 6e-6
76
- model_name: EleutherAI/pythia-12b-deduped
77
- output_dir: pythia_model_12b
78
- weight_decay: 0.0
79
- max_length: 2048
80
- warmup_steps: 100
81
- gradient_checkpointing: true
82
- gradient_accumulation_steps: 2
83
- per_device_train_batch_size: 4
84
- per_device_eval_batch_size: 4
85
- eval_steps: 100
86
- save_steps: 1000
87
- num_train_epochs: 8
88
- save_total_limit: 4
89
- ```
90
-
91
- zero config:
92
- ```
93
- {
94
- "fp16": {
95
- "enabled": "auto",
96
- "loss_scale": 0,
97
- "loss_scale_window": 1000,
98
- "initial_scale_power": 16,
99
- "hysteresis": 2,
100
- "min_loss_scale": 1
101
- },
102
- "bf16": {
103
- "enabled": "auto"
104
- },
105
- "optimizer": {
106
- "type": "AdamW",
107
- "params": {
108
- "lr": "auto",
109
- "betas": "auto",
110
- "eps": "auto",
111
- "weight_decay": "auto"
112
- }
113
- },
114
- "scheduler": {
115
- "type": "WarmupDecayLR",
116
- "params": {
117
- "warmup_min_lr": "auto",
118
- "warmup_max_lr": "auto",
119
- "warmup_num_steps": "auto",
120
- "total_num_steps": "auto"
121
- }
122
- },
123
- "zero_optimization": {
124
- "stage": 2,
125
- "allgather_partitions": true,
126
- "allgather_bucket_size": 1e9,
127
- "overlap_comm": false,
128
- "reduce_scatter": true,
129
- "reduce_bucket_size": 1e9,
130
- "contiguous_gradients": true
131
- },
132
- "gradient_accumulation_steps": "auto",
133
- "gradient_clipping": "auto",
134
- "steps_per_print": 2000,
135
- "train_batch_size": "auto",
136
- "train_micro_batch_size_per_gpu": "auto",
137
- "wall_clock_breakdown": false
138
- }
139
- ```