File size: 5,052 Bytes
8dc4741
0d7872b
 
 
 
 
 
 
8dc4741
 
0d7872b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
---
language: 
  - en
tags:
  - deepspeed
  - chatgpt
  - opt
  - sft
license: apache-2.0
---
---

# ChatGPT OPT 1.3B DeepSpeed Supervised fine tuning

*fsalab-chat-opt-1.3b-sft-deepspeed*

This model consists of the first step of a modified pipeline the to the traditional training process of Chat-GPT models, which is comprised of a three-step procedure of **supervised fine tuning**, [reward model](https://huggingface.co/FSALab/fsalab-chat-opt-350m-reward-deepspeed) and [RLHF](https://huggingface.co/FSALab/fsalab-chat-opt-1.3b-rlhf-deepspeed).

This project's main goal was to make proper use of existing frameworks that revolve around the minimisation of training costs and thus the eventual improvements towards both the feasibility and usability of ChatGPT-like models. The framework selected here is DeepSpeed which has been instrumental in the development of this model and through this framework it was possible to train the ChatGPT-like model on much larger data-sets with a reasonable number of GPUs and consequently achieve significantly better performance.

This model follows the blog of ChatGPT and the paper of InstructGPT and especially the [Microsoft DeepSpeed Chat Blog](https://github.com/microsoft/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat).

## Our Training Methodology and Speedup Recipes
  
This training process is broken up into three key steps:

1. **Supervised fine-tuning (SFT):** In the first step we perform supervised fine tuning by taking the pretrained models, configuring them to use smaller learning rates and then subsequently trained on a labelled data-set.

2. **Reward Model (RM) fine-tuning:** See  [here](https://huggingface.co/FSALab/fsalab-chat-opt-350m-reward-deepspeed) 

3. **Reinforcement-learning from Human feedback (RLHF) fine-tuning:** See  [here](https://huggingface.co/FSALab/fsalab-chat-opt-1.3b-rlhf-deepspeed) 

To view the details behind each step head into their respective links and view the model card there.
  
## Supervised fine tuning training configuration:

**Model Configurations:**

| Parameter              | Value |
|:-----------------------|:------|
| Parameters             | 1.3B  |
| Model type             | OPT   |
| FFN Dimensions         | 8192  |
| Hidden Size            | 2048  |
| Max Position Embedding | 2048  |
| Attention Heads        | 32    |
| Hidden layers          | 24    |


**Training Configurations:**
| Parameter              | Value |
|:-----------------------|:------|
| Train Batch size       | 32    |
| Train micro batch size | 4     |
| ZeRO stage             | 2     |
| FP16                   | True  |
| Gradient clipping      | 1.0   |
| Dropout                | 0.1   |
| Bias                   | True  |
| Prescale gradients     | True  |


## Installation

If using through the HuggingFace transformers library:

``` python
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("FSALab/deepspeed-chatgpt-opt1.3b-sft")

model = AutoModelForCausalLM.from_pretrained("FSALab/deepspeed-chatgpt-opt1.3b-sft")
```


If you would like to clone from source:
```bash
# Make sure you have git-lfs installed (https://git-lfs.github.com)
git lfs install
git clone https://huggingface.co/FSALab/deepspeed-chatgpt-opt1.3b-sft

# if you want to clone without large files – just their pointers
# prepend your git clone with the following env var:
GIT_LFS_SKIP_SMUDGE=1
```

## Why did we choose DeepSpeed?

**DeepSpeed Training:** 

The `main.py` Python code take the DeepSpeed config with the argument `--deepspeed_config ./ds_config.json`.

We read up on the DeepSpeed documentation and created a specific coniguration based on their work. The json file `ds_config.json` here is set to take the [ZeRO-2](https://www.microsoft.com/en-us/research/blog/ZeRO-2-deepspeed-shattering-barriers-of-deep-learning-speed-scale/) stage and FP16, allowing must faster training and GPU memory saving. Note that ZeRO-2 is just one of the examples using our DeepSpeed. You may use ZeRO-1, Zero-3, ZeRO-Offload and ZeRO-infinity. For more information on DeepSpeed ZeRO family, please see this [tutorial link](https://www.deepspeed.ai/tutorials/zero/) for Zero-1/2/3 and this [tutorial ](https://www.deepspeed.ai/tutorials/zero-offload/)for Zero-Offload.

To enable the DeepSpeed Zero family training, we injected several lines of code in order to enable this i.e.:

```python
model, optimizer, _, lr_scheduler = deepspeed.initialize(model=model, \
  optimizer=optimizer, \
  args=args,       \
  lr_scheduler=lr_scheduler,  \
  dist_init_required=True)
```



## **Acknowledgements**

We thank the following papers and open-source repositories. We especially thank DeepSpeed for their frameworks as well.

* [1] Schulman, John, et al. "Introducing ChatGPT", https://openai.com/blog/chatgpt (2022).
* [2] Transformers [Hugging Face (github.com)](https://github.com/huggingface)
* [3] DeepSpeed Chat [DeepSpeed Chat](https://github.com/microsoft/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat)