vedantpalit commited on
Commit
a88d836
·
verified ·
1 Parent(s): 207dd4d

RLHF model of StarCoder

Browse files
README.md CHANGED
@@ -49,5 +49,5 @@ The following hyperparameters were used during training:
49
 
50
  - Transformers 4.35.2
51
  - Pytorch 2.1.0+cu121
52
- - Datasets 2.16.1
53
  - Tokenizers 0.15.1
 
49
 
50
  - Transformers 4.35.2
51
  - Pytorch 2.1.0+cu121
52
+ - Datasets 2.17.0
53
  - Tokenizers 0.15.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2498dd934130cab6dfef49362869c4aa4353a760645bfb0d51b4f8eeacc48caf
3
  size 656601304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6ecfa625687ecbcce6ef6cf9e6699ce56d0ef0e1078eadb4c7463b3599517f3
3
  size 656601304
special_tokens_map.json CHANGED
@@ -22,6 +22,12 @@
22
  ],
23
  "bos_token": "<|endoftext|>",
24
  "eos_token": "<|endoftext|>",
25
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
26
  "unk_token": "<|endoftext|>"
27
  }
 
22
  ],
23
  "bos_token": "<|endoftext|>",
24
  "eos_token": "<|endoftext|>",
25
+ "pad_token": {
26
+ "content": "[PAD]",
27
+ "lstrip": false,
28
+ "normalized": false,
29
+ "rstrip": false,
30
+ "single_word": false
31
+ },
32
  "unk_token": "<|endoftext|>"
33
  }
tokenizer.json CHANGED
@@ -1,15 +1,20 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": {
5
  "strategy": {
6
- "Fixed": 32
7
  },
8
- "direction": "Left",
9
  "pad_to_multiple_of": null,
10
- "pad_id": 0,
11
  "pad_type_id": 0,
12
- "pad_token": "<|endoftext|>"
13
  },
14
  "added_tokens": [
15
  {
@@ -182,6 +187,15 @@
182
  "rstrip": false,
183
  "normalized": false,
184
  "special": true
 
 
 
 
 
 
 
 
 
185
  }
186
  ],
187
  "normalizer": null,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 256,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 256
12
  },
13
+ "direction": "Right",
14
  "pad_to_multiple_of": null,
15
+ "pad_id": 49152,
16
  "pad_type_id": 0,
17
+ "pad_token": "[PAD]"
18
  },
19
  "added_tokens": [
20
  {
 
187
  "rstrip": false,
188
  "normalized": false,
189
  "special": true
190
+ },
191
+ {
192
+ "id": 49152,
193
+ "content": "[PAD]",
194
+ "single_word": false,
195
+ "lstrip": false,
196
+ "rstrip": false,
197
+ "normalized": false,
198
+ "special": true
199
  }
200
  ],
201
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -152,6 +152,14 @@
152
  "rstrip": false,
153
  "single_word": false,
154
  "special": true
 
 
 
 
 
 
 
 
155
  }
156
  },
157
  "additional_special_tokens": [
@@ -179,8 +187,7 @@
179
  "clean_up_tokenization_spaces": true,
180
  "eos_token": "<|endoftext|>",
181
  "model_max_length": 1000000000000000019884624838656,
182
- "pad_token": "<|endoftext|>",
183
- "padding_side": "left",
184
  "tokenizer_class": "GPT2Tokenizer",
185
  "unk_token": "<|endoftext|>",
186
  "vocab_size": 49152
 
152
  "rstrip": false,
153
  "single_word": false,
154
  "special": true
155
+ },
156
+ "49152": {
157
+ "content": "[PAD]",
158
+ "lstrip": false,
159
+ "normalized": false,
160
+ "rstrip": false,
161
+ "single_word": false,
162
+ "special": true
163
  }
164
  },
165
  "additional_special_tokens": [
 
187
  "clean_up_tokenization_spaces": true,
188
  "eos_token": "<|endoftext|>",
189
  "model_max_length": 1000000000000000019884624838656,
190
+ "pad_token": "[PAD]",
 
191
  "tokenizer_class": "GPT2Tokenizer",
192
  "unk_token": "<|endoftext|>",
193
  "vocab_size": 49152
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dc9322b7a6d234418fd5eea0bd93c469cd4e6ee383d0fec25e325a499e75819
3
  size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:634f0d73674f856a40942422ae5689c52410dc84c10e66b5a50168bf529b2eb5
3
  size 4536