mayank-mishra
commited on
Commit
•
0a81cba
1
Parent(s):
862809c
update to r1.1
Browse files- config.json +2 -2
- generation_config.json +1 -1
- model-00001-of-00009.safetensors +1 -1
- model-00002-of-00009.safetensors +1 -1
- model-00003-of-00009.safetensors +1 -1
- model-00004-of-00009.safetensors +1 -1
- model-00005-of-00009.safetensors +1 -1
- model-00006-of-00009.safetensors +1 -1
- model-00007-of-00009.safetensors +1 -1
- model-00008-of-00009.safetensors +1 -1
- model-00009-of-00009.safetensors +1 -1
- tokenizer.json +1 -0
- tokenizer_config.json +1 -1
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"activation_function": "
|
3 |
"architectures": [
|
4 |
"GPTBigCodeForCausalLM"
|
5 |
],
|
@@ -22,7 +22,7 @@
|
|
22 |
"scale_attention_softmax_in_fp32": true,
|
23 |
"scale_attn_weights": true,
|
24 |
"torch_dtype": "bfloat16",
|
25 |
-
"transformers_version": "4.
|
26 |
"use_cache": true,
|
27 |
"vocab_size": 49152
|
28 |
}
|
|
|
1 |
{
|
2 |
+
"activation_function": "gelu",
|
3 |
"architectures": [
|
4 |
"GPTBigCodeForCausalLM"
|
5 |
],
|
|
|
22 |
"scale_attention_softmax_in_fp32": true,
|
23 |
"scale_attn_weights": true,
|
24 |
"torch_dtype": "bfloat16",
|
25 |
+
"transformers_version": "4.41.2",
|
26 |
"use_cache": true,
|
27 |
"vocab_size": 49152
|
28 |
}
|
generation_config.json
CHANGED
@@ -3,5 +3,5 @@
|
|
3 |
"bos_token_id": 0,
|
4 |
"eos_token_id": 0,
|
5 |
"pad_token_id": 0,
|
6 |
-
"transformers_version": "4.
|
7 |
}
|
|
|
3 |
"bos_token_id": 0,
|
4 |
"eos_token_id": 0,
|
5 |
"pad_token_id": 0,
|
6 |
+
"transformers_version": "4.41.2"
|
7 |
}
|
model-00001-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4952185256
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22b893e90be19cc42647f7d76caef1bd7101b6904df709b93472265fa6e41594
|
3 |
size 4952185256
|
model-00002-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4930227792
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc7f98d43d5aa8531381ffd2b7995008f1471a707a57becf638d089c549edec8
|
3 |
size 4930227792
|
model-00003-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4927118464
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1f56c8969bbdc7995f1c61123f4b63da7ff202a93e40fde98121ec9ccdc49556
|
3 |
size 4927118464
|
model-00004-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4930227864
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd7060b8fac28950098ce792fc518a0a8af296f47e22709ccaad1ce8216591ea
|
3 |
size 4930227864
|
model-00005-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4927118464
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bbd071ecf07518991d9d82749d9c864a7e923eb82450048565e9121604f4cbe2
|
3 |
size 4927118464
|
model-00006-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4930227864
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b3a0e75fbd4db5862a20e23ce6ffbf69a4a28d59505085b8d82c1d61620cbf98
|
3 |
size 4930227864
|
model-00007-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4927118464
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8fee8bebaf2202e65cd8253898b317fc4a03a5a5cfb5bd6a01ccbf76a65d7c8b
|
3 |
size 4927118464
|
model-00008-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4930227864
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1fa8ce48dbb754a3e762d10d3590c65fb0eb30b6f28044f37ad184c335ff6916
|
3 |
size 4930227864
|
model-00009-of-00009.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 679601176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:62f5a5189fb992fb52b72861f432dc19123bef17f78c42e1db8a89b208b4b5ab
|
3 |
size 679601176
|
tokenizer.json
CHANGED
@@ -206,6 +206,7 @@
|
|
206 |
"end_of_word_suffix": null,
|
207 |
"fuse_unk": false,
|
208 |
"byte_fallback": false,
|
|
|
209 |
"vocab": {
|
210 |
"<|endoftext|>": 0,
|
211 |
"<fim_prefix>": 1,
|
|
|
206 |
"end_of_word_suffix": null,
|
207 |
"fuse_unk": false,
|
208 |
"byte_fallback": false,
|
209 |
+
"ignore_merges": false,
|
210 |
"vocab": {
|
211 |
"<|endoftext|>": 0,
|
212 |
"<fim_prefix>": 1,
|
tokenizer_config.json
CHANGED
@@ -178,7 +178,7 @@
|
|
178 |
"bos_token": "<|endoftext|>",
|
179 |
"clean_up_tokenization_spaces": true,
|
180 |
"eos_token": "<|endoftext|>",
|
181 |
-
"model_max_length":
|
182 |
"pad_token": "<|endoftext|>",
|
183 |
"padding_side": "left",
|
184 |
"tokenizer_class": "GPT2Tokenizer",
|
|
|
178 |
"bos_token": "<|endoftext|>",
|
179 |
"clean_up_tokenization_spaces": true,
|
180 |
"eos_token": "<|endoftext|>",
|
181 |
+
"model_max_length": 8192,
|
182 |
"pad_token": "<|endoftext|>",
|
183 |
"padding_side": "left",
|
184 |
"tokenizer_class": "GPT2Tokenizer",
|