RaghulDevaraj commited on
Commit
a989aa9
1 Parent(s): 8282180

Adding ONNX file of this model

Browse files

Beep boop I am the [ONNX export bot 🤖🏎️](https://huggingface.co/spaces/onnx/export). On behalf of [RaghulDevaraj](https://huggingface.co/RaghulDevaraj), I would like to add to this repository the model converted to ONNX.

What is ONNX? It stands for "Open Neural Network Exchange", and is the most commonly used open standard for machine learning interoperability. You can find out more at [onnx.ai](https://onnx.ai/)!

The exported ONNX model can be then be consumed by various backends as TensorRT or TVM, or simply be used in a few lines with 🤗 Optimum through ONNX Runtime, check out how [here](https://huggingface.co/docs/optimum/main/en/onnxruntime/usage_guides/models)!

.gitattributes CHANGED
@@ -17,3 +17,10 @@
17
  pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
18
  rust_model.ot filter=lfs diff=lfs merge=lfs -text
19
  model.safetensors filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
17
  pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
18
  rust_model.ot filter=lfs diff=lfs merge=lfs -text
19
  model.safetensors filter=lfs diff=lfs merge=lfs -text
20
+ onnx/Constant_124_attr__value filter=lfs diff=lfs merge=lfs -text
21
+ onnx/Constant_268_attr__value filter=lfs diff=lfs merge=lfs -text
22
+ onnx/Constant_134_attr__value filter=lfs diff=lfs merge=lfs -text
23
+ onnx/Constant_280_attr__value filter=lfs diff=lfs merge=lfs -text
24
+ onnx/decoder_model_merged.onnx_data filter=lfs diff=lfs merge=lfs -text
25
+ onnx/decoder_with_past_model.onnx_data filter=lfs diff=lfs merge=lfs -text
26
+ onnx/decoder_model.onnx_data filter=lfs diff=lfs merge=lfs -text
onnx/Constant_124_attr__value ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83dd5f1a5a07912d2e7dbb73b75edeb2ce4c1f1d6ec48144a887bd4a4f72325b
3
+ size 4194304
onnx/Constant_134_attr__value ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83dd5f1a5a07912d2e7dbb73b75edeb2ce4c1f1d6ec48144a887bd4a4f72325b
3
+ size 4194304
onnx/Constant_268_attr__value ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6d5bc4e8a251949dd137f42edb53ba241856b24f1297d69ff00d93ee13599d8
3
+ size 4194304
onnx/Constant_280_attr__value ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6d5bc4e8a251949dd137f42edb53ba241856b24f1297d69ff00d93ee13599d8
3
+ size 4194304
onnx/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 50256
3
+ }
onnx/config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "EleutherAI/gpt-neo-1.3B",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPTNeoForCausalLM"
6
+ ],
7
+ "attention_dropout": 0,
8
+ "attention_layers": [
9
+ "global",
10
+ "local",
11
+ "global",
12
+ "local",
13
+ "global",
14
+ "local",
15
+ "global",
16
+ "local",
17
+ "global",
18
+ "local",
19
+ "global",
20
+ "local",
21
+ "global",
22
+ "local",
23
+ "global",
24
+ "local",
25
+ "global",
26
+ "local",
27
+ "global",
28
+ "local",
29
+ "global",
30
+ "local",
31
+ "global",
32
+ "local"
33
+ ],
34
+ "attention_types": [
35
+ [
36
+ [
37
+ "global",
38
+ "local"
39
+ ],
40
+ 12
41
+ ]
42
+ ],
43
+ "bos_token_id": 50256,
44
+ "classifier_dropout": 0.1,
45
+ "embed_dropout": 0,
46
+ "eos_token_id": 50256,
47
+ "gradient_checkpointing": false,
48
+ "hidden_size": 2048,
49
+ "initializer_range": 0.02,
50
+ "intermediate_size": null,
51
+ "layer_norm_epsilon": 1e-05,
52
+ "max_position_embeddings": 2048,
53
+ "model_type": "gpt_neo",
54
+ "num_heads": 16,
55
+ "num_layers": 24,
56
+ "resid_dropout": 0,
57
+ "summary_activation": null,
58
+ "summary_first_dropout": 0.1,
59
+ "summary_proj_to_labels": true,
60
+ "summary_type": "cls_index",
61
+ "summary_use_proj": true,
62
+ "task_specific_params": {
63
+ "text-generation": {
64
+ "do_sample": true,
65
+ "max_length": 50,
66
+ "temperature": 0.9
67
+ }
68
+ },
69
+ "tokenizer_class": "GPT2Tokenizer",
70
+ "transformers_version": "4.34.0",
71
+ "use_cache": true,
72
+ "vocab_size": 50257,
73
+ "window_size": 256
74
+ }
onnx/decoder_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df33d5e5d8b01d35cfad457684222a136ae3ce642511bd688e08e5e3b2ae6f0f
3
+ size 9058339
onnx/decoder_model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3809f77df9740bfcf5c379e53751ecf3c4bd37039ad7638d4a7afddad2e3357
3
+ size 5674008576
onnx/decoder_model_merged.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02ecbe162b7be1fe5175f029dae123d42e25d0ba609d528e3c798fd4a42df9d9
3
+ size 18111206
onnx/decoder_model_merged.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3809f77df9740bfcf5c379e53751ecf3c4bd37039ad7638d4a7afddad2e3357
3
+ size 5674008576
onnx/decoder_with_past_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a73309ee31f244ea0d34e05401664caecaca6fd5a689b4d20aef097bae726748
3
+ size 9076373
onnx/decoder_with_past_model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3809f77df9740bfcf5c379e53751ecf3c4bd37039ad7638d4a7afddad2e3357
3
+ size 5674008576
onnx/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.34.0"
6
+ }
onnx/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
onnx/special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "unk_token": "<|endoftext|>"
5
+ }
onnx/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
onnx/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "50256": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
+ },
14
+ "additional_special_tokens": [],
15
+ "bos_token": "<|endoftext|>",
16
+ "clean_up_tokenization_spaces": true,
17
+ "eos_token": "<|endoftext|>",
18
+ "errors": "replace",
19
+ "model_max_length": 2048,
20
+ "pad_token": null,
21
+ "tokenizer_class": "GPT2Tokenizer",
22
+ "unk_token": "<|endoftext|>"
23
+ }
onnx/vocab.json ADDED
The diff for this file is too large to render. See raw diff