ArthurZ HF staff commited on
Commit
104e274
1 Parent(s): c8485e9
Files changed (2) hide show
  1. config.json +21 -21
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,28 +1,28 @@
1
  {
 
2
  "activation_dropout": 0.0,
3
- "activation_function": "gelu",
4
  "architectures": [
5
  "OPTModel"
6
  ],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  "vocab_size": 50272,
8
- "max_position_embeddings":2048,
9
- "num_hidden_layers":24,
10
- "num_attention_heads":16,
11
- "ffn_dim":4096,
12
- "layerdrop":0.0,
13
- "d_model":1024,
14
- "embed_dim":512,
15
- "dropout":0.1,
16
- "attention_dropout":0.0,
17
- "init_std":0.02,
18
- "scale_embedding":false,
19
- "share_input_output_embed":true,
20
- "use_cache":false,
21
- "pad_token_id":1,
22
- "bos_token_id":0,
23
- "eos_token_id":2,
24
- "decoder_start_token_id":2,
25
- "forced_eos_token_id":2,
26
- "output_projection":true,
27
- "decoder_layernorm":false
28
  }
 
1
  {
2
+ "_name_or_path": "opt-350m",
3
  "activation_dropout": 0.0,
4
+ "activation_function": "relu",
5
  "architectures": [
6
  "OPTModel"
7
  ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 2,
10
+ "do_layer_norm_before": false,
11
+ "dropout": 0.1,
12
+ "eos_token_id": 2,
13
+ "ffn_dim": 4096,
14
+ "hidden_size": 1024,
15
+ "init_std": 0.02,
16
+ "layerdrop": 0.0,
17
+ "max_position_embeddings": 2048,
18
+ "model_type": "opt",
19
+ "num_attention_heads": 16,
20
+ "num_hidden_layers": 24,
21
+ "pad_token_id": 1,
22
+ "prefix": "</s>",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.21.0.dev0",
25
+ "use_cache": true,
26
  "vocab_size": 50272,
27
+ "word_embed_proj_dim": 512
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:866995abded477a0e8c7c05a492d607e97b5f168e02586f2bfbad23e50f763e9
3
- size 1427863863
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c450f836f017007a742704fc9f469acae8d94e2a6e785151016532c4b680f7ef
3
+ size 1324906489