collij22 commited on
Commit
183ce75
1 Parent(s): fe79109

Upload OPTForCausalLM

Browse files
config.json CHANGED
@@ -4,7 +4,7 @@
4
  "activation_dropout": 0.0,
5
  "activation_function": "relu",
6
  "architectures": [
7
- "OPTForSequenceClassification"
8
  ],
9
  "attention_dropout": 0.0,
10
  "bos_token_id": 2,
 
4
  "activation_dropout": 0.0,
5
  "activation_function": "relu",
6
  "architectures": [
7
+ "OPTForCausalLM"
8
  ],
9
  "attention_dropout": 0.0,
10
  "bos_token_id": 2,
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.38.2"
7
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc6c116400530888bf499ac6ad49716b27f9d723faefd12ced0386ebd05d5bd6
3
  size 4988544768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b1fd36712e0bf4b55a3ebe282d0995ae627d33f897f131c8bc1c179b90ddca5
3
  size 4988544768
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:487f528ca8ab122e7acd0e8060461f3d2934481394bb7cda7d2084dbea09bbef
3
- size 314720000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6055fb7460c7ff3ff1ee90b54ffe12b8f46668dc91a5851be72c3800b7d27053
3
+ size 314709672
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5303203840
4
  },
5
  "weight_map": {
6
  "model.decoder.embed_positions.weight": "model-00001-of-00002.safetensors",
@@ -518,7 +518,6 @@
518
  "model.decoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
519
  "model.decoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
520
  "model.decoder.layers.9.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
521
- "model.decoder.layers.9.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
522
- "score.weight": "model-00002-of-00002.safetensors"
523
  }
524
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 5303193600
4
  },
5
  "weight_map": {
6
  "model.decoder.embed_positions.weight": "model-00001-of-00002.safetensors",
 
518
  "model.decoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
519
  "model.decoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
520
  "model.decoder.layers.9.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
521
+ "model.decoder.layers.9.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors"
 
522
  }
523
  }