kmfoda commited on
Commit
835c702
1 Parent(s): a0f1bb7

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +4 -3
config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
- "_name_or_path": "google/pegasus-large",
3
  "activation_dropout": 0.1,
4
  "activation_function": "relu",
5
  "add_bias_logits": false,
6
  "add_final_layer_norm": true,
7
  "architectures": [
8
- "PegasusForConditionalGeneration"
9
  ],
10
  "attention_dropout": 0.1,
11
  "bos_token_id": 0,
@@ -117,7 +117,8 @@
117
  "max_position_embeddings": 512
118
  }
119
  },
120
- "transformers_version": "4.11.0.dev0",
 
121
  "use_cache": true,
122
  "vocab_size": 96103
123
  }
1
  {
2
+ "_name_or_path": "../output_dir/",
3
  "activation_dropout": 0.1,
4
  "activation_function": "relu",
5
  "add_bias_logits": false,
6
  "add_final_layer_norm": true,
7
  "architectures": [
8
+ "PegasusModel"
9
  ],
10
  "attention_dropout": 0.1,
11
  "bos_token_id": 0,
117
  "max_position_embeddings": 512
118
  }
119
  },
120
+ "torch_dtype": "float32",
121
+ "transformers_version": "4.16.0.dev0",
122
  "use_cache": true,
123
  "vocab_size": 96103
124
  }