system HF staff commited on
Commit
d03834d
1 Parent(s): 266c43a

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +3 -2
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "architectures": [
3
- "PhobertForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
@@ -25,7 +25,7 @@
25
  "length_penalty": 1.0,
26
  "max_length": 20,
27
  "max_position_embeddings": 258,
28
- "model_type": "phobert",
29
  "num_attention_heads": 12,
30
  "num_beams": 1,
31
  "num_hidden_layers": 12,
@@ -41,6 +41,7 @@
41
  "top_k": 50,
42
  "top_p": 1.0,
43
  "torchscript": false,
 
44
  "type_vocab_size": 1,
45
  "use_bfloat16": false,
46
  "vocab_size": 64001
 
1
  {
2
  "architectures": [
3
+ "RobertaForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
 
25
  "length_penalty": 1.0,
26
  "max_length": 20,
27
  "max_position_embeddings": 258,
28
+ "model_type": "roberta",
29
  "num_attention_heads": 12,
30
  "num_beams": 1,
31
  "num_hidden_layers": 12,
 
41
  "top_k": 50,
42
  "top_p": 1.0,
43
  "torchscript": false,
44
+ "tokenizer_class": "PhobertTokenizer",
45
  "type_vocab_size": 1,
46
  "use_bfloat16": false,
47
  "vocab_size": 64001