Vlasta commited on
Commit
696c6ed
1 Parent(s): 2b17428
Files changed (2) hide show
  1. config.json +12 -18
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,28 +1,22 @@
1
  {
 
2
  "architectures": [
3
- "DebertaForMaskedLM"
4
  ],
5
- "attention_probs_dropout_prob": 0.1,
6
- "hidden_act": "gelu",
7
- "hidden_dropout_prob": 0.1,
8
- "hidden_size": 768,
9
  "initializer_range": 0.02,
10
- "intermediate_size": 3072,
11
- "layer_norm_eps": 1e-07,
12
  "max_position_embeddings": 512,
13
- "max_relative_positions": -1,
14
- "model_type": "deberta",
15
- "num_attention_heads": 12,
16
- "num_hidden_layers": 6,
17
  "pad_token_id": 0,
18
- "pooler_dropout": 0,
19
- "pooler_hidden_act": "gelu",
20
- "pooler_hidden_size": 768,
21
- "pos_att_type": null,
22
- "position_biased_input": true,
23
- "relative_attention": false,
24
  "torch_dtype": "float32",
25
  "transformers_version": "4.19.2",
26
- "type_vocab_size": 0,
27
  "vocab_size": 4101
28
  }
 
1
  {
2
+ "activation": "gelu",
3
  "architectures": [
4
+ "DistilBertForMaskedLM"
5
  ],
6
+ "attention_dropout": 0.1,
7
+ "dim": 768,
8
+ "dropout": 0.1,
9
+ "hidden_dim": 3072,
10
  "initializer_range": 0.02,
 
 
11
  "max_position_embeddings": 512,
12
+ "model_type": "distilbert",
13
+ "n_heads": 12,
14
+ "n_layers": 6,
 
15
  "pad_token_id": 0,
16
+ "qa_dropout": 0.1,
17
+ "seq_classif_dropout": 0.2,
18
+ "sinusoidal_pos_embds": false,
 
 
 
19
  "torch_dtype": "float32",
20
  "transformers_version": "4.19.2",
 
21
  "vocab_size": 4101
22
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10d97ccab29d87b9ce72596d79552c93a0034aa0dae30ff9f79f881e15dabe95
3
- size 186685725
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd01904842e18a5e7a7a54a432190e4995d67b8281cc858409b3765cc79e419
3
+ size 186705839