File size: 1,209 Bytes
cbca73a
b05dfd5
cbca73a
 
b05dfd5
 
cbca73a
229fd89
cbca73a
 
c51c512
cbca73a
 
51aa5f0
229fd89
cbca73a
873c67f
cbca73a
b05dfd5
cbca73a
b05dfd5
51aa5f0
229fd89
cbca73a
df93d80
9bb0068
cbca73a
 
 
 
 
 
 
229fd89
cbca73a
 
 
 
 
b05dfd5
cbca73a
b05dfd5
cbca73a
b05dfd5
 
cbca73a
 
 
c51c512
cbca73a
68d27ac
cbca73a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
{
  "_num_labels": 3,
  "activation_dropout": 0.0,
  "activation_function": "gelu_new",
  "add_bias_logits": false,
  "add_final_layer_norm": false,
  "architectures": [
    "BartForConditionalGeneration"
  ],
  "attention_dropout": 0.0,
  "bos_token_id": 2,
  "classifier_dropout": 0.0,
  "d_model": 768,
  "decoder_attention_heads": 16,
  "decoder_ffn_dim": 4096,
  "decoder_layerdrop": 0.0,
  "decoder_layers": 34,
  "decoder_start_token_id": 2,
  "do_blenderbot_90_layernorm": false,
  "dropout": 0.1,
  "early_stopping": true,
  "encoder_attention_heads": 16,
  "encoder_ffn_dim": 4096,
  "encoder_layerdrop": 0.0,
  "encoder_layers": 3,
  "eos_token_id": 1,
  "forced_eos_token_id": 2,
  "id2label": {
    "0": "LABEL_0",
    "1": "LABEL_1",
    "2": "LABEL_2"
  },
  "init_std": 0.02,
  "is_encoder_decoder": true,
  "label2id": {
    "LABEL_0": 0,
    "LABEL_1": 1,
    "LABEL_2": 2
  },
  "max_length": 256,
  "max_position_embeddings": 1026,
  "min_length": 20,
  "model_type": "bart",
  "no_repeat_ngram_size": 3,
  "num_hidden_layers": 9,
  "pad_token_id": 3,
  "scale_embedding": false,
  "torch_dtype": "float32",
  "transformers_version": "4.26.0",
  "use_cache": true,
  "vocab_size": 30000
}