File size: 1,209 Bytes
cbca73a b05dfd5 cbca73a b05dfd5 cbca73a 229fd89 cbca73a c51c512 cbca73a 51aa5f0 229fd89 cbca73a 873c67f cbca73a b05dfd5 cbca73a b05dfd5 51aa5f0 229fd89 cbca73a df93d80 9bb0068 cbca73a 229fd89 cbca73a b05dfd5 cbca73a b05dfd5 cbca73a b05dfd5 cbca73a c51c512 cbca73a 68d27ac cbca73a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 |
{
"_num_labels": 3,
"activation_dropout": 0.0,
"activation_function": "gelu_new",
"add_bias_logits": false,
"add_final_layer_norm": false,
"architectures": [
"BartForConditionalGeneration"
],
"attention_dropout": 0.0,
"bos_token_id": 2,
"classifier_dropout": 0.0,
"d_model": 768,
"decoder_attention_heads": 16,
"decoder_ffn_dim": 4096,
"decoder_layerdrop": 0.0,
"decoder_layers": 34,
"decoder_start_token_id": 2,
"do_blenderbot_90_layernorm": false,
"dropout": 0.1,
"early_stopping": true,
"encoder_attention_heads": 16,
"encoder_ffn_dim": 4096,
"encoder_layerdrop": 0.0,
"encoder_layers": 3,
"eos_token_id": 1,
"forced_eos_token_id": 2,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1",
"2": "LABEL_2"
},
"init_std": 0.02,
"is_encoder_decoder": true,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1,
"LABEL_2": 2
},
"max_length": 256,
"max_position_embeddings": 1026,
"min_length": 20,
"model_type": "bart",
"no_repeat_ngram_size": 3,
"num_hidden_layers": 9,
"pad_token_id": 3,
"scale_embedding": false,
"torch_dtype": "float32",
"transformers_version": "4.26.0",
"use_cache": true,
"vocab_size": 30000
}
|