|
{ |
|
"_name_or_path": "/data/ckpt/converted/meter_pretrain_nlvr2_288", |
|
"architectures": [ |
|
"MeterForVisualReasoning" |
|
], |
|
"bos_token_id": 0, |
|
"contrast_method": "none", |
|
"draw_false_text": 0, |
|
"drop_rate": 0.1, |
|
"eos_token_id": 2, |
|
"freeze_patterns": null, |
|
"hidden_size": 768, |
|
"ib_kl": false, |
|
"image_size": 288, |
|
"initializer_range": 0.02, |
|
"itm_loss": 1, |
|
"keep_ratio": 0.5, |
|
"max_text_len": 50, |
|
"mlm_loss": 1, |
|
"mlm_prob": 0.15, |
|
"mlp_ratio": 4, |
|
"model_type": "meter", |
|
"mpp_loss": 0, |
|
"num_heads": 12, |
|
"num_layers": 12, |
|
"num_top_layer": 6, |
|
"pad_token_id": 1, |
|
"prune_layers": null, |
|
"prune_method": "mlp_states", |
|
"token_types": 3, |
|
"tokenizer": "roberta-base", |
|
"torch_dtype": "float32", |
|
"transformers_version": "4.25.1", |
|
"vit_patch_size": 16, |
|
"vocab_size": 50265, |
|
"vqa_label_size": 3129, |
|
"whole_word_masking": false |
|
} |
|
|