libra-11b-chat / config.json
yifanxu
model version 1.0
c6b22dd
{
"_name_or_path": "/home/yfxu/libra/CHECKPOINTS/Libra/instruction_1",
"addition_mode": false,
"architectures": [
"LibraForCausalLM"
],
"attn_pdrop": 0.0,
"bos_token_id": 1,
"bridge_rank": 8,
"concat_signals": true,
"contiguous_signal_size": 4096,
"embd_pdrop": 0.0,
"eos_token_id": 2,
"hidden_act": "silu",
"hidden_size": 4096,
"image_feature_resolution": 24,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 2048,
"max_vision_token_length": 578,
"model_type": "libra",
"newline_token_id": 13,
"norm_signals": true,
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"pad_token_id": 0,
"resid_pdrop": 0.0,
"rms_norm_eps": 1e-06,
"rope_theta": 10000.0,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.38.2",
"unified_head": false,
"use_2d_rope": false,
"use_bridge": true,
"use_cache": true,
"use_vision_position_embedding": false,
"vision_codebook_num": 2,
"vision_down_ratio": 4,
"vision_embd_pdrop": 0.0,
"vision_prediction_mode": "1d",
"vision_resid_pdrop": 0.0,
"vision_vocab_size": 514,
"vocab_size": 32000
}