bart_brio_v5 / config.txt
ntmkhanh's picture
Upload 22 files
747a654
raw
history blame contribute delete
No virus
3.75 kB
/content/drive/MyDrive/LuanVan/Bart-BRIO/brio_project-main/BRIO/main.py
Namespace(cuda=True, gpuid=[0], evaluate=False, do_reranking=False, do_generation=False, log=True, port=12355, model_pt='', config='', batch_size=1, epoch=1, report_freq=100, accumulate_step=8, margin=0.001, gold_margin=0, gold_weight=0, mle_weight=0.1, rank_weight=10, model_type='vinai/bartpho-word-base', warmup_steps=10000, normalize=True, grad_norm=0, seed=970903, no_gold=False, pretrained='./finetuned_model_v3/eval_bartpho_final', max_lr=0.002, scale=0.5, score_mode='log', datatype='diverse', dataset='cooking_bart', max_len=120, max_num=6, smooth=0.01, total_len=1024, length_penalty=2.0, do_sample=True, gen_max_len=140, gen_min_len=55, is_pegasus=False, adding=0, eval_interval=1000, num_beams=6)
BRIO(
(model): MBartScorer(
(model): CustomMBartModel(
(shared): Embedding(64001, 768, padding_idx=1)
(encoder): MBartEncoder(
(embed_tokens): Embedding(64001, 768, padding_idx=1)
(embed_positions): MBartLearnedPositionalEmbedding(1026, 768)
(layers): ModuleList(
(0-5): 6 x MBartEncoderLayer(
(self_attn): MBartAttention(
(k_proj): Linear(in_features=768, out_features=768, bias=True)
(v_proj): Linear(in_features=768, out_features=768, bias=True)
(q_proj): Linear(in_features=768, out_features=768, bias=True)
(out_proj): Linear(in_features=768, out_features=768, bias=True)
)
(self_attn_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
(activation_fn): GELUActivation()
(fc1): Linear(in_features=768, out_features=3072, bias=True)
(fc2): Linear(in_features=3072, out_features=768, bias=True)
(final_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
)
)
(layernorm_embedding): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
(layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
)
(decoder): MBartDecoder(
(embed_tokens): Embedding(64001, 768, padding_idx=1)
(embed_positions): MBartLearnedPositionalEmbedding(1026, 768)
(layers): ModuleList(
(0-5): 6 x MBartDecoderLayer(
(self_attn): MBartAttention(
(k_proj): Linear(in_features=768, out_features=768, bias=True)
(v_proj): Linear(in_features=768, out_features=768, bias=True)
(q_proj): Linear(in_features=768, out_features=768, bias=True)
(out_proj): Linear(in_features=768, out_features=768, bias=True)
)
(activation_fn): GELUActivation()
(self_attn_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
(encoder_attn): MBartAttention(
(k_proj): Linear(in_features=768, out_features=768, bias=True)
(v_proj): Linear(in_features=768, out_features=768, bias=True)
(q_proj): Linear(in_features=768, out_features=768, bias=True)
(out_proj): Linear(in_features=768, out_features=768, bias=True)
)
(encoder_attn_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
(fc1): Linear(in_features=768, out_features=3072, bias=True)
(fc2): Linear(in_features=3072, out_features=768, bias=True)
(final_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
)
)
(layernorm_embedding): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
(layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
)
)
(lm_head): Linear(in_features=768, out_features=64001, bias=False)
)
)