Edit model card

mbart-en-id-smaller-indo-amr-parsing-translated-nafkhan

This model was trained from scratch on the data dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1691
  • Smatch: 0.8299
  • Gen Len: 27.5333

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-06
  • train_batch_size: 5
  • eval_batch_size: 5
  • seed: 42
  • gradient_accumulation_steps: 5
  • total_train_batch_size: 25
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: polynomial
  • lr_scheduler_warmup_steps: 200
  • num_epochs: 640.0
  • label_smoothing_factor: 0.1

Training results

Training Loss Epoch Step Validation Loss Smatch Gen Len
0.9755 39.7351 3600 1.1801 0.7484 27.3333
0.9556 79.4702 7200 1.1545 0.7573 27.0667
0.8935 119.2053 10800 1.1454 0.7924 26.6667
0.8823 158.9404 14400 1.1351 0.7891 27.3333
0.8514 198.6755 18000 1.1308 0.828 26.6
0.8385 238.4106 21600 1.1385 0.8235 27.1
0.836 278.1457 25200 1.1407 0.8399 27.5
0.8503 317.8808 28800 1.1352 0.822 26.6
0.8349 357.6159 32400 1.1478 0.8326 27.2333
0.8296 397.3510 36000 1.1506 0.8201 27.3
0.871 437.0861 39600 1.1592 0.7992 28.3
0.834 476.8212 43200 1.1534 0.825 27.5667
0.8332 516.5563 46800 1.1615 0.8225 27.4
0.8513 556.2914 50400 1.1663 0.8417 27.3667
0.8249 596.0265 54000 1.1667 0.8358 27.4
0.8439 635.7616 57600 1.1691 0.8299 27.5333

Framework versions

  • Transformers 4.44.0
  • Pytorch 2.3.1+cu121
  • Datasets 2.20.0
  • Tokenizers 0.19.1
Downloads last month
25
Safetensors
Model size
394M params
Tensor type
F32
·
Inference API
Unable to determine this model's library. Check the docs .