kepinsam commited on
Commit
917a964
1 Parent(s): e190a5c

End of training

Browse files
Files changed (2) hide show
  1. README.md +10 -10
  2. pytorch_model.bin +1 -1
README.md CHANGED
@@ -22,7 +22,7 @@ model-index:
22
  metrics:
23
  - name: Sacrebleu
24
  type: sacrebleu
25
- value: 31.8044
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,9 +32,9 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [facebook/nllb-200-distilled-600M](https://huggingface.co/facebook/nllb-200-distilled-600M) on the nusatranslation_mt dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 1.1474
36
- - Sacrebleu: 31.8044
37
- - Gen Len: 45.2525
38
 
39
  ## Model description
40
 
@@ -55,7 +55,7 @@ More information needed
55
  The following hyperparameters were used during training:
56
  - learning_rate: 5e-05
57
  - train_batch_size: 4
58
- - eval_batch_size: 16
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: linear
@@ -66,11 +66,11 @@ The following hyperparameters were used during training:
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Sacrebleu | Gen Len |
68
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:|
69
- | 1.6014 | 1.0 | 1650 | 1.3826 | 29.3368 | 45.439 |
70
- | 1.2189 | 2.0 | 3300 | 1.2135 | 31.2236 | 45.2075 |
71
- | 1.0274 | 3.0 | 4950 | 1.1556 | 30.8443 | 45.661 |
72
- | 0.8923 | 4.0 | 6600 | 1.1471 | 31.6079 | 45.1555 |
73
- | 0.8094 | 5.0 | 8250 | 1.1474 | 31.8044 | 45.2525 |
74
 
75
 
76
  ### Framework versions
 
22
  metrics:
23
  - name: Sacrebleu
24
  type: sacrebleu
25
+ value: 31.4299
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [facebook/nllb-200-distilled-600M](https://huggingface.co/facebook/nllb-200-distilled-600M) on the nusatranslation_mt dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 1.1479
36
+ - Sacrebleu: 31.4299
37
+ - Gen Len: 45.1465
38
 
39
  ## Model description
40
 
 
55
  The following hyperparameters were used during training:
56
  - learning_rate: 5e-05
57
  - train_batch_size: 4
58
+ - eval_batch_size: 8
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: linear
 
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Sacrebleu | Gen Len |
68
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:|
69
+ | 1.6013 | 1.0 | 1650 | 1.3829 | 29.312 | 45.592 |
70
+ | 1.2324 | 2.0 | 3300 | 1.2091 | 30.6403 | 45.6005 |
71
+ | 1.0217 | 3.0 | 4950 | 1.1552 | 31.428 | 45.0845 |
72
+ | 0.8984 | 4.0 | 6600 | 1.1470 | 31.4125 | 45.365 |
73
+ | 0.8012 | 5.0 | 8250 | 1.1479 | 31.4299 | 45.1465 |
74
 
75
 
76
  ### Framework versions
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cf7a284027ed67db2eaca12c4f2992f1d080edf6a68bfa0ab301a12dd96bd03
3
  size 2460469182
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bea64cc6cc94e689d605700444cd22d5637352763b3f62542330dbb1f7f55916
3
  size 2460469182