Waquas commited on
Commit
e5f0131
1 Parent(s): 8ccf809

Training complete

Browse files
Files changed (2) hide show
  1. README.md +11 -10
  2. generation_config.json +1 -1
README.md CHANGED
@@ -2,6 +2,7 @@
2
  license: apache-2.0
3
  base_model: Helsinki-NLP/opus-mt-en-fr
4
  tags:
 
5
  - generated_from_trainer
6
  datasets:
7
  - kde4
@@ -22,7 +23,7 @@ model-index:
22
  metrics:
23
  - name: Bleu
24
  type: bleu
25
- value: 52.88398487672078
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +33,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.8556
36
- - Bleu: 52.8840
37
 
38
  ## Model description
39
 
@@ -53,12 +54,12 @@ More information needed
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 2e-05
56
- - train_batch_size: 32
57
- - eval_batch_size: 64
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
- - num_epochs: 3
62
  - mixed_precision_training: Native AMP
63
 
64
  ### Training results
@@ -67,7 +68,7 @@ The following hyperparameters were used during training:
67
 
68
  ### Framework versions
69
 
70
- - Transformers 4.35.2
71
- - Pytorch 2.1.0+cu121
72
- - Datasets 2.17.0
73
- - Tokenizers 0.15.2
 
2
  license: apache-2.0
3
  base_model: Helsinki-NLP/opus-mt-en-fr
4
  tags:
5
+ - waquas/lab1_finetuning
6
  - generated_from_trainer
7
  datasets:
8
  - kde4
 
23
  metrics:
24
  - name: Bleu
25
  type: bleu
26
+ value: 49.75159108718876
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
33
 
34
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 0.9966
37
+ - Bleu: 49.7516
38
 
39
  ## Model description
40
 
 
54
 
55
  The following hyperparameters were used during training:
56
  - learning_rate: 2e-05
57
+ - train_batch_size: 64
58
+ - eval_batch_size: 128
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: linear
62
+ - num_epochs: 1
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
 
68
 
69
  ### Framework versions
70
 
71
+ - Transformers 4.37.0
72
+ - Pytorch 2.1.2
73
+ - Datasets 2.1.0
74
+ - Tokenizers 0.15.1
generation_config.json CHANGED
@@ -12,5 +12,5 @@
12
  "num_beams": 4,
13
  "pad_token_id": 59513,
14
  "renormalize_logits": true,
15
- "transformers_version": "4.35.2"
16
  }
 
12
  "num_beams": 4,
13
  "pad_token_id": 59513,
14
  "renormalize_logits": true,
15
+ "transformers_version": "4.37.0"
16
  }