smitmenon commited on
Commit
35137cc
1 Parent(s): 6fbe757

Added Noise to english source, fine-tuned with denoising, version mbart_endenoised_v2

Browse files
Files changed (3) hide show
  1. README.md +5 -5
  2. generation_config.json +1 -1
  3. tokenizer_config.json +1 -0
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [facebook/mbart-large-50-one-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.4076
19
 
20
  ## Model description
21
 
@@ -46,13 +46,13 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | 2.9524 | 1.0 | 125 | 0.4337 |
50
- | 0.4483 | 2.0 | 250 | 0.4076 |
51
 
52
 
53
  ### Framework versions
54
 
55
- - Transformers 4.46.3
56
  - Pytorch 2.5.1+cu121
57
  - Datasets 3.1.0
58
- - Tokenizers 0.20.3
 
15
 
16
  This model is a fine-tuned version of [facebook/mbart-large-50-one-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.3923
19
 
20
  ## Model description
21
 
 
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
+ | 0.4815 | 1.0 | 625 | 0.3941 |
50
+ | 0.3489 | 2.0 | 1250 | 0.3923 |
51
 
52
 
53
  ### Framework versions
54
 
55
+ - Transformers 4.47.0
56
  - Pytorch 2.5.1+cu121
57
  - Datasets 3.1.0
58
+ - Tokenizers 0.21.0
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  "max_length": 200,
9
  "num_beams": 5,
10
  "pad_token_id": 1,
11
- "transformers_version": "4.46.3"
12
  }
 
8
  "max_length": 200,
9
  "num_beams": 5,
10
  "pad_token_id": 1,
11
+ "transformers_version": "4.47.0"
12
  }
tokenizer_config.json CHANGED
@@ -515,6 +515,7 @@
515
  "clean_up_tokenization_spaces": false,
516
  "cls_token": "<s>",
517
  "eos_token": "</s>",
 
518
  "language_codes": "ML50",
519
  "mask_token": "<mask>",
520
  "model_max_length": 1000000000000000019884624838656,
 
515
  "clean_up_tokenization_spaces": false,
516
  "cls_token": "<s>",
517
  "eos_token": "</s>",
518
+ "extra_special_tokens": {},
519
  "language_codes": "ML50",
520
  "mask_token": "<mask>",
521
  "model_max_length": 1000000000000000019884624838656,