n3wtou commited on
Commit
5374fd8
1 Parent(s): 2ea618f

Training in progress epoch 0

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -12,11 +12,11 @@ probably proofread and complete it, then remove this comment. -->
12
 
13
  # n3wtou/mt5-small-finedtuned-4-swahili
14
 
15
- This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on [csebuetnlp/xlsum](https://huggingface.co/datasets/csebuetnlp/xlsum/viewer/swahili/train?p=1) dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 3.1045
18
- - Validation Loss: 2.5859
19
- - Epoch: 9
20
 
21
  ## Model description
22
 
@@ -35,23 +35,14 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'WarmUp', 'config': {'initial_learning_rate': 5.6e-05, 'decay_schedule_fn': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5.6e-05, 'decay_steps': 15785, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, '__passive_serialization__': True}, 'warmup_steps': 5, 'power': 1.0, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.1}
39
  - training_precision: mixed_float16
40
 
41
  ### Training results
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
- | 6.8543 | 3.1033 | 0 |
46
- | 4.3991 | 2.9232 | 1 |
47
- | 3.9381 | 2.8241 | 2 |
48
- | 3.6672 | 2.7522 | 3 |
49
- | 3.4848 | 2.6935 | 4 |
50
- | 3.3552 | 2.6611 | 5 |
51
- | 3.2517 | 2.6296 | 6 |
52
- | 3.1818 | 2.6104 | 7 |
53
- | 3.1348 | 2.5930 | 8 |
54
- | 3.1045 | 2.5859 | 9 |
55
 
56
 
57
  ### Framework versions
 
12
 
13
  # n3wtou/mt5-small-finedtuned-4-swahili
14
 
15
+ This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 9.3513
18
+ - Validation Loss: 5.1821
19
+ - Epoch: 0
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5.6e-05, 'decay_steps': 987, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
39
  - training_precision: mixed_float16
40
 
41
  ### Training results
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
+ | 9.3513 | 5.1821 | 0 |
 
 
 
 
 
 
 
 
 
46
 
47
 
48
  ### Framework versions
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "eos_token": "</s>",
3
+ "pad_token": "<pad>",
4
+ "unk_token": "<unk>"
5
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:760f268b5d0ac4587ee568d2ed322832e4e6ce7216d0d1dd17caeaf0c808f902
3
  size 2225556280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3a4b7f26fa08ac4cb751ddfa52d4edd0306af8fc8ace26ef9ab33b3bf2d6dd
3
  size 2225556280
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da4980af4e0649bb07a8cffdad7344bba0401a39dc67fb0256b4da603aae65b9
3
+ size 16330466
tokenizer_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": null,
3
+ "clean_up_tokenization_spaces": true,
4
+ "eos_token": "</s>",
5
+ "extra_ids": 0,
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "pad_token": "<pad>",
8
+ "sp_model_kwargs": {},
9
+ "tokenizer_class": "T5Tokenizer",
10
+ "unk_token": "<unk>"
11
+ }