Uzair934 commited on
Commit
916d37d
1 Parent(s): 7f60191

End of training

Browse files
README.md CHANGED
@@ -1,9 +1,11 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: facebook/bart-base
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: cleaned_ds
9
  results: []
@@ -14,7 +16,14 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # cleaned_ds
16
 
17
- This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset.
 
 
 
 
 
 
 
18
 
19
  ## Model description
20
 
@@ -39,11 +48,21 @@ The following hyperparameters were used during training:
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
- - num_epochs: 1
43
  - mixed_precision_training: Native AMP
44
 
 
 
 
 
 
 
 
 
 
45
  ### Framework versions
46
 
47
  - Transformers 4.44.2
48
  - Pytorch 2.4.1+cu121
 
49
  - Tokenizers 0.19.1
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: sshleifer/distilbart-cnn-12-6
5
  tags:
6
  - generated_from_trainer
7
+ metrics:
8
+ - rouge
9
  model-index:
10
  - name: cleaned_ds
11
  results: []
 
16
 
17
  # cleaned_ds
18
 
19
+ This model is a fine-tuned version of [sshleifer/distilbart-cnn-12-6](https://huggingface.co/sshleifer/distilbart-cnn-12-6) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 4.2803
22
+ - Rouge1: 0.2705
23
+ - Rouge2: 0.0363
24
+ - Rougel: 0.1609
25
+ - Rougelsum: 0.1609
26
+ - Generated Length: 113.0
27
 
28
  ## Model description
29
 
 
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
+ - num_epochs: 3
52
  - mixed_precision_training: Native AMP
53
 
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Generated Length |
57
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:----------------:|
58
+ | No log | 1.0 | 1 | 4.5060 | 0.2826 | 0.0384 | 0.1694 | 0.1694 | 95.0 |
59
+ | No log | 2.0 | 2 | 4.3368 | 0.2832 | 0.0333 | 0.1701 | 0.1701 | 82.0 |
60
+ | No log | 3.0 | 3 | 4.2803 | 0.2705 | 0.0363 | 0.1609 | 0.1609 | 113.0 |
61
+
62
+
63
  ### Framework versions
64
 
65
  - Transformers 4.44.2
66
  - Pytorch 2.4.1+cu121
67
+ - Datasets 3.0.1
68
  - Tokenizers 0.19.1
generation_config.json CHANGED
@@ -5,6 +5,9 @@
5
  "eos_token_id": 2,
6
  "forced_bos_token_id": 0,
7
  "forced_eos_token_id": 2,
 
 
 
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 1,
 
5
  "eos_token_id": 2,
6
  "forced_bos_token_id": 0,
7
  "forced_eos_token_id": 2,
8
+ "length_penalty": 2.0,
9
+ "max_length": 142,
10
+ "min_length": 56,
11
  "no_repeat_ngram_size": 3,
12
  "num_beams": 4,
13
  "pad_token_id": 1,
runs/Oct03_02-19-24_4affa7904f21/events.out.tfevents.1727921966.4affa7904f21.169.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4f9a4c8baf36aa4ece9a2325a12615be74194ad91db626454f06f1f7cc87995
3
- size 7292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec8db7e358f80b808245106842b75a20b2319b96b3821e03f73271c3e4dcc3ae
3
+ size 8164