adamjweintraut commited on
Commit
9787ea3
1 Parent(s): 4cdbdbd

bart-finetuned-loaf-512-lyrictoplan

Browse files
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [facebook/bart-large](https://huggingface.co/facebook/bart-large) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.8160
19
 
20
  ## Model description
21
 
@@ -40,25 +40,24 @@ The following hyperparameters were used during training:
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
- - num_epochs: 4
 
44
 
45
  ### Training results
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | 1.4176 | 0.45 | 500 | 1.1073 |
50
- | 1.1228 | 0.89 | 1000 | 0.9915 |
51
- | 0.9568 | 1.34 | 1500 | 0.9347 |
52
- | 0.888 | 1.78 | 2000 | 0.8795 |
53
- | 0.8179 | 2.23 | 2500 | 0.8774 |
54
- | 0.7929 | 2.67 | 3000 | 0.8551 |
55
- | 0.7256 | 3.12 | 3500 | 0.8213 |
56
- | 0.6979 | 3.56 | 4000 | 0.8160 |
57
 
58
 
59
  ### Framework versions
60
 
61
- - Transformers 4.39.3
62
  - Pytorch 2.2.1+cu121
63
  - Datasets 2.18.0
64
  - Tokenizers 0.15.2
 
15
 
16
  This model is a fine-tuned version of [facebook/bart-large](https://huggingface.co/facebook/bart-large) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.4458
19
 
20
  ## Model description
21
 
 
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
+ - num_epochs: 3
44
+ - mixed_precision_training: Native AMP
45
 
46
  ### Training results
47
 
48
  | Training Loss | Epoch | Step | Validation Loss |
49
  |:-------------:|:-----:|:----:|:---------------:|
50
+ | 1.0218 | 0.45 | 500 | 0.5545 |
51
+ | 0.5582 | 0.89 | 1000 | 0.5083 |
52
+ | 0.4996 | 1.34 | 1500 | 0.4844 |
53
+ | 0.4744 | 1.78 | 2000 | 0.4609 |
54
+ | 0.449 | 2.23 | 2500 | 0.4560 |
55
+ | 0.4287 | 2.67 | 3000 | 0.4458 |
 
 
56
 
57
 
58
  ### Framework versions
59
 
60
+ - Transformers 4.38.2
61
  - Pytorch 2.2.1+cu121
62
  - Datasets 2.18.0
63
  - Tokenizers 0.15.2
generation_config.json CHANGED
@@ -1,13 +1,25 @@
1
  {
2
- "_from_model_config": true,
3
  "bos_token_id": 0,
 
4
  "decoder_start_token_id": 2,
 
5
  "early_stopping": true,
6
  "eos_token_id": 2,
7
  "forced_bos_token_id": 0,
8
  "forced_eos_token_id": 2,
 
 
 
9
  "no_repeat_ngram_size": 3,
10
  "num_beams": 4,
11
  "pad_token_id": 1,
12
- "transformers_version": "4.39.3"
 
 
 
 
 
 
 
 
13
  }
 
1
  {
 
2
  "bos_token_id": 0,
3
+ "clean_up_tokenization_spaces": true,
4
  "decoder_start_token_id": 2,
5
+ "do_sample": true,
6
  "early_stopping": true,
7
  "eos_token_id": 2,
8
  "forced_bos_token_id": 0,
9
  "forced_eos_token_id": 2,
10
+ "max_new_tokens": 512,
11
+ "min_new_tokens": 128,
12
+ "n_examples": null,
13
  "no_repeat_ngram_size": 3,
14
  "num_beams": 4,
15
  "pad_token_id": 1,
16
+ "padding": "max_length",
17
+ "renormalize_logits": true,
18
+ "repetition_penalty": 0.6,
19
+ "skip_special_tokens": true,
20
+ "temperature": 0.85,
21
+ "top_k": 0,
22
+ "top_p": 0.9,
23
+ "transformers_version": "4.38.2",
24
+ "truncation": true
25
  }
runs/Apr09_04-20-42_870b8dce1033/events.out.tfevents.1712636443.870b8dce1033.342.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21300487002646825696ff09f4e5df7e9bb1f783d0edb5d5dfa9429694293e0e
3
- size 9033
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:501c2bfdf543df9dbdb9cf1578e328c24a523ad8d320ceb5d758a5f80b528617
3
+ size 9387