ubaada commited on
Commit
17508c9
1 Parent(s): 04d0ff3

End of training

Browse files
README.md CHANGED
@@ -1,37 +1,66 @@
1
  ---
 
2
  tags:
3
- - summarization
4
- - summary
5
- - booksum
6
- - long-document
7
- - long-form
8
- datasets:
9
- - ubaada/booksum-complete-cleaned
10
- language:
11
- - en
12
- pipeline_tag: summarization
13
  metrics:
14
  - rouge
15
  model-index:
16
- - name: ubaada/lsg-bart-large-4096-booksum
17
- results:
18
- - task:
19
- type: summarization
20
- name: Summarization
21
- dataset:
22
- name: BookSum
23
- type: ubaada/booksum-complete-cleaned
24
- config: ubaada--booksum
25
- split: test
26
- metrics:
27
- - type: rouge
28
- value: 35.116342
29
- name: ROUGE-1
30
- verified: false
31
- - type: rouge
32
- value: 8.574494
33
- name: ROUGE-2
34
- verified: false
35
  ---
36
 
37
- Trained for 1 epoch on ubaada/booksum-complete-cleaned.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: ubaada/lsg-bart-large-4096-booksum
3
  tags:
4
+ - generated_from_trainer
 
 
 
 
 
 
 
 
 
5
  metrics:
6
  - rouge
7
  model-index:
8
+ - name: lsg-bart-large-4096-booksum
9
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # lsg-bart-large-4096-booksum
16
+
17
+ This model is a fine-tuned version of [ubaada/lsg-bart-large-4096-booksum](https://huggingface.co/ubaada/lsg-bart-large-4096-booksum) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 2.0742
20
+ - Rouge1: 0.4145
21
+ - Rouge2: 0.0797
22
+ - Rougel: 0.1541
23
+
24
+ ## Model description
25
+
26
+ More information needed
27
+
28
+ ## Intended uses & limitations
29
+
30
+ More information needed
31
+
32
+ ## Training and evaluation data
33
+
34
+ More information needed
35
+
36
+ ## Training procedure
37
+
38
+ ### Training hyperparameters
39
+
40
+ The following hyperparameters were used during training:
41
+ - learning_rate: 8e-05
42
+ - train_batch_size: 8
43
+ - eval_batch_size: 1
44
+ - seed: 42
45
+ - gradient_accumulation_steps: 4
46
+ - total_train_batch_size: 32
47
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
+ - lr_scheduler_type: linear
49
+ - num_epochs: 3
50
+ - mixed_precision_training: Native AMP
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel |
55
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|
56
+ | 1.3801 | 1.0 | 1251 | 2.0441 | 0.4223 | 0.0811 | 0.1532 |
57
+ | 1.2385 | 2.0 | 2502 | 2.0753 | 0.3995 | 0.0751 | 0.1512 |
58
+ | 0.9542 | 3.0 | 3753 | 2.0742 | 0.4145 | 0.0797 | 0.1541 |
59
+
60
+
61
+ ### Framework versions
62
+
63
+ - Transformers 4.40.2
64
+ - Pytorch 2.2.0
65
+ - Datasets 2.19.1
66
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -87,7 +87,7 @@
87
  }
88
  },
89
  "torch_dtype": "float32",
90
- "transformers_version": "4.19.2",
91
  "use_cache": true,
92
  "vocab_size": 50265
93
  }
 
87
  }
88
  },
89
  "torch_dtype": "float32",
90
+ "transformers_version": "4.40.2",
91
  "use_cache": true,
92
  "vocab_size": 50265
93
  }
generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 0,
3
+ "decoder_start_token_id": 2,
4
+ "eos_token_id": 2,
5
+ "forced_bos_token_id": 0,
6
+ "forced_eos_token_id": 2,
7
+ "max_length": 2048,
8
+ "no_repeat_ngram_size": 3,
9
+ "num_beams": 5,
10
+ "pad_token_id": 1,
11
+ "repetition_penalty": 2.0,
12
+ "transformers_version": "4.40.2"
13
+ }
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ t
3
  Ġ a
4
  h e
 
1
+ #version: 0.2
2
  Ġ t
3
  Ġ a
4
  h e
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b71c0580da2a3fb1df066ddda88b99824ba70dadd4fbe0ff2e0ae03406b9f04c
3
- size 1650592852
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9f6853726455387d47ecbc7f94866d2dcca5695e87fb950c563e151e1aba420
3
+ size 1652690124
tokenizer_config.json CHANGED
@@ -48,10 +48,17 @@
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
 
51
  "model_max_length": 4096,
 
52
  "pad_token": "<pad>",
 
 
53
  "sep_token": "</s>",
 
54
  "tokenizer_class": "BartTokenizer",
55
  "trim_offsets": true,
 
 
56
  "unk_token": "<unk>"
57
  }
 
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
+ "max_length": 1024,
52
  "model_max_length": 4096,
53
+ "pad_to_multiple_of": null,
54
  "pad_token": "<pad>",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
  "sep_token": "</s>",
58
+ "stride": 0,
59
  "tokenizer_class": "BartTokenizer",
60
  "trim_offsets": true,
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
  "unk_token": "<unk>"
64
  }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed859e39a3a215c49bc9630595a01eb1a9472169f53e6b0e128aceb793c9d52c
3
+ size 6712