noahkim commited on
Commit
f7cbb70
โ€ข
1 Parent(s): 405de3f

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -20
README.md CHANGED
@@ -1,27 +1,67 @@
1
  ---
2
- language: ko
3
- inference: false
4
- widget:
5
- - text: 'hello this is an example'
6
  tags:
7
- - summarization
8
- - bigbird
9
- - EncoderDecoderModel
10
-
 
 
11
  ---
12
- - This model is a [kobigbird-bert-base](https://huggingface.co/monologg/kobigbird-bert-base) finetuned on the [naver-news-summarization-ko](https://huggingface.co/datasets/daekeun-ml/naver-news-summarization-ko)
13
 
14
- <<20220916 Commit>>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- ๊ธด ๋ฌธ์žฅ์˜ ์š”์•ฝ ๋ชจ๋ธ ํŠนํ™”๋œ ๋ชจ๋ธ์„ ๋งŒ๋“ค๊ธฐ ์œ„ํ•ด KoBigBird ๋ชจ๋ธ์„ Encoder Decoder ํ˜•ํƒœ๋กœ ๋ณ€ํ™˜ํ•œ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
17
- monologg๋‹˜์ด ๋งŒ๋“ค์–ด ๋†“์œผ์‹  KoBigBird ๋ชจ๋ธ์€ bert ๊ธฐ๋ฐ˜์œผ๋กœ ๋งŒ๋“ค์–ด์ ธ ์žˆ๊ธฐ ๋•Œ๋ฌธ์— Q&A ๋“ฑ ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ Task์— ์ข‹์€ ์„ฑ๋Šฅ์„ ๋ณด์ž…๋‹ˆ๋‹ค.
18
- ์ €๋Š” ์š”์•ฝ(Summarization) Task๋ฅผ ์œ„ํ•ด์„œ Encoder-Decoder ํ˜•ํƒœ์˜ ๋ชจ์–‘์œผ๋กœ ๋ฐ”๊ฟจ์Šต๋‹ˆ๋‹ค.
19
- ์ถ”ํ›„ ์ง€์†์ ์ธ ์—…๋ฐ์ดํŠธ๋กœ ๋ณ„๋„์˜ ๋ถˆํŽธํ•จ ์—†์ด ์‚ฌ์šฉํ•  ์ˆ˜ ์žˆ๋„๋ก ๋ฐ”๊พธ๊ฒ ์Šต๋‹ˆ๋‹ค.
20
 
21
- <pre><code>
22
- # Python Code
23
- from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
24
 
25
- tokenizer = AutoTokenizer.from_pretrained("noahkim/kobigbird-finetuned-Encoder-Decoder")
26
- model = AutoModelForSeq2SeqLM.from_pretrained("noahkim/kobigbird-finetuned-Encoder-Decoder")
27
- </pre></code>
 
 
1
  ---
 
 
 
 
2
  tags:
3
+ - generated_from_trainer
4
+ metrics:
5
+ - rouge
6
+ model-index:
7
+ - name: kobigbird-finetuned-Encoder-Decoder
8
+ results: []
9
  ---
 
10
 
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # kobigbird-finetuned-Encoder-Decoder
15
+
16
+ This model was trained from scratch on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 4.4011
19
+ - Rouge1: 12.1436
20
+ - Rouge2: 2.2747
21
+ - Rougel: 11.7428
22
+ - Rougelsum: 11.7408
23
+ - Gen Len: 20.0
24
+
25
+ ## Model description
26
+
27
+ More information needed
28
+
29
+ ## Intended uses & limitations
30
+
31
+ More information needed
32
+
33
+ ## Training and evaluation data
34
+
35
+ More information needed
36
+
37
+ ## Training procedure
38
+
39
+ ### Training hyperparameters
40
+
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 2e-05
43
+ - train_batch_size: 1
44
+ - eval_batch_size: 1
45
+ - seed: 42
46
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
+ - lr_scheduler_type: linear
48
+ - num_epochs: 5
49
+ - mixed_precision_training: Native AMP
50
+
51
+ ### Training results
52
+
53
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
54
+ |:-------------:|:-----:|:------:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:|
55
+ | 5.8126 | 1.0 | 22194 | 5.7258 | 8.857 | 0.2266 | 8.6768 | 8.703 | 20.0 |
56
+ | 5.1675 | 2.0 | 44388 | 5.0540 | 7.0446 | 0.5937 | 6.8255 | 6.8637 | 20.0 |
57
+ | 4.5552 | 3.0 | 66582 | 4.6871 | 10.3238 | 1.1363 | 9.9598 | 9.9394 | 20.0 |
58
+ | 4.3369 | 4.0 | 88776 | 4.4789 | 11.0189 | 1.6226 | 10.7301 | 10.6951 | 20.0 |
59
+ | 4.0251 | 5.0 | 110970 | 4.4011 | 12.1436 | 2.2747 | 11.7428 | 11.7408 | 20.0 |
60
 
 
 
 
 
61
 
62
+ ### Framework versions
 
 
63
 
64
+ - Transformers 4.22.0
65
+ - Pytorch 1.12.1+cu113
66
+ - Datasets 2.4.0
67
+ - Tokenizers 0.12.1