louis030195 commited on
Commit
7f9c9d2
1 Parent(s): 732dafc

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -19
README.md CHANGED
@@ -1,34 +1,20 @@
1
  ---
2
- language:
3
- - en
4
  tags:
5
- - text-generation
6
  - generated_from_trainer
7
- datasets:
8
- - Langame/starter
9
- license: apache-2.0
10
  model-index:
11
  - name: distilgpt2-starter
12
  results: []
13
- widget:
14
- - text: "ice breaker,travel ###"
15
- example_title: "Ice breaker"
16
- - text: "philosophy,science,physic,mathematic,biology ###"
17
- example_title: "Scientific"
18
- - text: "big talk,deep talk ###"
19
- example_title: "Big talk"
20
  ---
21
- # Generate conversation starters, from [Langame](https://langa.me) :p
22
-
23
 
24
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
25
  should probably proofread and complete it, then remove this comment. -->
26
 
27
  # distilgpt2-starter
28
 
29
- This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the Langame/starter dataset.
30
  It achieves the following results on the evaluation set:
31
- - Loss: 3.3145
32
 
33
  ## Model description
34
 
@@ -56,15 +42,24 @@ The following hyperparameters were used during training:
56
  - total_train_batch_size: 8
57
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
58
  - lr_scheduler_type: linear
59
- - num_epochs: 50.0
60
 
61
  ### Training results
62
 
 
 
 
 
 
 
 
 
 
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.16.0.dev0
68
  - Pytorch 1.10.0+cu111
69
  - Datasets 1.18.1
70
  - Tokenizers 0.11.0
 
1
  ---
2
+ license: apache-2.0
 
3
  tags:
 
4
  - generated_from_trainer
 
 
 
5
  model-index:
6
  - name: distilgpt2-starter
7
  results: []
 
 
 
 
 
 
 
8
  ---
 
 
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
  # distilgpt2-starter
14
 
15
+ This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 5.9961
18
 
19
  ## Model description
20
 
 
42
  - total_train_batch_size: 8
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
+ - num_epochs: 500.0
46
 
47
  ### Training results
48
 
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:------:|:----:|:---------------:|
51
+ | No log | 66.67 | 200 | 3.6445 |
52
+ | No log | 133.33 | 400 | 4.5703 |
53
+ | 1.0101 | 200.0 | 600 | 5.2109 |
54
+ | 1.0101 | 266.67 | 800 | 5.5430 |
55
+ | 0.0681 | 333.33 | 1000 | 5.7227 |
56
+ | 0.0681 | 400.0 | 1200 | 5.8672 |
57
+ | 0.0681 | 466.67 | 1400 | 5.9961 |
58
 
59
 
60
  ### Framework versions
61
 
62
+ - Transformers 4.17.0.dev0
63
  - Pytorch 1.10.0+cu111
64
  - Datasets 1.18.1
65
  - Tokenizers 0.11.0