nomsgadded commited on
Commit
80cb67e
1 Parent(s): d5ad055

Model save

Browse files
Files changed (1) hide show
  1. README.md +15 -13
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- license: apache-2.0
3
- base_model: distilgpt2
4
  tags:
5
  - generated_from_trainer
6
  datasets:
@@ -8,13 +8,13 @@ datasets:
8
  metrics:
9
  - accuracy
10
  model-index:
11
- - name: clm
12
  results:
13
  - task:
14
- name: Causal Language Modeling
15
- type: text-generation
16
  dataset:
17
- name: wikitext wikitext-2-raw-v1
18
  type: wikitext
19
  config: wikitext-2-raw-v1
20
  split: validation
@@ -22,18 +22,18 @@ model-index:
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
- value: 0.37187601824698596
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
  should probably proofread and complete it, then remove this comment. -->
30
 
31
- # clm
32
 
33
- This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the wikitext wikitext-2-raw-v1 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 3.4802
36
- - Accuracy: 0.3719
37
 
38
  ## Model description
39
 
@@ -61,13 +61,15 @@ The following hyperparameters were used during training:
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_ratio: 0.1
64
- - num_epochs: 1.0
65
 
66
  ### Training results
67
 
68
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
69
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
70
- | 3.6158 | 0.99 | 72 | 3.4802 | 0.3719 |
 
 
71
 
72
 
73
  ### Framework versions
 
1
  ---
2
+ license: mit
3
+ base_model: roberta-base
4
  tags:
5
  - generated_from_trainer
6
  datasets:
 
8
  metrics:
9
  - accuracy
10
  model-index:
11
+ - name: mlm
12
  results:
13
  - task:
14
+ name: Masked Language Modeling
15
+ type: fill-mask
16
  dataset:
17
+ name: wikitext
18
  type: wikitext
19
  config: wikitext-2-raw-v1
20
  split: validation
 
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
+ value: 0.7278010101558682
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
  should probably proofread and complete it, then remove this comment. -->
30
 
31
+ # mlm
32
 
33
+ This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the wikitext dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 1.2607
36
+ - Accuracy: 0.7278
37
 
38
  ## Model description
39
 
 
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_ratio: 0.1
64
+ - num_epochs: 3.0
65
 
66
  ### Training results
67
 
68
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
69
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
70
+ | 1.3758 | 1.0 | 150 | 1.2826 | 0.7277 |
71
+ | 1.3763 | 2.0 | 300 | 1.2747 | 0.7272 |
72
+ | 1.3558 | 3.0 | 450 | 1.2607 | 0.7278 |
73
 
74
 
75
  ### Framework versions