tianyisun commited on
Commit
ff923c1
1 Parent(s): 0267c57

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -30
README.md CHANGED
@@ -3,24 +3,9 @@ tags:
3
  - generated_from_trainer
4
  datasets:
5
  - glue
6
- metrics:
7
- - matthews_correlation
8
  model-index:
9
  - name: gpt2-finetuned-cola
10
- results:
11
- - task:
12
- name: Text Classification
13
- type: text-classification
14
- dataset:
15
- name: glue
16
- type: glue
17
- config: cola
18
- split: train
19
- args: cola
20
- metrics:
21
- - name: Matthews Correlation
22
- type: matthews_correlation
23
- value: 0.43512191500404157
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -29,9 +14,6 @@ should probably proofread and complete it, then remove this comment. -->
29
  # gpt2-finetuned-cola
30
 
31
  This model is a fine-tuned version of [microsoft/DialogRPT-updown](https://huggingface.co/microsoft/DialogRPT-updown) on the glue dataset.
32
- It achieves the following results on the evaluation set:
33
- - Loss: 0.9720
34
- - Matthews Correlation: 0.4351
35
 
36
  ## Model description
37
 
@@ -58,17 +40,6 @@ The following hyperparameters were used during training:
58
  - lr_scheduler_type: linear
59
  - num_epochs: 5
60
 
61
- ### Training results
62
-
63
- | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation |
64
- |:-------------:|:-----:|:----:|:---------------:|:--------------------:|
65
- | 0.5599 | 1.0 | 535 | 0.5520 | 0.3379 |
66
- | 0.4011 | 2.0 | 1070 | 0.5644 | 0.4304 |
67
- | 0.284 | 3.0 | 1605 | 0.6289 | 0.4232 |
68
- | 0.2214 | 4.0 | 2140 | 0.8217 | 0.4320 |
69
- | 0.1779 | 5.0 | 2675 | 0.9720 | 0.4351 |
70
-
71
-
72
  ### Framework versions
73
 
74
  - Transformers 4.22.0.dev0
 
3
  - generated_from_trainer
4
  datasets:
5
  - glue
 
 
6
  model-index:
7
  - name: gpt2-finetuned-cola
8
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
14
  # gpt2-finetuned-cola
15
 
16
  This model is a fine-tuned version of [microsoft/DialogRPT-updown](https://huggingface.co/microsoft/DialogRPT-updown) on the glue dataset.
 
 
 
17
 
18
  ## Model description
19
 
 
40
  - lr_scheduler_type: linear
41
  - num_epochs: 5
42
 
 
 
 
 
 
 
 
 
 
 
 
43
  ### Framework versions
44
 
45
  - Transformers 4.22.0.dev0