khrousdevil commited on
Commit
5836c62
1 Parent(s): 1d6a4b3

Training completed!

Browse files
Files changed (4) hide show
  1. README.md +81 -0
  2. model.safetensors +1 -1
  3. trainer_state.json +18 -18
  4. training_args.bin +1 -1
README.md ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: distilbert-base-uncased
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - emotion
8
+ metrics:
9
+ - accuracy
10
+ - f1
11
+ model-index:
12
+ - name: distilbert-base-uncased-finetuned-emotion
13
+ results:
14
+ - task:
15
+ name: Text Classification
16
+ type: text-classification
17
+ dataset:
18
+ name: emotion
19
+ type: emotion
20
+ config: split
21
+ split: validation
22
+ args: split
23
+ metrics:
24
+ - name: Accuracy
25
+ type: accuracy
26
+ value: 0.922
27
+ - name: F1
28
+ type: f1
29
+ value: 0.9219196681926928
30
+ ---
31
+
32
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
+ should probably proofread and complete it, then remove this comment. -->
34
+
35
+ # distilbert-base-uncased-finetuned-emotion
36
+
37
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
38
+ It achieves the following results on the evaluation set:
39
+ - Loss: 0.2162
40
+ - Accuracy: 0.922
41
+ - F1: 0.9219
42
+
43
+ ## Model description
44
+
45
+ More information needed
46
+
47
+ ## Intended uses & limitations
48
+
49
+ More information needed
50
+
51
+ ## Training and evaluation data
52
+
53
+ More information needed
54
+
55
+ ## Training procedure
56
+
57
+ ### Training hyperparameters
58
+
59
+ The following hyperparameters were used during training:
60
+ - learning_rate: 2e-05
61
+ - train_batch_size: 64
62
+ - eval_batch_size: 64
63
+ - seed: 42
64
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
+ - lr_scheduler_type: linear
66
+ - num_epochs: 2
67
+
68
+ ### Training results
69
+
70
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
71
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
72
+ | 0.8226 | 1.0 | 250 | 0.3270 | 0.9035 | 0.9024 |
73
+ | 0.2533 | 2.0 | 500 | 0.2162 | 0.922 | 0.9219 |
74
+
75
+
76
+ ### Framework versions
77
+
78
+ - Transformers 4.37.2
79
+ - Pytorch 2.2.1+cu121
80
+ - Datasets 2.17.0
81
+ - Tokenizers 0.15.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f69bb94f5679e22f8e88404f97a740e2ce37469d1d53c77b5f363f747f022b6a
3
  size 267844872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abda52c12383258863e78b0e5a36987639ae2e055704fd6abc9b9cc6963efa43
3
  size 267844872
trainer_state.json CHANGED
@@ -11,43 +11,43 @@
11
  {
12
  "epoch": 1.0,
13
  "learning_rate": 1e-05,
14
- "loss": 0.8202,
15
  "step": 250
16
  },
17
  {
18
  "epoch": 1.0,
19
- "eval_accuracy": 0.901,
20
- "eval_f1": 0.9000590964742174,
21
- "eval_loss": 0.3247230350971222,
22
- "eval_runtime": 4.5139,
23
- "eval_samples_per_second": 443.079,
24
- "eval_steps_per_second": 7.089,
25
  "step": 250
26
  },
27
  {
28
  "epoch": 2.0,
29
  "learning_rate": 0.0,
30
- "loss": 0.2547,
31
  "step": 500
32
  },
33
  {
34
  "epoch": 2.0,
35
- "eval_accuracy": 0.9205,
36
- "eval_f1": 0.9204480932224948,
37
- "eval_loss": 0.21922191977500916,
38
- "eval_runtime": 4.5051,
39
- "eval_samples_per_second": 443.938,
40
- "eval_steps_per_second": 7.103,
41
  "step": 500
42
  },
43
  {
44
  "epoch": 2.0,
45
  "step": 500,
46
  "total_flos": 720342861696000.0,
47
- "train_loss": 0.5374606018066406,
48
- "train_runtime": 260.6287,
49
- "train_samples_per_second": 122.78,
50
- "train_steps_per_second": 1.918
51
  }
52
  ],
53
  "logging_steps": 250,
 
11
  {
12
  "epoch": 1.0,
13
  "learning_rate": 1e-05,
14
+ "loss": 0.8226,
15
  "step": 250
16
  },
17
  {
18
  "epoch": 1.0,
19
+ "eval_accuracy": 0.9035,
20
+ "eval_f1": 0.9023761650285413,
21
+ "eval_loss": 0.326994925737381,
22
+ "eval_runtime": 4.5192,
23
+ "eval_samples_per_second": 442.555,
24
+ "eval_steps_per_second": 7.081,
25
  "step": 250
26
  },
27
  {
28
  "epoch": 2.0,
29
  "learning_rate": 0.0,
30
+ "loss": 0.2533,
31
  "step": 500
32
  },
33
  {
34
  "epoch": 2.0,
35
+ "eval_accuracy": 0.922,
36
+ "eval_f1": 0.9219196681926928,
37
+ "eval_loss": 0.21616625785827637,
38
+ "eval_runtime": 4.4445,
39
+ "eval_samples_per_second": 449.996,
40
+ "eval_steps_per_second": 7.2,
41
  "step": 500
42
  },
43
  {
44
  "epoch": 2.0,
45
  "step": 500,
46
  "total_flos": 720342861696000.0,
47
+ "train_loss": 0.5379159927368165,
48
+ "train_runtime": 263.9893,
49
+ "train_samples_per_second": 121.217,
50
+ "train_steps_per_second": 1.894
51
  }
52
  ],
53
  "logging_steps": 250,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:98ab995a6f6b3efbd34a24c2da3e03ebd6d386366b4c58a3f9f5eac9ece73f5a
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91a91b9edad39c6efaf98433b769dcfef6b569ac32ebfca2ad691c24a3333fb9
3
  size 4728