SimoneJLaudani commited on
Commit
a8c1494
1 Parent(s): 920d44d

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: distilbert-base-cased
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -18,13 +18,13 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # trainerH
20
 
21
- This model is a fine-tuned version of [distilbert-base-cased](https://huggingface.co/distilbert-base-cased) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.9755
24
- - Precision: 0.8239
25
- - Recall: 0.8179
26
- - F1: 0.8178
27
- - Accuracy: 0.8179
28
 
29
  ## Model description
30
 
@@ -55,48 +55,48 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
- | 1.9268 | 0.14 | 30 | 1.8275 | 0.1219 | 0.2549 | 0.1507 | 0.2549 |
59
- | 1.7885 | 0.27 | 60 | 1.5537 | 0.5383 | 0.4678 | 0.4354 | 0.4678 |
60
- | 1.3895 | 0.41 | 90 | 1.2517 | 0.6341 | 0.5770 | 0.5469 | 0.5770 |
61
- | 1.1972 | 0.54 | 120 | 1.0482 | 0.7148 | 0.7115 | 0.7048 | 0.7115 |
62
- | 1.0679 | 0.68 | 150 | 0.8604 | 0.7629 | 0.7507 | 0.7449 | 0.7507 |
63
- | 0.9503 | 0.81 | 180 | 0.7366 | 0.7749 | 0.7619 | 0.7585 | 0.7619 |
64
- | 0.7036 | 0.95 | 210 | 0.7657 | 0.7898 | 0.7479 | 0.7457 | 0.7479 |
65
- | 0.5612 | 1.08 | 240 | 0.7082 | 0.7806 | 0.7675 | 0.7665 | 0.7675 |
66
- | 0.5023 | 1.22 | 270 | 0.6190 | 0.8138 | 0.8039 | 0.8042 | 0.8039 |
67
- | 0.3955 | 1.35 | 300 | 0.6376 | 0.8129 | 0.7955 | 0.7977 | 0.7955 |
68
- | 0.3902 | 1.49 | 330 | 0.6429 | 0.7922 | 0.7815 | 0.7820 | 0.7815 |
69
- | 0.4563 | 1.62 | 360 | 0.7129 | 0.8281 | 0.8067 | 0.8064 | 0.8067 |
70
- | 0.396 | 1.76 | 390 | 0.6885 | 0.8137 | 0.8067 | 0.8049 | 0.8067 |
71
- | 0.2764 | 1.89 | 420 | 0.6671 | 0.7962 | 0.7871 | 0.7867 | 0.7871 |
72
- | 0.298 | 2.03 | 450 | 0.6989 | 0.8315 | 0.8235 | 0.8239 | 0.8235 |
73
- | 0.148 | 2.16 | 480 | 0.7212 | 0.8242 | 0.8207 | 0.8196 | 0.8207 |
74
- | 0.1483 | 2.3 | 510 | 0.8080 | 0.8060 | 0.8011 | 0.8002 | 0.8011 |
75
- | 0.1393 | 2.43 | 540 | 0.8290 | 0.8143 | 0.8039 | 0.8039 | 0.8039 |
76
- | 0.2032 | 2.57 | 570 | 0.7937 | 0.8179 | 0.8123 | 0.8127 | 0.8123 |
77
- | 0.2178 | 2.7 | 600 | 0.7673 | 0.8284 | 0.8235 | 0.8236 | 0.8235 |
78
- | 0.0704 | 2.84 | 630 | 0.7686 | 0.8302 | 0.8263 | 0.8267 | 0.8263 |
79
- | 0.0938 | 2.97 | 660 | 0.8731 | 0.8152 | 0.8095 | 0.8087 | 0.8095 |
80
- | 0.1216 | 3.11 | 690 | 0.8420 | 0.8268 | 0.8235 | 0.8233 | 0.8235 |
81
- | 0.0542 | 3.24 | 720 | 0.8717 | 0.8187 | 0.8123 | 0.8124 | 0.8123 |
82
- | 0.017 | 3.38 | 750 | 0.8219 | 0.8305 | 0.8263 | 0.8269 | 0.8263 |
83
- | 0.0117 | 3.51 | 780 | 0.9020 | 0.8119 | 0.8067 | 0.8062 | 0.8067 |
84
- | 0.0381 | 3.65 | 810 | 0.8809 | 0.8412 | 0.8347 | 0.8350 | 0.8347 |
85
- | 0.0348 | 3.78 | 840 | 0.9674 | 0.8273 | 0.8179 | 0.8185 | 0.8179 |
86
- | 0.0338 | 3.92 | 870 | 0.9815 | 0.8274 | 0.8207 | 0.8206 | 0.8207 |
87
- | 0.0452 | 4.05 | 900 | 0.9594 | 0.8259 | 0.8207 | 0.8209 | 0.8207 |
88
- | 0.0027 | 4.19 | 930 | 0.9452 | 0.8274 | 0.8235 | 0.8237 | 0.8235 |
89
- | 0.0193 | 4.32 | 960 | 0.9259 | 0.8426 | 0.8375 | 0.8379 | 0.8375 |
90
- | 0.0313 | 4.46 | 990 | 0.9691 | 0.8334 | 0.8263 | 0.8264 | 0.8263 |
91
- | 0.0027 | 4.59 | 1020 | 0.9779 | 0.8210 | 0.8151 | 0.8151 | 0.8151 |
92
- | 0.0153 | 4.73 | 1050 | 0.9779 | 0.8172 | 0.8123 | 0.8120 | 0.8123 |
93
- | 0.0312 | 4.86 | 1080 | 0.9776 | 0.8222 | 0.8179 | 0.8179 | 0.8179 |
94
- | 0.017 | 5.0 | 1110 | 0.9755 | 0.8239 | 0.8179 | 0.8178 | 0.8179 |
95
 
96
 
97
  ### Framework versions
98
 
99
- - Transformers 4.39.1
100
  - Pytorch 2.2.1+cu121
101
  - Datasets 2.18.0
102
  - Tokenizers 0.15.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  metrics:
 
18
 
19
  # trainerH
20
 
21
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.9240
24
+ - Precision: 0.8148
25
+ - Recall: 0.8095
26
+ - F1: 0.8085
27
+ - Accuracy: 0.8095
28
 
29
  ## Model description
30
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
+ | 1.9362 | 0.14 | 30 | 1.8590 | 0.2709 | 0.2521 | 0.2166 | 0.2521 |
59
+ | 1.7693 | 0.27 | 60 | 1.5403 | 0.4339 | 0.4538 | 0.4032 | 0.4538 |
60
+ | 1.4217 | 0.41 | 90 | 1.3401 | 0.4658 | 0.4706 | 0.3921 | 0.4706 |
61
+ | 1.3121 | 0.54 | 120 | 1.1863 | 0.6652 | 0.6190 | 0.6092 | 0.6190 |
62
+ | 1.1994 | 0.68 | 150 | 0.9959 | 0.7354 | 0.7115 | 0.7070 | 0.7115 |
63
+ | 0.9937 | 0.81 | 180 | 0.8418 | 0.7401 | 0.7227 | 0.7093 | 0.7227 |
64
+ | 0.7953 | 0.95 | 210 | 0.7477 | 0.7756 | 0.7563 | 0.7491 | 0.7563 |
65
+ | 0.6038 | 1.08 | 240 | 0.6583 | 0.7882 | 0.7843 | 0.7827 | 0.7843 |
66
+ | 0.5213 | 1.22 | 270 | 0.6735 | 0.7966 | 0.7843 | 0.7840 | 0.7843 |
67
+ | 0.4313 | 1.35 | 300 | 0.6661 | 0.8142 | 0.7871 | 0.7855 | 0.7871 |
68
+ | 0.4077 | 1.49 | 330 | 0.6784 | 0.7938 | 0.7731 | 0.7732 | 0.7731 |
69
+ | 0.4095 | 1.62 | 360 | 0.6522 | 0.8064 | 0.7983 | 0.7978 | 0.7983 |
70
+ | 0.3463 | 1.76 | 390 | 0.6827 | 0.8198 | 0.8095 | 0.8081 | 0.8095 |
71
+ | 0.3221 | 1.89 | 420 | 0.7073 | 0.8190 | 0.7983 | 0.7947 | 0.7983 |
72
+ | 0.3645 | 2.03 | 450 | 0.6673 | 0.8246 | 0.8179 | 0.8175 | 0.8179 |
73
+ | 0.1191 | 2.16 | 480 | 0.7730 | 0.8005 | 0.7955 | 0.7928 | 0.7955 |
74
+ | 0.1524 | 2.3 | 510 | 0.9105 | 0.8086 | 0.8011 | 0.7994 | 0.8011 |
75
+ | 0.1701 | 2.43 | 540 | 0.8482 | 0.8003 | 0.7955 | 0.7939 | 0.7955 |
76
+ | 0.2633 | 2.57 | 570 | 0.7490 | 0.8121 | 0.8039 | 0.8035 | 0.8039 |
77
+ | 0.191 | 2.7 | 600 | 0.7528 | 0.8016 | 0.7955 | 0.7959 | 0.7955 |
78
+ | 0.103 | 2.84 | 630 | 0.7891 | 0.8212 | 0.8151 | 0.8131 | 0.8151 |
79
+ | 0.1179 | 2.97 | 660 | 0.7611 | 0.8128 | 0.8067 | 0.8048 | 0.8067 |
80
+ | 0.065 | 3.11 | 690 | 0.7939 | 0.8208 | 0.8151 | 0.8141 | 0.8151 |
81
+ | 0.088 | 3.24 | 720 | 0.8769 | 0.8099 | 0.7955 | 0.7941 | 0.7955 |
82
+ | 0.1397 | 3.38 | 750 | 0.7566 | 0.8306 | 0.8263 | 0.8259 | 0.8263 |
83
+ | 0.0526 | 3.51 | 780 | 0.7722 | 0.8176 | 0.8151 | 0.8136 | 0.8151 |
84
+ | 0.0073 | 3.65 | 810 | 0.8817 | 0.8223 | 0.8123 | 0.8110 | 0.8123 |
85
+ | 0.0564 | 3.78 | 840 | 0.9068 | 0.8277 | 0.8179 | 0.8155 | 0.8179 |
86
+ | 0.0481 | 3.92 | 870 | 0.8703 | 0.8115 | 0.8067 | 0.8051 | 0.8067 |
87
+ | 0.0573 | 4.05 | 900 | 0.8761 | 0.8300 | 0.8207 | 0.8205 | 0.8207 |
88
+ | 0.0253 | 4.19 | 930 | 0.8925 | 0.8287 | 0.8207 | 0.8201 | 0.8207 |
89
+ | 0.0067 | 4.32 | 960 | 0.9603 | 0.8069 | 0.7983 | 0.7964 | 0.7983 |
90
+ | 0.0279 | 4.46 | 990 | 0.9283 | 0.8034 | 0.7983 | 0.7970 | 0.7983 |
91
+ | 0.0133 | 4.59 | 1020 | 0.9180 | 0.8119 | 0.8067 | 0.8053 | 0.8067 |
92
+ | 0.0154 | 4.73 | 1050 | 0.9232 | 0.8086 | 0.8039 | 0.8024 | 0.8039 |
93
+ | 0.0207 | 4.86 | 1080 | 0.9236 | 0.8086 | 0.8039 | 0.8024 | 0.8039 |
94
+ | 0.0305 | 5.0 | 1110 | 0.9240 | 0.8148 | 0.8095 | 0.8085 | 0.8095 |
95
 
96
 
97
  ### Framework versions
98
 
99
+ - Transformers 4.39.2
100
  - Pytorch 2.2.1+cu121
101
  - Datasets 2.18.0
102
  - Tokenizers 0.15.2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "distilbert-base-cased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
@@ -31,7 +31,6 @@
31
  "model_type": "distilbert",
32
  "n_heads": 12,
33
  "n_layers": 6,
34
- "output_past": true,
35
  "pad_token_id": 0,
36
  "problem_type": "single_label_classification",
37
  "qa_dropout": 0.1,
@@ -39,6 +38,6 @@
39
  "sinusoidal_pos_embds": false,
40
  "tie_weights_": true,
41
  "torch_dtype": "float32",
42
- "transformers_version": "4.39.1",
43
- "vocab_size": 28996
44
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
 
31
  "model_type": "distilbert",
32
  "n_heads": 12,
33
  "n_layers": 6,
 
34
  "pad_token_id": 0,
35
  "problem_type": "single_label_classification",
36
  "qa_dropout": 0.1,
 
38
  "sinusoidal_pos_embds": false,
39
  "tie_weights_": true,
40
  "torch_dtype": "float32",
41
+ "transformers_version": "4.39.2",
42
+ "vocab_size": 30522
43
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb7e3a4fb1337e91a995d17104e983c643c0a64a424fabab891ee8a0519eac74
3
- size 263160068
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69d2daf8f453723480d2c5d837cdb9099dc77133b8bb6271dc8558017422722d
3
+ size 267847948
runs/Mar29_17-19-55_ef5b0f37f27c/events.out.tfevents.1711732822.ef5b0f37f27c.1234.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d2e3c019c19324664e326ee47a83adb1c14f7987682a4e9ab7c8facea492c55
3
+ size 30328
runs/Mar29_17-19-55_ef5b0f37f27c/events.out.tfevents.1711733155.ef5b0f37f27c.1234.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3380acd839be3aeb0795ef2d5c99975691f6a68728094a5500feda02b8c02fd2
3
+ size 560
tokenizer_config.json CHANGED
@@ -43,9 +43,11 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
- "do_lower_case": false,
 
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
  "model_max_length": 512,
50
+ "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
53
  "strip_accents": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a21352c6380d4c37e10dd68f77416be4478d0d3b78212dc6a198a87e50a32840
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e4d5e579cb54793fe56674437c0d06448c302eaa2901282a3480ccd8bd2121c
3
  size 4920
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff