karoldobiczek
commited on
Commit
•
b9dcbaf
1
Parent(s):
04e2ca6
End of training
Browse files- README.md +32 -32
- model.safetensors +1 -1
- tokenizer.json +14 -2
README.md
CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
-
- Loss: 0.
|
21 |
-
- Accuracy: 0.
|
22 |
|
23 |
## Model description
|
24 |
|
@@ -50,36 +50,36 @@ The following hyperparameters were used during training:
|
|
50 |
|
51 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
52 |
|:-------------:|:------:|:----:|:---------------:|:--------:|
|
53 |
-
| No log | 0.0083 | 1 | 1.
|
54 |
-
| 1.
|
55 |
-
| 1.
|
56 |
-
| 1.
|
57 |
-
| 1.
|
58 |
-
|
|
59 |
-
|
|
60 |
-
|
|
61 |
-
|
|
62 |
-
| 0.
|
63 |
-
| 0.
|
64 |
-
| 0.
|
65 |
-
| 0.
|
66 |
-
| 0.
|
67 |
-
| 0.
|
68 |
-
| 0.
|
69 |
-
| 0.
|
70 |
-
| 0.
|
71 |
-
| 0.
|
72 |
-
| 0.
|
73 |
-
| 0.
|
74 |
-
| 0.
|
75 |
-
| 0.
|
76 |
-
| 0.
|
77 |
-
| 0.
|
78 |
-
| 0.
|
79 |
-
| 0.
|
80 |
-
| 0.
|
81 |
-
| 0.
|
82 |
-
| 0.
|
83 |
|
84 |
|
85 |
### Framework versions
|
|
|
17 |
|
18 |
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 0.7874
|
21 |
+
- Accuracy: 0.6660
|
22 |
|
23 |
## Model description
|
24 |
|
|
|
50 |
|
51 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
52 |
|:-------------:|:------:|:----:|:---------------:|:--------:|
|
53 |
+
| No log | 0.0083 | 1 | 1.0582 | 0.4980 |
|
54 |
+
| 1.0574 | 0.2149 | 26 | 1.0428 | 0.4980 |
|
55 |
+
| 1.0702 | 0.4215 | 51 | 1.0500 | 0.4980 |
|
56 |
+
| 1.1065 | 0.6281 | 76 | 1.0374 | 0.4980 |
|
57 |
+
| 1.0241 | 0.8347 | 101 | 1.0391 | 0.4980 |
|
58 |
+
| 1.0324 | 1.0 | 121 | 1.0191 | 0.4980 |
|
59 |
+
| 1.0324 | 1.0413 | 126 | 1.0097 | 0.4980 |
|
60 |
+
| 0.9751 | 1.2479 | 151 | 1.0542 | 0.4737 |
|
61 |
+
| 1.0134 | 1.4545 | 176 | 0.9746 | 0.5931 |
|
62 |
+
| 0.9276 | 1.6612 | 201 | 0.8633 | 0.5648 |
|
63 |
+
| 0.8469 | 1.8678 | 226 | 0.7729 | 0.6538 |
|
64 |
+
| 0.7992 | 2.0 | 242 | 0.7874 | 0.6660 |
|
65 |
+
| 0.8853 | 2.0744 | 251 | 0.8597 | 0.6680 |
|
66 |
+
| 0.6466 | 2.2810 | 276 | 0.7767 | 0.6498 |
|
67 |
+
| 0.778 | 2.4876 | 301 | 1.0588 | 0.6498 |
|
68 |
+
| 0.7202 | 2.6942 | 326 | 0.7493 | 0.6721 |
|
69 |
+
| 0.7108 | 2.9008 | 351 | 0.8892 | 0.6397 |
|
70 |
+
| 0.6354 | 3.0 | 363 | 0.8265 | 0.6579 |
|
71 |
+
| 0.7704 | 3.1074 | 376 | 0.7833 | 0.6781 |
|
72 |
+
| 0.6867 | 3.3140 | 401 | 0.9702 | 0.6478 |
|
73 |
+
| 0.6973 | 3.5207 | 426 | 1.0300 | 0.6700 |
|
74 |
+
| 0.6682 | 3.7273 | 451 | 0.8206 | 0.6781 |
|
75 |
+
| 0.6605 | 3.9339 | 476 | 0.8862 | 0.6822 |
|
76 |
+
| 0.8521 | 4.0 | 484 | 0.8093 | 0.6316 |
|
77 |
+
| 0.6442 | 4.1405 | 501 | 0.9483 | 0.6437 |
|
78 |
+
| 0.577 | 4.3471 | 526 | 0.8860 | 0.6883 |
|
79 |
+
| 0.5252 | 4.5537 | 551 | 0.8797 | 0.7045 |
|
80 |
+
| 0.5274 | 4.7603 | 576 | 0.7289 | 0.7024 |
|
81 |
+
| 0.467 | 4.9669 | 601 | 0.8224 | 0.6903 |
|
82 |
+
| 0.467 | 5.0 | 605 | 0.8218 | 0.6903 |
|
83 |
|
84 |
|
85 |
### Framework versions
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1421499516
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee757e7d76ccb4b73122bae3db08be0e759e57e44335c85ee0c7b7a9e9b35b7c
|
3 |
size 1421499516
|
tokenizer.json
CHANGED
@@ -1,7 +1,19 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
"added_tokens": [
|
6 |
{
|
7 |
"id": 0,
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": {
|
4 |
+
"direction": "Right",
|
5 |
+
"max_length": 256,
|
6 |
+
"strategy": "LongestFirst",
|
7 |
+
"stride": 0
|
8 |
+
},
|
9 |
+
"padding": {
|
10 |
+
"strategy": "BatchLongest",
|
11 |
+
"direction": "Right",
|
12 |
+
"pad_to_multiple_of": null,
|
13 |
+
"pad_id": 1,
|
14 |
+
"pad_type_id": 0,
|
15 |
+
"pad_token": "<pad>"
|
16 |
+
},
|
17 |
"added_tokens": [
|
18 |
{
|
19 |
"id": 0,
|