shirzady1934 commited on
Commit
0f38953
1 Parent(s): 9976cc7

End of training

Browse files
Files changed (5) hide show
  1. README.md +22 -17
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. tokenizer.json +7 -5
  5. training_args.bin +1 -1
README.md CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.2420
21
- - Accuracy: 0.9000
22
 
23
  ## Model description
24
 
@@ -37,33 +37,38 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 0.001
41
- - train_batch_size: 16
42
- - eval_batch_size: 16
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - num_epochs: 10
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
- | No log | 1.0 | 23 | 0.4131 | 0.875 |
53
- | No log | 2.0 | 46 | 0.4599 | 0.8250 |
54
- | No log | 3.0 | 69 | 0.3552 | 0.8250 |
55
- | No log | 4.0 | 92 | 0.3887 | 0.8500 |
56
- | No log | 5.0 | 115 | 0.3795 | 0.875 |
57
- | No log | 6.0 | 138 | 0.3703 | 0.875 |
58
- | No log | 7.0 | 161 | 0.2565 | 0.875 |
59
- | No log | 8.0 | 184 | 0.4069 | 0.875 |
60
- | No log | 9.0 | 207 | 0.2420 | 0.9000 |
61
- | No log | 10.0 | 230 | 0.2521 | 0.9000 |
 
 
 
 
 
62
 
63
 
64
  ### Framework versions
65
 
66
- - Transformers 4.36.2
67
  - Pytorch 1.13.1+cu117
68
  - Datasets 2.16.1
69
  - Tokenizers 0.15.0
 
17
 
18
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.1949
21
+ - Accuracy: 0.875
22
 
23
  ## Model description
24
 
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - learning_rate: 0.0005
41
+ - train_batch_size: 32
42
+ - eval_batch_size: 32
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - num_epochs: 15
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | No log | 1.0 | 12 | 0.4873 | 0.7250 |
53
+ | No log | 2.0 | 24 | 0.3075 | 0.8000 |
54
+ | No log | 3.0 | 36 | 0.4185 | 0.8000 |
55
+ | No log | 4.0 | 48 | 0.3031 | 0.8250 |
56
+ | No log | 5.0 | 60 | 0.3392 | 0.8250 |
57
+ | No log | 6.0 | 72 | 0.3420 | 0.8500 |
58
+ | No log | 7.0 | 84 | 0.3509 | 0.8250 |
59
+ | No log | 8.0 | 96 | 0.3089 | 0.8500 |
60
+ | No log | 9.0 | 108 | 0.2797 | 0.8250 |
61
+ | No log | 10.0 | 120 | 0.2378 | 0.8000 |
62
+ | No log | 11.0 | 132 | 0.2622 | 0.875 |
63
+ | No log | 12.0 | 144 | 0.2334 | 0.9000 |
64
+ | No log | 13.0 | 156 | 0.2314 | 0.9000 |
65
+ | No log | 14.0 | 168 | 0.1987 | 0.875 |
66
+ | No log | 15.0 | 180 | 0.1949 | 0.875 |
67
 
68
 
69
  ### Framework versions
70
 
71
+ - Transformers 4.37.2
72
  - Pytorch 1.13.1+cu117
73
  - Datasets 2.16.1
74
  - Tokenizers 0.15.0
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.36.2",
24
  "type_vocab_size": 1,
25
  "use_cache": true,
26
  "vocab_size": 50265
 
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "torch_dtype": "float32",
23
+ "transformers_version": "4.37.2",
24
  "type_vocab_size": 1,
25
  "use_cache": true,
26
  "vocab_size": 50265
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18a7926350cb93a28c2976739952431bfa430489a521341bfbaecc683d9fff0c
3
  size 499796292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0178adadfbe5914188c90e24aaac5f1cfe755dcddbf09a04ceba3be5ff269334
3
  size 499796292
tokenizer.json CHANGED
@@ -1,12 +1,14 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
 
 
4
  "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
 
8
  },
9
- "padding": null,
10
  "added_tokens": [
11
  {
12
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": {
5
+ "strategy": "BatchLongest",
6
  "direction": "Right",
7
+ "pad_to_multiple_of": null,
8
+ "pad_id": 1,
9
+ "pad_type_id": 0,
10
+ "pad_token": "<pad>"
11
  },
 
12
  "added_tokens": [
13
  {
14
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d88e308546b4db6eb88f2047f60247fa7326cf3893313af5e058760ab76febf
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c4f43df0841f80e5cefa0d1c31b75a750c448e281b48fe55e88d50acca17bd7
3
  size 4283