A-Bar commited on
Commit
517dddb
1 Parent(s): 972f209

Model save

Browse files
README.md ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: hongpingjun98/BioMedNLP_DeBERTa
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - sem_eval_2024_task_2
8
+ metrics:
9
+ - accuracy
10
+ - precision
11
+ - recall
12
+ - f1
13
+ model-index:
14
+ - name: BioMedNLP_DeBERTa_all_updates
15
+ results:
16
+ - task:
17
+ name: Text Classification
18
+ type: text-classification
19
+ dataset:
20
+ name: sem_eval_2024_task_2
21
+ type: sem_eval_2024_task_2
22
+ config: sem_eval_2024_task_2_source
23
+ split: validation
24
+ args: sem_eval_2024_task_2_source
25
+ metrics:
26
+ - name: Accuracy
27
+ type: accuracy
28
+ value: 0.655
29
+ - name: Precision
30
+ type: precision
31
+ value: 0.6551396256630968
32
+ - name: Recall
33
+ type: recall
34
+ value: 0.655
35
+ - name: F1
36
+ type: f1
37
+ value: 0.6549223575304444
38
+ ---
39
+
40
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
41
+ should probably proofread and complete it, then remove this comment. -->
42
+
43
+ # BioMedNLP_DeBERTa_all_updates
44
+
45
+ This model is a fine-tuned version of [hongpingjun98/BioMedNLP_DeBERTa](https://huggingface.co/hongpingjun98/BioMedNLP_DeBERTa) on the sem_eval_2024_task_2 dataset.
46
+ It achieves the following results on the evaluation set:
47
+ - Loss: 2.5118
48
+ - Accuracy: 0.655
49
+ - Precision: 0.6551
50
+ - Recall: 0.655
51
+ - F1: 0.6549
52
+
53
+ ## Model description
54
+
55
+ More information needed
56
+
57
+ ## Intended uses & limitations
58
+
59
+ More information needed
60
+
61
+ ## Training and evaluation data
62
+
63
+ More information needed
64
+
65
+ ## Training procedure
66
+
67
+ ### Training hyperparameters
68
+
69
+ The following hyperparameters were used during training:
70
+ - learning_rate: 5e-05
71
+ - train_batch_size: 16
72
+ - eval_batch_size: 16
73
+ - seed: 42
74
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
75
+ - lr_scheduler_type: linear
76
+ - lr_scheduler_warmup_steps: 500
77
+ - num_epochs: 50
78
+ - mixed_precision_training: Native AMP
79
+
80
+ ### Training results
81
+
82
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
83
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
84
+ | No log | 1.0 | 9 | 0.6482 | 0.62 | 0.6403 | 0.62 | 0.6058 |
85
+ | 0.7604 | 2.0 | 18 | 0.6376 | 0.635 | 0.6515 | 0.635 | 0.6248 |
86
+ | 0.7485 | 3.0 | 27 | 0.6256 | 0.655 | 0.6672 | 0.655 | 0.6486 |
87
+ | 0.7114 | 4.0 | 36 | 0.6188 | 0.675 | 0.6790 | 0.675 | 0.6732 |
88
+ | 0.6906 | 5.0 | 45 | 0.6181 | 0.705 | 0.7050 | 0.705 | 0.7050 |
89
+ | 0.5355 | 6.0 | 54 | 0.6257 | 0.68 | 0.6803 | 0.6800 | 0.6799 |
90
+ | 0.5411 | 7.0 | 63 | 0.6258 | 0.675 | 0.6754 | 0.675 | 0.6748 |
91
+ | 0.4849 | 8.0 | 72 | 0.6376 | 0.665 | 0.6670 | 0.665 | 0.6640 |
92
+ | 0.4386 | 9.0 | 81 | 0.6507 | 0.68 | 0.6826 | 0.6800 | 0.6788 |
93
+ | 0.3565 | 10.0 | 90 | 0.6631 | 0.685 | 0.6850 | 0.685 | 0.6850 |
94
+ | 0.3565 | 11.0 | 99 | 0.7089 | 0.66 | 0.6616 | 0.6600 | 0.6591 |
95
+ | 0.2992 | 12.0 | 108 | 0.7791 | 0.67 | 0.6717 | 0.6700 | 0.6692 |
96
+ | 0.2092 | 13.0 | 117 | 0.8224 | 0.68 | 0.6803 | 0.6800 | 0.6799 |
97
+ | 0.1643 | 14.0 | 126 | 0.9128 | 0.675 | 0.6750 | 0.675 | 0.6750 |
98
+ | 0.0811 | 15.0 | 135 | 1.0458 | 0.67 | 0.6701 | 0.67 | 0.6700 |
99
+ | 0.0502 | 16.0 | 144 | 1.2061 | 0.67 | 0.6701 | 0.67 | 0.6700 |
100
+ | 0.011 | 17.0 | 153 | 1.3763 | 0.655 | 0.6558 | 0.655 | 0.6546 |
101
+ | 0.0261 | 18.0 | 162 | 1.4862 | 0.655 | 0.6558 | 0.655 | 0.6546 |
102
+ | 0.0057 | 19.0 | 171 | 1.5609 | 0.665 | 0.6651 | 0.665 | 0.6649 |
103
+ | 0.0026 | 20.0 | 180 | 1.6435 | 0.655 | 0.6550 | 0.655 | 0.6550 |
104
+ | 0.0026 | 21.0 | 189 | 1.7122 | 0.655 | 0.6550 | 0.655 | 0.6550 |
105
+ | 0.0019 | 22.0 | 198 | 1.7682 | 0.655 | 0.6550 | 0.655 | 0.6550 |
106
+ | 0.0016 | 23.0 | 207 | 1.8163 | 0.655 | 0.6550 | 0.655 | 0.6550 |
107
+ | 0.0013 | 24.0 | 216 | 1.8590 | 0.655 | 0.6550 | 0.655 | 0.6550 |
108
+ | 0.0012 | 25.0 | 225 | 1.8883 | 0.66 | 0.6601 | 0.66 | 0.6600 |
109
+ | 0.001 | 26.0 | 234 | 1.9199 | 0.665 | 0.6651 | 0.665 | 0.6649 |
110
+ | 0.0008 | 27.0 | 243 | 1.9548 | 0.665 | 0.6651 | 0.665 | 0.6649 |
111
+ | 0.0007 | 28.0 | 252 | 1.9958 | 0.665 | 0.6658 | 0.665 | 0.6646 |
112
+ | 0.0007 | 29.0 | 261 | 2.0427 | 0.665 | 0.6658 | 0.665 | 0.6646 |
113
+ | 0.0006 | 30.0 | 270 | 2.0890 | 0.66 | 0.6601 | 0.66 | 0.6600 |
114
+ | 0.0006 | 31.0 | 279 | 2.1265 | 0.66 | 0.6601 | 0.66 | 0.6600 |
115
+ | 0.0005 | 32.0 | 288 | 2.1537 | 0.66 | 0.6601 | 0.66 | 0.6600 |
116
+ | 0.0077 | 33.0 | 297 | 2.1871 | 0.655 | 0.6550 | 0.655 | 0.6550 |
117
+ | 0.0004 | 34.0 | 306 | 2.2152 | 0.66 | 0.66 | 0.66 | 0.66 |
118
+ | 0.0004 | 35.0 | 315 | 2.2393 | 0.66 | 0.6601 | 0.66 | 0.6600 |
119
+ | 0.0003 | 36.0 | 324 | 2.2641 | 0.66 | 0.6601 | 0.66 | 0.6600 |
120
+ | 0.0003 | 37.0 | 333 | 2.2881 | 0.66 | 0.6601 | 0.66 | 0.6600 |
121
+ | 0.0008 | 38.0 | 342 | 2.3215 | 0.645 | 0.6462 | 0.645 | 0.6443 |
122
+ | 0.0005 | 39.0 | 351 | 2.3445 | 0.665 | 0.6650 | 0.665 | 0.6650 |
123
+ | 0.0426 | 40.0 | 360 | 2.3033 | 0.68 | 0.6818 | 0.6800 | 0.6792 |
124
+ | 0.0426 | 41.0 | 369 | 2.3582 | 0.66 | 0.6601 | 0.66 | 0.6600 |
125
+ | 0.0005 | 42.0 | 378 | 2.3550 | 0.66 | 0.6603 | 0.66 | 0.6599 |
126
+ | 0.0402 | 43.0 | 387 | 2.3575 | 0.665 | 0.6654 | 0.665 | 0.6648 |
127
+ | 0.0003 | 44.0 | 396 | 2.3372 | 0.675 | 0.6752 | 0.675 | 0.6749 |
128
+ | 0.0135 | 45.0 | 405 | 2.3467 | 0.66 | 0.6603 | 0.66 | 0.6599 |
129
+ | 0.0007 | 46.0 | 414 | 2.3033 | 0.685 | 0.6859 | 0.685 | 0.6846 |
130
+ | 0.0003 | 47.0 | 423 | 2.2770 | 0.675 | 0.6764 | 0.675 | 0.6743 |
131
+ | 0.0003 | 48.0 | 432 | 2.3131 | 0.68 | 0.6807 | 0.6800 | 0.6797 |
132
+ | 0.0002 | 49.0 | 441 | 2.4371 | 0.66 | 0.6601 | 0.66 | 0.6600 |
133
+ | 0.0004 | 50.0 | 450 | 2.5118 | 0.655 | 0.6551 | 0.655 | 0.6549 |
134
+
135
+
136
+ ### Framework versions
137
+
138
+ - Transformers 4.35.2
139
+ - Pytorch 2.1.0+cu121
140
+ - Datasets 2.16.1
141
+ - Tokenizers 0.15.0
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "hongpingjun98/BioMedNLP_DeBERTa",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.35.2",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 28895
27
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:919e12c6ceff6826eb2fa11610eaa7b2a9ef9bd8045b5313a061ee32db77110f
3
+ size 432960488
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 1000000000000000019884624838656,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37263d7e04c5fb332911d41572e95a31ef1c1eae6601fc6e15ddf4b1281ad4c6
3
+ size 4536
vocab.txt ADDED
The diff for this file is too large to render. See raw diff