akmaldika commited on
Commit
bd307fc
·
verified ·
1 Parent(s): 1440dd2

akmaldika/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext-medical-question-classifier

Browse files
Files changed (4) hide show
  1. README.md +64 -0
  2. config.json +70 -0
  3. model.safetensors +3 -0
  4. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: mit
4
+ base_model: microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext
5
+ tags:
6
+ - generated_from_trainer
7
+ metrics:
8
+ - f1
9
+ model-index:
10
+ - name: results
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # results
18
+
19
+ This model is a fine-tuned version of [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.0078
22
+ - F1: 0.9961
23
+
24
+ ## Model description
25
+
26
+ More information needed
27
+
28
+ ## Intended uses & limitations
29
+
30
+ More information needed
31
+
32
+ ## Training and evaluation data
33
+
34
+ More information needed
35
+
36
+ ## Training procedure
37
+
38
+ ### Training hyperparameters
39
+
40
+ The following hyperparameters were used during training:
41
+ - learning_rate: 5e-05
42
+ - train_batch_size: 16
43
+ - eval_batch_size: 16
44
+ - seed: 42
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
+ - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_steps: 500
48
+ - num_epochs: 3
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss | F1 |
53
+ |:-------------:|:-----:|:----:|:---------------:|:------:|
54
+ | 0.0215 | 1.0 | 897 | 0.0172 | 0.9848 |
55
+ | 0.0014 | 2.0 | 1794 | 0.0149 | 0.9870 |
56
+ | 0.0006 | 3.0 | 2691 | 0.0078 | 0.9961 |
57
+
58
+
59
+ ### Framework versions
60
+
61
+ - Transformers 4.46.3
62
+ - Pytorch 2.5.1+cu121
63
+ - Datasets 3.2.0
64
+ - Tokenizers 0.20.3
config.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "causes",
13
+ "1": "complications",
14
+ "2": "considerations",
15
+ "3": "desc",
16
+ "4": "dont answer",
17
+ "5": "exams and tests",
18
+ "6": "frequency",
19
+ "7": "general",
20
+ "8": "genetic changes",
21
+ "9": "information",
22
+ "10": "inheritance",
23
+ "11": "outlook",
24
+ "12": "precaution",
25
+ "13": "prevention",
26
+ "14": "research",
27
+ "15": "stages",
28
+ "16": "support groups",
29
+ "17": "susceptibility",
30
+ "18": "symptoms",
31
+ "19": "treatment"
32
+ },
33
+ "initializer_range": 0.02,
34
+ "intermediate_size": 3072,
35
+ "label2id": {
36
+ "causes": 0,
37
+ "complications": 1,
38
+ "considerations": 2,
39
+ "desc": 3,
40
+ "dont answer": 4,
41
+ "exams and tests": 5,
42
+ "frequency": 6,
43
+ "general": 7,
44
+ "genetic changes": 8,
45
+ "information": 9,
46
+ "inheritance": 10,
47
+ "outlook": 11,
48
+ "precaution": 12,
49
+ "prevention": 13,
50
+ "research": 14,
51
+ "stages": 15,
52
+ "support groups": 16,
53
+ "susceptibility": 17,
54
+ "symptoms": 18,
55
+ "treatment": 19
56
+ },
57
+ "layer_norm_eps": 1e-12,
58
+ "max_position_embeddings": 512,
59
+ "model_type": "bert",
60
+ "num_attention_heads": 12,
61
+ "num_hidden_layers": 12,
62
+ "pad_token_id": 0,
63
+ "position_embedding_type": "absolute",
64
+ "problem_type": "single_label_classification",
65
+ "torch_dtype": "float32",
66
+ "transformers_version": "4.46.3",
67
+ "type_vocab_size": 2,
68
+ "use_cache": true,
69
+ "vocab_size": 30522
70
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:159793ceeb76e559c05b6a484acd28845a6d62508e4753e875911c822c7d7509
3
+ size 438014016
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cb13d1ecb68e1c4e02e1458bf7b5342efd919f74be3c1aed1eb841ef70d0de2
3
+ size 5240