oracat commited on
Commit
5eed682
1 Parent(s): d59d395

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -25
README.md CHANGED
@@ -9,27 +9,15 @@ model-index:
9
  results: []
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
  # bert-paper-classifier
16
 
17
- This model is a fine-tuned version of [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract) on the None dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 1.1069
20
- - Accuracy: 0.6475
21
-
22
- ## Model description
23
-
24
- More information needed
25
 
26
  ## Intended uses & limitations
27
 
28
- More information needed
29
 
30
- ## Training and evaluation data
31
-
32
- More information needed
33
 
34
  ## Training procedure
35
 
@@ -37,25 +25,17 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 5e-05
40
- - train_batch_size: 256
41
  - eval_batch_size: 32
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - num_epochs: 3
46
 
47
- ### Training results
48
-
49
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
50
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
51
- | 1.2101 | 1.0 | 704 | 1.1445 | 0.6359 |
52
- | 1.01 | 2.0 | 1408 | 1.1027 | 0.6472 |
53
- | 0.8619 | 3.0 | 2112 | 1.1069 | 0.6475 |
54
-
55
 
56
  ### Framework versions
57
 
58
  - Transformers 4.28.1
59
  - Pytorch 2.0.0+cu117
60
  - Datasets 2.11.0
61
- - Tokenizers 0.13.3
 
9
  results: []
10
  ---
11
 
 
 
 
12
  # bert-paper-classifier
13
 
14
+ This model is a fine-tuned version of [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract) on the dataset from [González-Márquez et al., 2023](https://www.biorxiv.org/content/10.1101/2023.04.10.536208v1).
 
 
 
 
 
 
 
15
 
16
  ## Intended uses & limitations
17
 
18
+ This model is intended to predict the category given the paper title (and optionally its abstract) — for the biomedical papers. For example, it is likely to predict `virology` as a category for the paper with a title containing `COVID-19`.
19
 
20
+ So far only a subset of the PubMed dataset has been used for training. Future improvements to this model can come with using the full dataset with a combination of titles and abstracts for the fine-tuning as well as extending the training set to the preprints from bioRxiv and/or arXiv.
 
 
21
 
22
  ## Training procedure
23
 
 
25
 
26
  The following hyperparameters were used during training:
27
  - learning_rate: 5e-05
28
+ - train_batch_size: 128
29
  - eval_batch_size: 32
30
  - seed: 42
31
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
32
  - lr_scheduler_type: linear
33
  - num_epochs: 3
34
 
 
 
 
 
 
 
 
 
35
 
36
  ### Framework versions
37
 
38
  - Transformers 4.28.1
39
  - Pytorch 2.0.0+cu117
40
  - Datasets 2.11.0
41
+ - Tokenizers 0.13.3