mfarrington
commited on
End of training
Browse files
README.md
CHANGED
@@ -1,66 +1,74 @@
|
|
1 |
---
|
2 |
-
|
3 |
-
|
4 |
-
-
|
5 |
metrics:
|
|
|
|
|
|
|
6 |
- accuracy
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
---
|
11 |
-
# Model Card for Model ID
|
12 |
-
|
13 |
-
Pretrained BioBERT Model for Performing Named Entity Recognition (NER) of Medical Device Names, Components and Part Numbers.
|
14 |
-
|
15 |
-
## Model Details
|
16 |
-
|
17 |
-
### Abstract
|
18 |
-
*FDA Medical Device recalls are critical and time-sensitive events, requiring
|
19 |
-
swift identification of impacted devices to inform the public of a recall event and
|
20 |
-
ensure patient safety. The OpenFDA device recall dataset contains valuable
|
21 |
-
information about ongoing device recall actions, but manually extracting relevant
|
22 |
-
device information from the recall action summaries is a time-consuming task.
|
23 |
-
Named Entity Recognition (NER) is a task in Natural Language Processing
|
24 |
-
(NLP) that involves identifying and categorizing named entities in unstructured text.
|
25 |
-
Existing NER models, including domain-specific models like BioBERT, struggle
|
26 |
-
to correctly identify medical device trade names, part numbers and component terms within these
|
27 |
-
summaries. To address this, we propose DeviceBERT, a medical device annotation, pre-processing and enrichment pipeline, which builds on BioBERT to identify and label medical device terminology in the device recall summaries with improved accuracy. Furthermore, we demonstrate that our approach can be applied effectively for performing entity recognition tasks where training data is limited or sparse.*
|
28 |
-
|
29 |
-
### Model Description
|
30 |
|
31 |
-
This model
|
32 |
-
|
33 |
|
34 |
-
|
35 |
|
36 |
-
|
37 |
-
|
|
|
|
|
|
|
|
|
|
|
38 |
|
39 |
-
|
40 |
-
- **Model type:** Deep Learning Language Model/LLM
|
41 |
-
- **Language(s) (NLP):** Python, TensorFlow
|
42 |
-
- **License:** MIT
|
43 |
-
- **Finetuned from model [optional]:** BioBERT (dmis-lab/biobert-base-cased-v1.2)
|
44 |
|
45 |
-
|
46 |
|
47 |
-
|
48 |
-
- **Paper:** "DeviceBERT: Applied Transfer Learning With Targeted Annotations and Vocabulary Enrichment to Identify Medical Device and Component Terminology in FDA Recall Summaries"
|
49 |
|
50 |
-
|
51 |
|
52 |
-
|
53 |
|
54 |
-
|
55 |
-
The model can be applied further to generate human feedback loops using inferencing to generate additional NER data for more complex downstream tasks or additional finetuning.
|
56 |
|
57 |
-
## Training
|
58 |
|
59 |
-
### Training
|
60 |
|
61 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
|
|
|
63 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
|
65 |
|
|
|
66 |
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model: dmis-lab/biobert-base-cased-v1.2
|
3 |
+
tags:
|
4 |
+
- generated_from_trainer
|
5 |
metrics:
|
6 |
+
- precision
|
7 |
+
- recall
|
8 |
+
- f1
|
9 |
- accuracy
|
10 |
+
model-index:
|
11 |
+
- name: devicebert-base-cased-v1.0
|
12 |
+
results: []
|
13 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
16 |
+
should probably proofread and complete it, then remove this comment. -->
|
17 |
|
18 |
+
# devicebert-base-cased-v1.0
|
19 |
|
20 |
+
This model is a fine-tuned version of [dmis-lab/biobert-base-cased-v1.2](https://huggingface.co/dmis-lab/biobert-base-cased-v1.2) on an unknown dataset.
|
21 |
+
It achieves the following results on the evaluation set:
|
22 |
+
- Loss: nan
|
23 |
+
- Precision: 0.6816
|
24 |
+
- Recall: 0.6691
|
25 |
+
- F1: 0.6753
|
26 |
+
- Accuracy: 0.8547
|
27 |
|
28 |
+
## Model description
|
|
|
|
|
|
|
|
|
29 |
|
30 |
+
More information needed
|
31 |
|
32 |
+
## Intended uses & limitations
|
|
|
33 |
|
34 |
+
More information needed
|
35 |
|
36 |
+
## Training and evaluation data
|
37 |
|
38 |
+
More information needed
|
|
|
39 |
|
40 |
+
## Training procedure
|
41 |
|
42 |
+
### Training hyperparameters
|
43 |
|
44 |
+
The following hyperparameters were used during training:
|
45 |
+
- learning_rate: 1e-05
|
46 |
+
- train_batch_size: 16
|
47 |
+
- eval_batch_size: 16
|
48 |
+
- seed: 42
|
49 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
50 |
+
- lr_scheduler_type: linear
|
51 |
+
- num_epochs: 10
|
52 |
|
53 |
+
### Training results
|
54 |
|
55 |
+
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
56 |
+
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
57 |
+
| No log | 1.0 | 101 | nan | 0.5981 | 0.5740 | 0.5858 | 0.8131 |
|
58 |
+
| No log | 2.0 | 202 | nan | 0.6673 | 0.6197 | 0.6427 | 0.8424 |
|
59 |
+
| No log | 3.0 | 303 | nan | 0.6926 | 0.6673 | 0.6797 | 0.8498 |
|
60 |
+
| No log | 4.0 | 404 | nan | 0.686 | 0.6271 | 0.6552 | 0.8473 |
|
61 |
+
| 0.3891 | 5.0 | 505 | nan | 0.6853 | 0.6490 | 0.6667 | 0.8539 |
|
62 |
+
| 0.3891 | 6.0 | 606 | nan | 0.6857 | 0.7020 | 0.6938 | 0.8563 |
|
63 |
+
| 0.3891 | 7.0 | 707 | nan | 0.6900 | 0.6673 | 0.6784 | 0.8580 |
|
64 |
+
| 0.3891 | 8.0 | 808 | nan | 0.6795 | 0.6782 | 0.6789 | 0.8514 |
|
65 |
+
| 0.3891 | 9.0 | 909 | nan | 0.6906 | 0.6691 | 0.6797 | 0.8571 |
|
66 |
+
| 0.1315 | 10.0 | 1010 | nan | 0.6816 | 0.6691 | 0.6753 | 0.8547 |
|
67 |
|
68 |
|
69 |
+
### Framework versions
|
70 |
|
71 |
+
- Transformers 4.41.2
|
72 |
+
- Pytorch 2.3.0+cu121
|
73 |
+
- Datasets 2.19.2
|
74 |
+
- Tokenizers 0.19.1
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 928741200
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d5ac909e05c9af2fedd3552dc8c5b4c4357d3155f34bac3d225d04bce007404
|
3 |
size 928741200
|
runs/Jun07_18-40-05_6d0acad3a460/events.out.tfevents.1717785615.6d0acad3a460.17459.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6fbb28ea6c9a34b9f30d89c36c94159ccd47563a6dd8cd9486ab0aed78bf260
|
3 |
+
size 10490
|