learnrr's picture
Update README.md
4df126f verified
metadata
language: en
license: mit
base_model: roberta-base
tags:
  - token-classification
  - ner
  - named-entity-recognition
datasets:
  - conll2003
metrics:
  - f1
  - precision
  - recall
  - accuracy
model-index:
  - name: RoBERTa-base-NER-CoNLL2003
    results:
      - task:
          type: token-classification
          name: Named Entity Recognition
        dataset:
          type: conll2003
          name: CoNLL-2003 (English)
        metrics:
          - type: f1
            value: 95.99

Model description

This model is a fine-tuned version of roberta-base for the Named Entity Recognition (NER) task using the CoNLL-2003 dataset. It can identify four types of entities: Persons (PER), Organizations (ORG), Locations (LOC), and Miscellaneous (MISC).

Training procedure

  • Hardware: NVIDIA V100 GPU
  • Optimizer: AdamW
  • Learning Rate: 2e-5
  • Batch Size: 16
  • Weight Decay: 0.01
  • Epochs: 5
  • Mixed Precision Training: FP16 enabled

Evaluation Results

Metric) Value
F1 Score 95.99%
Precision 95.61%
Recall 96.38%
Accuracy 99.29%
Eval Loss 0.0464

How to use

from transformers import pipeline
model_id = "learnrr/roberta-NER-conll2003"
text = "Apple is looking at buying U.K. startup for $1 billion"
results = nlp(text)
for entity in results:
print(f"entity: {entity['word']} | class: {entity['entity_group']} | confidence: {entity['score']:.4f}")