IreNkweke's picture
Update README.md
5211f31 verified
|
raw
history blame
3.51 kB
metadata
license: mit
base_model: dslim/bert-base-NER
tags:
  - generated_from_trainer
metrics:
  - precision
  - recall
  - f1
  - accuracy
model-index:
  - name: bert-finetuned-ner-ontonotes5
    results: []

Model Description

This model is a fine-tuned version of dslim/bert-base-NER on OntoNotes 5 dataset and is designed to identify and classify various types of entities in text, including persons, organizations, locations, dates, and more. It achieves the following results on the evaluation set:

  • Loss: 0.1634
  • Precision: 0.8620
  • Recall: 0.8849
  • F1: 0.8733
  • Accuracy: 0.9758

Intended uses & limitations

The model is intended for use in applications requiring NER, such as information extraction, text classification, and enhancing search capabilities by identifying key entities within the text. It can be used to identify entities in any English text, including news articles, social media posts, and legal documents.

Training and evaluation data

Training Data The model was fine-tuned on the OntoNotes 5 dataset. This dataset includes multiple types of named entities and is widely used for NER tasks. The dataset is annotated with the following entity tags:

CARDINAL: Numerical values

DATE: References to dates and periods

PERSON: Names of people

NORP: Nationalities, religious groups, political groups

GPE: Countries, cities, states

LAW: Named documents and legal entities

ORG: Organizations

PERCENT: Percentage values

ORDINAL: Ordinal numbers

MONEY: Monetary values

WORK_OF_ART: Titles of creative works

FAC: Facilities

TIME: Times smaller than a day

LOC: Non-GPE locations, mountain ranges, bodies of water

QUANTITY: Measurements, as of weight or distance

PRODUCT: Objects, vehicles, foods, etc. (not services)

EVENT: Named events

LANGUAGE: Named languages

Model Configuration

Base Model: dslim/bert-base-NER

Number of Labels: 37 (including the "O" tag for outside any named entity)

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Precision Recall F1 Accuracy
0.0372 1.0 7491 0.1188 0.8392 0.8799 0.8591 0.9738
0.04 2.0 14982 0.1182 0.8562 0.8824 0.8691 0.9754
0.0164 3.0 22473 0.1380 0.8561 0.8835 0.8696 0.9752
0.0117 4.0 29964 0.1531 0.8618 0.8833 0.8724 0.9758
0.0054 5.0 37455 0.1634 0.8620 0.8849 0.8733 0.9758

Framework versions

  • Transformers 4.41.2
  • Pytorch 2.3.0+cu121
  • Datasets 2.20.0
  • Tokenizers 0.19.1

Contact Information

For questions, comments, or issues with the model, please contact:

Name: [Irechukwu Nkweke]

Email: [mnkweke@yahoo.com]

GitHub: [https://github.com/mnkweke]

Acknowledgments

This model was trained using the Hugging Face transformers library and the OntoNotes 5 dataset.