YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Bengali NER Model

Model Description

This model is fine-tuned XLM-RoBERTa for Bengali Named Entity Recognition with proper tokenization handling.

Architecture Details

  • Base model: xlm-roberta-base
  • Task: Named Entity Recognition
  • Language: Bengali
  • Tokenization: XLM-RoBERTa tokenizer with proper Bengali word handling
  • Architecture:
    • BiLSTM layers: 2
    • Hidden size: 256
    • CRF decoder
  • Training Date: 2025-01-13

Tokenization

  • Uses XLM-RoBERTa tokenizer
  • Handles Bengali word splitting
  • Maintains proper label alignment for split words
  • Uses word_ids for token-to-word mapping

Performance Metrics

  • final_train_loss: 1.2787
  • final_val_loss: 3.5568
  • best_val_loss: 3.0780
  • epochs_trained: 5
  • training_args: {'epochs': 5, 'batch_size': 16, 'learning_rate': 2e-05, 'warmup_ratio': 0.1, 'gradient_accumulation_steps': 2, 'max_length': 91, 'lstm_hidden_size': 256, 'lstm_layers': 2, 'lstm_dropout': 0.2}
Downloads last month
38
Safetensors
Model size
559M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.