Edit model card

Model Card for Model ID

This model is designed to detect bias in text data. It analyzes text inputs to identify and classify types of biases, aiding in the development of more inclusive and fair AI systems. The model is fine-tuned from valurank/distilroberta-bias model for research purpose. The model is able to detect bias in formal language since the training corpus is news titles.

Model Details

Model Description

The data used for fine-tuning is MBIC dataset, which contains texts with bias labels.

The model is capable of classifying any text into Biased or Non_biased. Max length set for the tokenizer is 512.

  • Developed by: [More Information Needed]
  • Model type: DistillRoBERTa transformer
  • Language(s) (NLP): English
  • License: Apache 2.0
  • Finetuned from model: valurank/distilroberta-bias
  • Repository: To be uploaded

The following sections are under construction...

How to Get Started with the Model

Use the code below to get started with the model.

Link to the github demo page to be included

[More Information Needed]

Training Details


Size of the Dataset: 1700 entries

Preprocessing Steps: Tokenization using a pre-specified tokenizer, padding, and truncation to convert text to numerical features. Classes are encoded numerically.

Data Splitting Strategy: 80% training, 20% validation split, with a random state for reproducibility.

Optimization Algorithm: AdamW

Loss Function: CrossEntropyLoss, weighted by class frequencies to address class imbalance.

Learning Rate: 1e-5

Number of Epochs: 3

Batch Size: 16

Regularization Techniques: Gradient clipping is applied with a max norm of 1.0.

Model-Specific Hyperparameters: Scheduler with step size of 3 and gamma of 0.1 for learning rate decay.

Training time: around 150 iterations/s under CUDA pytorch, less than 10 minutes for training.

Monitoring Strategies: Training and validation losses and validation accuracy are monitored.

Details on the Validation Dataset: Generated from the same DataFrame df using a train-test split.

Techniques Used for Fine-tuning: Learning rate scheduler for adjusting the learning rate.

Challenges and Solutions

Challenges Faced During Training: Class imbalance is addressed through weighted CrossEntropyLoss.

Solutions and Techniques Applied: Calculation of class weights from the training data and applying gradient clipping.

Metrics

[More Information Needed]

Results

[More Information Needed]

Summary

Model Update Log

Downloads last month
48
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train D1V1DE/bias-detection