Edit model card

Model Name: DistilBERT for Sentiment Analysis

Model Description

Overview

This model is a fine-tuned version of distilbert-base-uncased on a social media dataset for the purpose of sentiment analysis. It can classify text into non-negative and negative sentiments.

Intended Use

This model is intended for sentiment analysis tasks, particularly for analyzing social media texts.

Model Architecture

This model is based on the DistilBertForSequenceClassification architecture, a distilled version of BERT that maintains comparable performance on downstream tasks while being more computationally efficient.

Training

Training Data

The model was trained on a dataset consisting of social media posts, surveys and interviews, labeled for sentiment (non-negative and negative). The dataset includes texts from a variety of sources and demographics.

Training Procedure

The model was trained using the following parameters:

  • Optimizer: AdamW
  • Learning Rate: 5e-5
  • Batch Size: 32
  • Epochs: 30

Training was conducted on Kaggle, utilizing two GPUs for accelerated training.

Downloads last month
167
Safetensors
Model size
67M params
Tensor type
F32
·