Edit model card

bert-base-uncased-finetuned-sst2-v2

BERT ("bert-base-uncased") finetuned on SST-2 (Stanford Sentiment Treebank Binary).

This model pertains to the "Try it out!" exercise in section 4 of chapter 3 of the Hugging Face "NLP Course" (https://huggingface.co/learn/nlp-course/chapter3/4).

It was trained using a custom PyTorch loop without Hugging Face Accelerate.

Code: https://github.com/sambitmukherjee/hf-nlp-course-exercises/blob/main/chapter3/section4.ipynb

Experiment tracking: https://wandb.ai/sadhaklal/bert-base-uncased-finetuned-sst2-v2

Usage

from transformers import pipeline

classifier = pipeline("text-classification", model="sadhaklal/bert-base-uncased-finetuned-sst2-v2")
print(classifier("uneasy mishmash of styles and genres ."))
print(classifier("by the end of no such thing the audience , like beatrice , has a watchful affection for the monster ."))

Dataset

From the dataset page:

The Stanford Sentiment Treebank is a corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language...

Binary classification experiments on full sentences (negative or somewhat negative vs somewhat positive or positive with neutral sentences discarded) refer to the dataset as SST-2 or SST binary.

Examples: https://huggingface.co/datasets/sst2/viewer

Metric

Accuracy on the 'validation' split of SST-2: 0.9278

Downloads last month
7
Safetensors
Model size
109M params
Tensor type
F32
·

Dataset used to train sadhaklal/bert-base-uncased-finetuned-sst2-v2