|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- sst2 |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
library_name: transformers |
|
pipeline_tag: text-classification |
|
widget: |
|
- text: "this film 's relationship to actual tension is the same as what christmas-tree flocking in a spray can is to actual snow : a poor -- if durable -- imitation ." |
|
example_title: "negative" |
|
- text: "director rob marshall went out gunning to make a great one ." |
|
example_title: "positive" |
|
--- |
|
|
|
# bert-base-uncased-finetuned-sst2-v2 |
|
|
|
BERT (`"bert-base-uncased"`) finetuned on SST-2 (Stanford Sentiment Treebank Binary). |
|
|
|
This model pertains to the "Try it out!" exercise in section 4 of chapter 3 of the Hugging Face "NLP Course" (https://huggingface.co/learn/nlp-course/chapter3/4). |
|
|
|
It was trained using a custom PyTorch loop without Hugging Face Accelerate. |
|
|
|
Code: https://github.com/sambitmukherjee/hf-nlp-course-exercises/blob/main/chapter3/section4.ipynb |
|
|
|
Experiment tracking: https://wandb.ai/sadhaklal/bert-base-uncased-finetuned-sst2-v2 |
|
|
|
## Usage |
|
|
|
``` |
|
from transformers import pipeline |
|
|
|
classifier = pipeline("text-classification", model="sadhaklal/bert-base-uncased-finetuned-sst2-v2") |
|
print(classifier("uneasy mishmash of styles and genres .")) |
|
print(classifier("by the end of no such thing the audience , like beatrice , has a watchful affection for the monster .")) |
|
``` |
|
|
|
## Dataset |
|
|
|
From the dataset page: |
|
|
|
> The Stanford Sentiment Treebank is a corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language... |
|
|
|
> Binary classification experiments on full sentences (negative or somewhat negative vs somewhat positive or positive with neutral sentences discarded) refer to the dataset as SST-2 or SST binary. |
|
|
|
Examples: https://huggingface.co/datasets/sst2/viewer |
|
|
|
## Metric |
|
|
|
Accuracy on the `'validation'` split of SST-2: 0.9278 |
|
|