Edit model card

nickwong64/bert-base-uncased-poems-sentiment

Bert is a Transformer Bidirectional Encoder based Architecture trained on MLM(Mask Language Modeling) objective. bert-base-uncased finetuned on the poem_sentiment dataset using HuggingFace Trainer with below training parameters.

 learning rate 2e-5, 
 batch size 8,
 num_train_epochs=8,

Model Performance

Epoch Training Loss Validation Loss Accuracy F1
8 0.468200 0.458632 0.904762 0.899756

How to Use the Model

from transformers import pipeline
nlp = pipeline(task='text-classification', 
               model='nickwong64/bert-base-uncased-poems-sentiment')
p1 = "No man is an island, Entire of itself, Every man is a piece of the continent, A part of the main."
p2 = "Ten years, dead and living dim and draw apart. I don’t try to remember, But forgetting is hard."
p3 = "My mind to me a kingdom is; Such present joys therein I find,That it excels all other bliss"
print(nlp(p1))
print(nlp(p2))
print(nlp(p3))
"""
output:
[{'label': 'no_impact', 'score': 0.9982421398162842}]
[{'label': 'negative', 'score': 0.9856176972389221}]
[{'label': 'positive', 'score': 0.9931322932243347}]
"""

Dataset

poem_sentiment

Labels

{0: 'negative', 1: 'positive', 2: 'no_impact', 3: 'mixed'}

Evaluation

{'test_loss': 0.4359096586704254,
 'test_accuracy': 0.9142857142857143,
 'test_f1': 0.9120554830816401,
 'test_runtime': 0.5689,
 'test_samples_per_second': 184.582,
 'test_steps_per_second': 24.611}
Downloads last month
176
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.