Edit model card

BatterySciBERT-uncased for Battery Abstract Multi-label Classification

This new model is a fine-tuned version of the BatterySciBERT-uncased model on a few-sample dataset of 1140 abstract of paper.
This model is uncased.

Hyperparameters

batch_size = 4
n_epochs = 16
base_LM_model = "batteryscibert-uncased"
learning_rate = 3e-5

Performance

"Validation Micro F1-score": 94.54,
"Test Micro F1-score": 93.42,

Details on the test set

Predicted Label Precision Recall F1 score
Coating 95.83% 76.67% 85.19%
Computation 86.96% 90.90% 88.89%
Doping 96.30% 100% 98.11%
Experiment 98.02% 93.40% 95.65%
Sodium layered oxide cathode 93.75% 91.84% 92.78%
Aggregate Metric Micro average 95.51% 91.42% 93.42%
Macro average 94.17% 90.56% 92.12%

Use in Transformers

from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline
model_name = "NoWayBack/batteryscibert-uncased-abstract-mtc"

# Get predictions
nlp = pipeline('text-classification', model=model_name, tokenizer=model_name, top_k=5)
input_string = "Sodium-ion batteries are among the most promising alternatives to lithium-based " \
               "technologies for grid and other energy storage applications due to their cost benefits " \
               "and sustainable resource supply. For the cathode—the component that largely determines the " \
               "energy density of a sodium-ion battery cell—one major category of materials is P2-type layered oxides."
res = nlp(input_string)

# Load model & tokenizer
model = AutoModelForSequenceClassification.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
Downloads last month
14
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.