Model Card for Model ID

This model has been created with Argilla, trained with Transformers.

This is a sample model finetuned from prajjwal1/bert-tiny.

Model training

Training the model using the ArgillaTrainer:

# Load the dataset:
dataset = FeedbackDataset.from_huggingface("argilla/emotion")

# Create the training task:
task = TrainingTask.for_text_classification(text=dataset.field_by_name("text"), label=dataset.question_by_name("label"))

# Create the ArgillaTrainer:
trainer = ArgillaTrainer(
    dataset=dataset,
    task=task,
    framework="transformers",
    model="prajjwal1/bert-tiny",
)

trainer.update_config({
    "logging_steps": 1,
    "num_train_epochs": 1,
    "output_dir": "tmp"
})

trainer.train(output_dir="None")

You can test the type of predictions of this model like so:

trainer.predict("This is awesome!")

Model Details

Model Description

Model trained with ArgillaTrainer for demo purposes

  • Developed by: [More Information Needed]
  • Shared by [optional]: [More Information Needed]
  • Model type: Finetuned version of prajjwal1/bert-tiny for demo purposes
  • Language(s) (NLP): ['en']
  • License: apache-2.0
  • Finetuned from model [optional]: prajjwal1/bert-tiny

Model Sources [optional]

  • Repository: N/A

Technical Specifications [optional]

Framework Versions

  • Python: 3.10.7
  • Argilla: 1.19.0-dev
Downloads last month
105
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The HF Inference API does not support text-classification models for Transformers library.