YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

TextAttack Model Card

This bert-base-uncased model was fine-tuned for sequence classification using TextAttack and the glue dataset loaded using the nlp library. The model was fine-tuned for 5 epochs with a batch size of 16, a learning rate of 2e-05, and a maximum sequence length of 256. Since this was a classification task, the model was trained with a cross-entropy loss function. The best score the model achieved on this task was 0.8774509803921569, as measured by the eval set accuracy, found after 1 epoch.

For more information, check out TextAttack on Github.

Downloads last month
528
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Spaces using textattack/bert-base-uncased-MRPC 4