Edit model card

Model Card of instructionRoberta-base for Bertology

roBERTa illustration

A minimalistic instruction model with an already good analysed and pretrained encoder like roBERTa. So we can research the Bertology with instruction-tuned models, look at the attention and investigate what happens to BERT embeddings during fine-tuning.

The training code is released at the instructionBERT repository. We used the Huggingface API for warm-starting BertGeneration with Encoder-Decoder-Models for this purpose.

Run the model with a longer output

from transformers import AutoTokenizer, EncoderDecoderModel
# load the fine-tuned seq2seq model and corresponding tokenizer
model_name = "Bachstelze/instructionRoberta-base"
model = EncoderDecoderModel.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
input = "Write a poem about love, peace and pancake."
input_ids = tokenizer(input, return_tensors="pt").input_ids
output_ids = model.generate(input_ids, max_new_tokens=200)
print(tokenizer.decode(output_ids[0]))

Training parameters

  • base model: "roberta-base"
  • trained for 1 epoche
  • batch size of 16
  • 20000 warm-up steps
  • learning rate of 0.0001

Purpose of instructionRoberta-base

InstructionBERT is intended for research purposes. The model-generated text should be treated as a starting point rather than a definitive solution for potential use cases. Users should be cautious when employing these models in their applications.

Downloads last month
46
Safetensors
Model size
154M params
Tensor type
F32
·

Datasets used to train Bachstelze/instructionRoberta-base