roberta-swahili / flax_to_torch.py
Patrick von Platen
model with clean data
a71d002
raw
history blame
137 Bytes
from transformers import RobertaForMaskedLM
model = RobertaForMaskedLM.from_pretrained("./", from_flax=True)
model.save_pretrained("./")