EconBERTa - RoBERTa further trained for 25k steps (T=512, batch_size = 256) on text sourced from economics books. Example usage for MLM: ```python from transformers import RobertaTokenizer, RobertaForMaskedLM from transformers import pipeline tokenizer = RobertaTokenizer.from_pretrained('roberta-base') model = RobertaForMaskedLM.from_pretrained('models').cpu() model.eval() mlm = pipeline('fill-mask', model = model, tokenizer = tokenizer) test = "ECB - euro, FED - , BoJ - yen" print(mlm(test)[:2]) [{'sequence': 'ECB - euro, FED - dollar, BoJ - yen', 'score': 0.7342271208763123, 'token': 1404, 'token_str': ' dollar'}, {'sequence': 'ECB - euro, FED - dollars, BoJ - yen', 'score': 0.10828445851802826, 'token': 1932, 'token_str': ' dollars'}] ```