--- license: mit widget: - text: "I like you. I love you." --- ## roberta-large-mnli Trained by Facebook, [original source](https://github.com/pytorch/fairseq/tree/master/examples/roberta) ```bibtex @article{liu2019roberta, title = {RoBERTa: A Robustly Optimized BERT Pretraining Approach}, author = {Yinhan Liu and Myle Ott and Naman Goyal and Jingfei Du and Mandar Joshi and Danqi Chen and Omer Levy and Mike Lewis and Luke Zettlemoyer and Veselin Stoyanov}, journal={arXiv preprint arXiv:1907.11692}, year = {2019}, } ```