DeBERTa trained from scratch continued training from https://huggingface.co/mikesong724/deberta-wiki-2006 Source data: https://dumps.wikimedia.org/archive/2010/ Tools used: https://github.com/mikesong724/Point-in-Time-Language-Model 2010 wiki archive 6.1 GB trained 18 epochs = 108GB + 2006 (65GB) GLUE benchmark cola (3e): matthews corr: 0.3640 sst2 (3e): acc: 0.9106 mrpc (5e): F1: 0.8505, acc: 0.7794 stsb (3e): pearson: 0.8339, spearman: 0.8312 qqp (3e): acc: 0.8965, F1: 0.8604 mnli (3e): acc_mm: 0.8023 qnli (3e): acc: 0.8889 rte (3e): acc: 0.5271 wnli (5e): acc: 0.3380