![](https://cdn-avatars.huggingface.co/v1/production/uploads/653153455002276ffc0f6a2c/qoqMuEmDCrdYNA20vV0Ot.jpeg)
giant-oak/lsg-roberta-base-4096
Fill-Mask
•
Updated
•
29
Various efficient attention encoder-style architectures distilled into student models with half the hidden layers, plus a long-context NER dataset