license: mit | |
language: protein | |
tags: | |
- protein language model | |
datasets: | |
- Uniref50 | |
# DistilProtBert model | |
Distilled protein language of [ProtBert](https://huggingface.co/Rostlab/prot_bert). | |
In addition to cross entropy and cosine teacher-student losses, DistilProtBert was pretrained on a masked language modeling (MLM) objective and it only works with capital letter amino acids. | |
# Model description | |