bizhen commited on
Commit
d27d23e
1 Parent(s): a7723a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -1
README.md CHANGED
@@ -1 +1,26 @@
1
- Pretrained model on protein sequences using masked language modeling (MLM) and knowledge embedding (KE) objective objective. It was introduced in [this paper](https://openreview.net/pdf?id=yfe1VMYAXa4) and first released in [this repository](https://github.com/zjunlp/OntoProtein). This model is trained on uppercase amino acids: it only works with capital letter amino acids.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: protein
3
+ tags:
4
+ - protein language model
5
+ datasets:
6
+ - ProteinKG25
7
+ widget:
8
+ - text: "D L I P T S S K L V V [MASK] D T S L Q V K K A F F A L V T"
9
+
10
+ ---
11
+
12
+ # OntoProtein model
13
+ Pretrained model on protein sequences using masked language modeling (MLM) and knowledge embedding (KE) objective objective. It was introduced in [this paper](https://openreview.net/pdf?id=yfe1VMYAXa4) and first released in [this repository](https://github.com/zjunlp/OntoProtein). This model is trained on uppercase amino acids: it only works with capital letter amino acids.
14
+
15
+ ## Model description
16
+ OntoProtein is the first general framework that makes use of structure in GO (Gene Ontology) into protein pre-training models. We construct a novel large-scale knowledge graph that consists of GO and its related proteins, and gene annotation texts or protein sequences describe all nodes in the graph. We propose novel contrastive learning with knowledge-aware negative sampling to jointly optimize the knowledge graph and protein embedding during pre-training.
17
+
18
+ ### BibTeX entry and citation info
19
+ ```bibtex
20
+ @article{zhang2022ontoprotein,
21
+ title={OntoProtein: Protein Pretraining With Gene Ontology Embedding},
22
+ author={Zhang, Ningyu and Bi, Zhen and Liang, Xiaozhuan and Cheng, Siyuan and Hong, Haosen and Deng, Shumin and Lian, Jiazhang and Zhang, Qiang and Chen, Huajun},
23
+ journal={arXiv preprint arXiv:2201.11147},
24
+ year={2022}
25
+ }
26
+ ```