Update README.md
Browse files
README.md
CHANGED
@@ -14,7 +14,7 @@ tags:
|
|
14 |
|
15 |
VHHBERT is a RoBERTa-based model pre-trained on two million VHH sequences in [VHHCorpus-2M](https://huggingface.co/datasets/COGNANO/VHHCorpus-2M).
|
16 |
VHHBERT has the same model parameters as RoBERTa<sub>BASE</sub>, except that it used positional embeddings with a length of 185 to cover the maximum sequence length of 179 in VHHCorpus-2M.
|
17 |
-
Further details on VHHBERT are described in our paper "A SARS-CoV-2 Interaction Dataset and VHH Sequence Corpus for Antibody Language Models.”
|
18 |
|
19 |
## Usage
|
20 |
|
@@ -30,12 +30,17 @@ model = RobertaModel.from_pretrained("tsurubee/VHHBERT")
|
|
30 |
|
31 |
- Pre-training Corpus: https://huggingface.co/datasets/COGNANO/VHHCorpus-2M
|
32 |
- Code: https://github.com/cognano/AVIDa-SARS-CoV-2
|
33 |
-
- Paper:
|
34 |
|
35 |
## Citation
|
36 |
|
37 |
If you use VHHBERT in your research, please cite the following paper.
|
38 |
|
39 |
```bibtex
|
40 |
-
|
|
|
|
|
|
|
|
|
|
|
41 |
```
|
|
|
14 |
|
15 |
VHHBERT is a RoBERTa-based model pre-trained on two million VHH sequences in [VHHCorpus-2M](https://huggingface.co/datasets/COGNANO/VHHCorpus-2M).
|
16 |
VHHBERT has the same model parameters as RoBERTa<sub>BASE</sub>, except that it used positional embeddings with a length of 185 to cover the maximum sequence length of 179 in VHHCorpus-2M.
|
17 |
+
Further details on VHHBERT are described in our paper "[A SARS-CoV-2 Interaction Dataset and VHH Sequence Corpus for Antibody Language Models](https://arxiv.org/abs/2405.18749).”
|
18 |
|
19 |
## Usage
|
20 |
|
|
|
30 |
|
31 |
- Pre-training Corpus: https://huggingface.co/datasets/COGNANO/VHHCorpus-2M
|
32 |
- Code: https://github.com/cognano/AVIDa-SARS-CoV-2
|
33 |
+
- Paper: https://arxiv.org/abs/2405.18749
|
34 |
|
35 |
## Citation
|
36 |
|
37 |
If you use VHHBERT in your research, please cite the following paper.
|
38 |
|
39 |
```bibtex
|
40 |
+
@article{tsuruta2024sars,
|
41 |
+
title={A {SARS}-{C}o{V}-2 Interaction Dataset and {VHH} Sequence Corpus for Antibody Language Models},
|
42 |
+
author={Hirofumi Tsuruta and Hiroyuki Yamazaki and Ryota Maeda and Ryotaro Tamura and Akihiro Imura},
|
43 |
+
journal={arXiv preprint arXiv:2405.18749},
|
44 |
+
year={2024}
|
45 |
+
}
|
46 |
```
|