File size: 1,143 Bytes
7b8c550 3e28fec 7b8c550 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 |
---
license: apache-2.0
language:
- zh
---
# Chinese MentalBERT, a pre-trained language model specifically designed for mental tasks.
In this study, we employ a domain-adaptive pretraining model, and introduce a novel lexicon guided masking machanism strategy based on the Chinese depression lexicon.
## How to use
```bash
from transformers import BertTokenizer, BertForMaskedLM
tokenizer = BertTokenizer.from_pretrained('zwzzz/Chinese-MentalBERT')
model = BertForMaskedLM.from_pretrained('zwzzz/Chinese-MentalBERT')
```
## Citation
If you find the technical report or resource is useful, please cite the following technical report in your paper.
Article address:[https://arxiv.org/pdf/2402.09151.pdf](https://arxiv.org/pdf/2402.09151.pdf)
```bash
@misc{zhai2024chinese,
title={Chinese MentalBERT: Domain-Adaptive Pre-training on Social Media for Chinese Mental Health Text Analysis},
author={Wei Zhai and Hongzhi Qi and Qing Zhao and Jianqiang Li and Ziqi Wang and Han Wang and Bing Xiang Yang and Guanghui Fu},
year={2024},
eprint={2402.09151},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` |