--- license: apache-2.0 language: - zh --- # Chinese MentalBERT, a pre-trained language model specifically designed for mental tasks. In this study, we employ a domain-adaptive pretraining model, and introduce a novel lexicon guided masking machanism strategy based on the Chinese depression lexicon. ## How to use ```bash from transformers import BertTokenizer, BertForMaskedLM tokenizer = BertTokenizer.from_pretrained('zwzzz/Chinese-MentalBERT') model = BertForMaskedLM.from_pretrained('zwzzz/Chinese-MentalBERT') ``` ## Citation If you find the technical report or resource is useful, please cite the following technical report in your paper. Article address:[https://arxiv.org/pdf/2402.09151.pdf](https://arxiv.org/pdf/2402.09151.pdf) ```bash @misc{zhai2024chinese, title={Chinese MentalBERT: Domain-Adaptive Pre-training on Social Media for Chinese Mental Health Text Analysis}, author={Wei Zhai and Hongzhi Qi and Qing Zhao and Jianqiang Li and Ziqi Wang and Han Wang and Bing Xiang Yang and Guanghui Fu}, year={2024}, eprint={2402.09151}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```