hfl-rc commited on
Commit
21b753b
•
1 Parent(s): 20a6d59

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -10,12 +10,15 @@ language:
10
  license: "apache-2.0"
11
  ---
12
 
 
13
 
14
- ### Chinese ELECTRA
 
 
15
 
16
- Google and Stanford University released a new pre-trained model called ELECTRA, which has a much compact model size and relatively competitive performance compared to BERT and its variants. For further accelerating the research of the Chinese pre-trained model, the Joint Laboratory of HIT and iFLYTEK Research (HFL) has released the Chinese ELECTRA models based on the official code of ELECTRA. ELECTRA-small could reach similar or even higher scores on several NLP tasks with only 1/10 parameters compared to BERT and its variants.
17
 
18
- This project is based on the official code of ELECTRA: https://github.com/google-research/electra
19
 
20
  You may also interested in,
21
 
 
10
  license: "apache-2.0"
11
  ---
12
 
13
+ ## CINO: Pre-trained Language Models for Chinese Minority
14
 
15
+ Multilingual Pre-trained Language Model, such as mBERT, XLM-R, provide multilingual and cross-lingual ability for language understanding.
16
+ We have seen rapid progress on building multilingual PLMs in recent year.
17
+ However, there is a lack of contributions on building PLMs on Chines minority languages, which hinders researchers from building powerful NLP systems.
18
 
19
+ To address the absence of Chinese minority PLMs, Joint Laboratory of HIT and iFLYTEK Research (HFL) proposes CINO (Chinese-miNOrity pre-trained language model), which is built on XLM-R with additional pre-training using Chinese minority corpus, such as Tibetan, Mongolian (Uighur form), Uyghur, Kazakh (Arabic form), Korean, Zhuang, Cantonese, etc.
20
 
21
+ Please read our GitHub repository for more details (Chinese): https://github.com/ymcui/Chinese-Minority-PLM
22
 
23
  You may also interested in,
24