File size: 431 Bytes
13f2cc0
 
 
2a0703b
d1771e2
 
8582164
1
2
3
4
5
6
7
---
language:
- bo
---
T‑LLaMA: a Tibetan large language model based on LLaMA2

In this study, we built a corpus containing 2.2 billion Tibetan characters and trained Tibetan LLaMA based on LLaMA2 7B. We achieved state-of-the-art performance in the text classification task using the open-source TNCC dataset, with an accuracy of 79.8%. Additionally, we obtained promising results in text generation and text summarization tasks.