A continually pre-trained model based on Llama-2-7b-hf.
We use the Traditional Mongolian texts in MC^2 and English texts in RedPajama with a proportion of 4:1 for training.
Base model