Update README.md
Browse files
README.md
CHANGED
@@ -45,16 +45,7 @@ pipeline_tag: text-generation
|
|
45 |
|
46 |
# 1. Model Introduction
|
47 |
|
48 |
-
- Orion-14B-Chat is
|
49 |
-
|
50 |
-
- The 850,000 fine-tuning corpus comprises two parts: approximately 220,000 manually curated high-quality datasets and 630,000 entries selected and semantically deduplicated from open-source data through model filtering. Among these, the Japanese and Korean data, totaling 70,000 entries, have only undergone basic cleaning and deduplication.
|
51 |
-
|
52 |
-
- The Orion-14B series models exhibit the following features:
|
53 |
-
- Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
|
54 |
-
- Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
|
55 |
-
- The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
|
56 |
-
- The long-chat version supports extremely long texts, extending up to 200K tokens.
|
57 |
-
- The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
|
58 |
<div align="center">
|
59 |
<img src="./assets/imgs/model_cap_en.png" alt="model_cap_en" width="50%" />
|
60 |
</div>
|
|
|
45 |
|
46 |
# 1. Model Introduction
|
47 |
|
48 |
+
- Orion-14B-Chat-Int4 is quantized using awq from Orion-14B-Chat while reducing model size by 70% and improving inference speed by 30%, with performance loss less than 1%.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
<div align="center">
|
50 |
<img src="./assets/imgs/model_cap_en.png" alt="model_cap_en" width="50%" />
|
51 |
</div>
|