DachengZhang
commited on
Commit
•
008e8fd
1
Parent(s):
15a22b4
Update README.md
Browse files
README.md
CHANGED
@@ -49,7 +49,7 @@ pipeline_tag: text-generation
|
|
49 |
- Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
|
50 |
- Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
|
51 |
- The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
|
52 |
-
- The long-chat version supports extremely long texts,
|
53 |
- The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
|
54 |
<table style="border-collapse: collapse; width: 100%;">
|
55 |
<tr>
|
@@ -65,7 +65,7 @@ pipeline_tag: text-generation
|
|
65 |
- Orion-14B series models including:
|
66 |
- **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
|
67 |
- **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
|
68 |
-
- **Orion-14B-LongChat:**
|
69 |
- **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
|
70 |
- **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
|
71 |
- **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
|
|
|
49 |
- Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
|
50 |
- Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
|
51 |
- The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
|
52 |
+
- The long-chat version supports extremely long texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
|
53 |
- The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
|
54 |
<table style="border-collapse: collapse; width: 100%;">
|
55 |
<tr>
|
|
|
65 |
- Orion-14B series models including:
|
66 |
- **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
|
67 |
- **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
|
68 |
+
- **Orion-14B-LongChat:** The long-context version excels at handling extremely lengthy texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
|
69 |
- **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
|
70 |
- **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
|
71 |
- **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
|