DachengZhang commited on
Commit
008e8fd
1 Parent(s): 15a22b4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -49,7 +49,7 @@ pipeline_tag: text-generation
49
  - Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
50
  - Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
51
  - The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
52
- - The long-chat version supports extremely long texts, extending up to 200K tokens.
53
  - The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
54
  <table style="border-collapse: collapse; width: 100%;">
55
  <tr>
@@ -65,7 +65,7 @@ pipeline_tag: text-generation
65
  - Orion-14B series models including:
66
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
67
  - **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
68
- - **Orion-14B-LongChat:** This model is optimized for long context lengths more than 200k tokens and demonstrates performance comparable to proprietary models on long context evaluation sets.
69
  - **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
70
  - **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
71
  - **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
 
49
  - Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
50
  - Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
51
  - The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
52
+ - The long-chat version supports extremely long texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
53
  - The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
54
  <table style="border-collapse: collapse; width: 100%;">
55
  <tr>
 
65
  - Orion-14B series models including:
66
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
67
  - **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
68
+ - **Orion-14B-LongChat:** The long-context version excels at handling extremely lengthy texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
69
  - **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
70
  - **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
71
  - **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.