Text Generation
Transformers
Safetensors
Chinese
English
qwen
conversational
custom_code
yuyijiong commited on
Commit
6fb372c
1 Parent(s): 8a80f7a

Upload 2 files

Browse files
Files changed (2) hide show
  1. README.md +1 -1
  2. README_en.md +1 -1
README.md CHANGED
@@ -57,7 +57,7 @@ Qwen-14b-chat-yarn-32k经过微调后,在多文档问答(或检索)任务
57
 
58
  # "大海捞针"测试结果
59
  ![](大海捞针50k.png)
60
- * 可以发现即使在50k长度下(即使训练样本不大于32k),检索信息的准确率依然极高,证明此模型确实拥有强大的长上下文能力,极大缓解“lost in the middle”问题,并且拥有极大的扩展潜力。
61
  * 而且此模型在推理时,不需要进行"原文复述",只需要给出问题,并让模型直接回答问题,模型就能给出正确的答案。(相对的,claude2.1-200k 需要特定的 prompt才能正确回答)这也证明了此模型的强大能力。
62
 
63
  <br>
 
57
 
58
  # "大海捞针"测试结果
59
  ![](大海捞针50k.png)
60
+ * 可以发现即使在**50k**长度下(即使训练样本不大于32k),检索信息的准确率依然极高,证明此模型确实拥有强大的长上下文能力,极大缓解“lost in the middle”问题,并且拥有极大的扩展潜力。
61
  * 而且此模型在推理时,不需要进行"原文复述",只需要给出问题,并让模型直接回答问题,模型就能给出正确的答案。(相对的,claude2.1-200k 需要特定的 prompt才能正确回答)这也证明了此模型的强大能力。
62
 
63
  <br>
README_en.md CHANGED
@@ -56,7 +56,7 @@ Qwen-14b-chat-yarn-32k has shown significant improvement in multi-document quest
56
 
57
  # Test Results for "Needle in a Haystack"
58
  ![](大海捞针50k.png)
59
- * The model can accurately retrieve the needle in a haystack even when the context length is 50k or longer, proving that the model does have strong long-context capabilities, which greatly alleviates the "lost in the middle" problem.
60
  * In addition, the model does not need to paraphrase the original text during inference, it only needs to give the question and let the model answer the question directly, and the model can give the correct answer. (In contrast, claude2.1-200k needs a specific prompt to answer correctly) This also proves the powerful ability of this model.
61
 
62
 
 
56
 
57
  # Test Results for "Needle in a Haystack"
58
  ![](大海捞针50k.png)
59
+ * The model can accurately retrieve the needle in a haystack even when the context length is **50k** or longer, proving that the model does have strong long-context capabilities, which greatly alleviates the "lost in the middle" problem.
60
  * In addition, the model does not need to paraphrase the original text during inference, it only needs to give the question and let the model answer the question directly, and the model can give the correct answer. (In contrast, claude2.1-200k needs a specific prompt to answer correctly) This also proves the powerful ability of this model.
61
 
62