OrionStarAI/OrionStar-Yi-34B-Chat-Llama

This model is identical to OrionStarAI/OrionStar-Yi-34B with the only difference being that the tensors have been renamed to follow the LLaMA format for automatic evaluation on the HF leaderboard.

Model Introduction

  • OrionStar-Yi-34B-Chat from OrionStarAI is based on the open-source Yi-34B model, fine-tuned on a high-quality corpus of over 15 million sentences. OrionStar-Yi-34B-Chat aims to provide an excellent interactive experience for users in the large model community.

  • The Yi series models, open-sourced by the 01-ai team, have shown impressive performance on various benchmarks in Chinese, English, and general domains. OrionStar-Yi-34B-Chat further explores the potential of Yi-34B. Through extensive fine-tuning on a large and high-quality corpus, OrionStar-Yi-34B-Chat performs exceptionally well on evaluation data. We strive to make it an outstanding open-source alternative in the ChatGPT domain!

  • Our fine-tuned model is completely open for academic research, but please adhere to the agreement and the Yi License.

  • Model Evaluation Results

We use opencompass to perform 5-shot on the following general domain datasets Testing. The evaluation results of other models are taken from opencompass leaderboard.

C-Eval MMLU CMMLU
GPT-4 69.9 83 71
ChatGPT 52.5 69.1 53.9
Claude-1 52 65.7 -
TigerBot-70B-Chat-V2 57.7 65.9 59.9
WeMix-LLaMA2-70B 55.2 71.3 56
LLaMA-2-70B-Chat 44.3 63.8 43.3
Qwen-14B-Chat 71.7 66.4 70
Baichuan2-13B-Chat 56.7 57 58.4
OrionStar-Yi-34B-Chat 77.71 78.32 73.52

Discord Link: https://discord.gg/zumjDWgdAs

Downloads last month
1,257
Safetensors
Model size
34.4B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for OrionStarAI/OrionStar-Yi-34B-Chat-Llama

Quantizations
6 models