Papers
arxiv:2401.12246

Orion-14B: Open-source Multilingual Large Language Models

Published on Jan 20
· Featured in Daily Papers on Jan 24
Authors:
,
,
,
,

Abstract

In this study, we introduce Orion-14B, a collection of multilingual large language models with 14 billion parameters. We utilize a data scheduling approach to train a foundational model on a diverse corpus of 2.5 trillion tokens, sourced from texts in English, Chinese, Japanese, Korean, and other languages. Additionally, we fine-tuned a series of models tailored for conversational applications and other specific use cases. Our evaluation results demonstrate that Orion-14B achieves state-of-the-art performance across a broad spectrum of tasks. We make the Orion-14B model family and its associated code publicly accessible https://github.com/OrionStarAI/Orion, aiming to inspire future research and practical applications in the field.

Community

Looks good, but the licence is misleading. GitHub says Apache, but read further and you'll see only the code is Apache. The models weights require a commercial licence. The license is also revocable.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

Sign up or log in to comment

Models citing this paper 1

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2401.12246 in a dataset README.md to link it from this page.

Spaces citing this paper 2

Collections including this paper 3