YC-Chen commited on
Commit
abb39c0
1 Parent(s): 5bc7700

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -0
README.md CHANGED
@@ -7,6 +7,30 @@ language:
7
 
8
  # Model Card for Breeze-7B-Instruct-v0.1
9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  Breeze-7B is a language model that builds upon the foundation of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically enhanced for Traditional Chinese.
11
 
12
  [Breeze-7B-Base-v0.1](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0.1) introduces an expanded vocabulary with additional 30,000 Traditional Chinese tokens and
 
7
 
8
  # Model Card for Breeze-7B-Instruct-v0.1
9
 
10
+ Breeze-7B is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1). By additionally pretraining Mistral 7B with 250GB of Traditional Chinese content, Breeze is specifically intended for Traditional Chinese use.
11
+
12
+ [Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0.1) is the base model for the Breeze series. It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
13
+
14
+ [Breeze-7B-Instruct](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v0.1) derives from the base model Breeze-7B-Base and has undergone supervised fine-tuning with over 1 million instances, making the resulting model amenable to be used as-is for commonly seen tasks.
15
+
16
+ [Breeze-7B-Instruct-64k](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-64k-v0.1) is a slightly modified version of Breeze-7B-Instruct to enable a 64k-token context length. Roughly speaking, that is equivalent to 88k Traditional Chinese characters.
17
+
18
+ The current release version of Breeze is v0.1.
19
+
20
+ Practicality-wise:
21
+
22
+ - Breeze expands the original vocabulary with additional 30,000 Traditional Chinese tokens. With the expanded vocabulary, everything else being equal, Breeze operates at twice the inference speed for Traditional Chinese to Mistral-7B and Llama 7B. [See Inference Performance.]
23
+ - Breeze-Instruct can be used as is for common tasks such as Q&A, RAG, multi-round chat, and summarization.
24
+ - In particular, Breeze-Instruct-64k can perform tasks at a document level, not a chapter level.
25
+
26
+ Performance-wise:
27
+
28
+ - Breeze demonstrates impressive performance in benchmarks for Traditional Chinese, when compared to similar sized open-source contemporaries such as Taiwan-LLM, QWen, and Yi. [See Chat Model Performance.]
29
+ - Breeze shows comparable results to Mistral-7B-Instruct-v0.1 on the MMLU and MT-Bench benchmarks. [See Chat Model Performance.]
30
+
31
+
32
+
33
+
34
  Breeze-7B is a language model that builds upon the foundation of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically enhanced for Traditional Chinese.
35
 
36
  [Breeze-7B-Base-v0.1](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0.1) introduces an expanded vocabulary with additional 30,000 Traditional Chinese tokens and