pochunhsu commited on
Commit
b336863
1 Parent(s): 3e9f922

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -6,9 +6,9 @@ language:
6
  - en
7
  ---
8
 
9
- # Model Card for Breeze-7B-Base-v0_1
10
 
11
- Breeze-7B is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically intended for Traditional Chinese use.
12
 
13
  [Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0_1) is the base model for the Breeze-7B series.
14
  It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
@@ -20,7 +20,7 @@ Breeze-7B-Instruct to enable a 64k-token context length. Roughly speaking, that
20
 
21
  *Update (Feb. 21st, 2024): Breeze-7B-Instruct-64k-v0_1 has been temporarily removed from Hugging Face due to its actual performance in long context tests not meeting expectations.*
22
 
23
- The current release version of Breeze-7B is v0.1.
24
 
25
  Practicality-wise:
26
  - Breeze-7B-Base expands the original vocabulary with additional 30,000 Traditional Chinese tokens. With the expanded vocabulary, everything else being equal, Breeze-7B operates at twice the inference speed for Traditional Chinese to Mistral-7B and Llama 7B. [See [Inference Performance](#inference-performance).]
 
6
  - en
7
  ---
8
 
9
+ # Model Card for MediaTek Research Breeze-7B-Base-v0_1
10
 
11
+ MediaTek Research Breeze-7B (hereinafter referred to as Breeze-7B) is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically intended for Traditional Chinese use.
12
 
13
  [Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0_1) is the base model for the Breeze-7B series.
14
  It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
 
20
 
21
  *Update (Feb. 21st, 2024): Breeze-7B-Instruct-64k-v0_1 has been temporarily removed from Hugging Face due to its actual performance in long context tests not meeting expectations.*
22
 
23
+ *Update (Mar. 7th, 2024): The current release version of Breeze-7B is v1.0. See [Breeze-7B-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v1_0).*
24
 
25
  Practicality-wise:
26
  - Breeze-7B-Base expands the original vocabulary with additional 30,000 Traditional Chinese tokens. With the expanded vocabulary, everything else being equal, Breeze-7B operates at twice the inference speed for Traditional Chinese to Mistral-7B and Llama 7B. [See [Inference Performance](#inference-performance).]