Update README.md
Browse files
README.md
CHANGED
@@ -7,14 +7,12 @@ language:
|
|
7 |
|
8 |
# Model Card for Breeze-7B-Instruct-v0.1
|
9 |
|
10 |
-
Breeze-7B is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1).
|
11 |
-
By additionally pretraining Mistral 7B with 250GB of Traditional Chinese content, Breeze is specifically intended for Traditional Chinese use.
|
12 |
|
13 |
[Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0.1) is the base model for the Breeze series.
|
14 |
It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
|
15 |
|
16 |
-
[Breeze-7B-Instruct](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v0.1) derives from the base model Breeze-7B-Base
|
17 |
-
undergone supervised fine-tuning with over 1 million instances, making the resulting model amenable to be used as-is for commonly seen tasks.
|
18 |
|
19 |
[Breeze-7B-Instruct-64k](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-64k-v0.1) is a slightly modified version of
|
20 |
Breeze-7B-Instruct to enable a 64k-token context length. Roughly speaking, that is equivalent to 88k Traditional Chinese characters.
|
|
|
7 |
|
8 |
# Model Card for Breeze-7B-Instruct-v0.1
|
9 |
|
10 |
+
Breeze-7B is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically intended for Traditional Chinese use.
|
|
|
11 |
|
12 |
[Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0.1) is the base model for the Breeze series.
|
13 |
It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
|
14 |
|
15 |
+
[Breeze-7B-Instruct](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v0.1) derives from the base model Breeze-7B-Base, making the resulting model amenable to be used as-is for commonly seen tasks.
|
|
|
16 |
|
17 |
[Breeze-7B-Instruct-64k](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-64k-v0.1) is a slightly modified version of
|
18 |
Breeze-7B-Instruct to enable a 64k-token context length. Roughly speaking, that is equivalent to 88k Traditional Chinese characters.
|