|
--- |
|
license: apache-2.0 |
|
tags: |
|
- merge |
|
- mergekit |
|
- lazymergekit |
|
- MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
--- |
|
|
|
# Breeze-13B-32k-Base-v1_0 |
|
|
|
Breeze-13B-32k-Base-v1_0 is a merge of the following models using [mergekit](https://github.com/cg123/mergekit): |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
* [MediaTek-Research/Breeze-7B-32k-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-32k-Base-v1_0) |
|
|
|
## 🧩 Configuration |
|
|
|
```yaml |
|
dtype: bfloat16 |
|
merge_method: linear |
|
slices: |
|
- sources: |
|
- layer_range: [0, 8] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [0, 8] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
- sources: |
|
- layer_range: [4, 12] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [4, 12] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
- sources: |
|
- layer_range: [8, 16] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [8, 16] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
- sources: |
|
- layer_range: [12, 20] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [12, 20] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
- sources: |
|
- layer_range: [16, 24] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [16, 24] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
- sources: |
|
- layer_range: [20, 28] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [20, 28] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
- sources: |
|
- layer_range: [24, 32] |
|
model: MediaTek-Research/Breeze-7B-32k-Base-v1_0 |
|
- layer_range: [24, 32] |
|
model: meta-llama/Meta-Llama-3-8B |
|
parameters: |
|
weight: 0 |
|
tokenizer_source: union |
|
|
|
``` |