Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,22 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
datasets:
|
| 4 |
+
- HuggingFaceFW/fineweb-edu
|
| 5 |
+
language:
|
| 6 |
+
- en
|
| 7 |
+
base_model:
|
| 8 |
+
- Qwen/Qwen3-1.7B
|
| 9 |
+
tags:
|
| 10 |
+
- linear-attention
|
| 11 |
+
- hybrid
|
| 12 |
+
- rnn
|
| 13 |
+
- distillation
|
| 14 |
+
---
|
| 15 |
+
|
| 16 |
+
Links:
|
| 17 |
+
|
| 18 |
+
- GitHub repo: <https://github.com/thunlp/hybrid-linear-attention>
|
| 19 |
+
- Paper: <https://arxiv.org/abs/2601.22156>
|
| 20 |
+
|
| 21 |
+
This is the final HypeNet-2B checkpoint from the paper [Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts](https://arxiv.org/pdf/2601.22156), distilled from Qwen3-1.7B using the HALO pipeline proposed in our paper. For more information, please refer to our GitHub repo.
|
| 22 |
+
|