chen-yingfa commited on
Commit
81a6e00
·
verified ·
1 Parent(s): 1e81bfd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -3
README.md CHANGED
@@ -1,3 +1,22 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - HuggingFaceFW/fineweb-edu
5
+ language:
6
+ - en
7
+ base_model:
8
+ - Qwen/Qwen3-1.7B
9
+ tags:
10
+ - linear-attention
11
+ - hybrid
12
+ - rnn
13
+ - distillation
14
+ ---
15
+
16
+ Links:
17
+
18
+ - GitHub repo: <https://github.com/thunlp/hybrid-linear-attention>
19
+ - Paper: <https://arxiv.org/abs/2601.22156>
20
+
21
+ This is the final HypeNet-2B checkpoint from the paper [Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts](https://arxiv.org/pdf/2601.22156), distilled from Qwen3-1.7B using the HALO pipeline proposed in our paper. For more information, please refer to our GitHub repo.
22
+