These language model checkpoints are trained at the 360M and 1.3Bn parameter scales for up to 50Bn tokens on the Pile corpus, for research purposes.
HazyResearch
community
AI & ML interests
None defined yet.
Collections
3
models
19
hazyresearch/mamba-360M-30B
Updated
hazyresearch/based-360M-30B
Updated
•
1
hazyresearch/attn-360M-30B
Updated
hazyresearch/M2-BERT-8k-Retrieval-Encoder-V1
Fill-Mask
•
Updated
•
37
•
1
hazyresearch/M2-BERT-2k-Retrieval-Encoder-V1
Fill-Mask
•
Updated
•
27
•
1
hazyresearch/M2-BERT-32K-Retrieval-Encoder-V1
Fill-Mask
•
Updated
•
63
•
1
hazyresearch/M2-BERT-128-Retrieval-Encoder-V1
Fill-Mask
•
Updated
•
19
•
2
hazyresearch/based-1b-50b
Updated
•
167
•
1
hazyresearch/attn-1b-50bn
Updated
•
186
hazyresearch/mamba-360m
Updated
•
52
datasets
14
hazyresearch/based_triviaqa
Viewer
•
Updated
•
14
hazyresearch/based_drop
Viewer
•
Updated
•
10
hazyresearch/based_nq_2048
Viewer
•
Updated
•
10
hazyresearch/based_nq_512
Viewer
•
Updated
hazyresearch/based_nq_1024
Viewer
•
Updated
•
10
hazyresearch/based-squad
Viewer
•
Updated
•
2.13k
hazyresearch/based-swde
Viewer
•
Updated
•
192
•
2
hazyresearch/based-fda
Viewer
•
Updated
•
1.35k
•
3
hazyresearch/LoCoV1-Queries
Viewer
•
Updated
•
364
hazyresearch/LoCoV1-Documents
Viewer
•
Updated
•
395
•
1