-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 564 -
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 172 -
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
Paper • 2402.19427 • Published • 48 -
ResLoRA: Identity Residual Mapping in Low-Rank Adaption
Paper • 2402.18039 • Published • 10
RachidAR
RachidAR
AI & ML interests
1.58 bit LLM
Organizations
Collections
1
spaces
1
models
21
RachidAR/Llama-3-8B-Instruct-DPO-v0.3-Q6_K-GGUF
Text Generation
•
Updated
•
62
RachidAR/Waktaverse-Llama-3-KO-8B-Instruct-Q6_K-GGUF
Updated
•
17
RachidAR/llama-3-indotuned-v0-Q6_K-GGUF
Updated
•
15
RachidAR/saiga_llama3_8b-Q6_K-GGUF
Updated
•
27
RachidAR/Llama-3-8B-saiga-suzume-ties-Q6_K-GGUF
Text Generation
•
Updated
•
54
•
1
RachidAR/wiz-llama3-8B-Q6_K-GGUF
Updated
•
47
RachidAR/ablation-model-fineweb-v1-Q6_K-GGUF
Updated
•
157
RachidAR/Llama-3-8B-Instruct-Physics-5k-Scar-Q6_K-GGUF
Updated
•
23
RachidAR/NorskGPT-Llama3-8b-Q6_K-GGUF
Updated
•
41
RachidAR/llama3-Mirage-Walker-8b-v0.2-slerp-Q6_K-GGUF
Updated
•
24
datasets
None public yet