alokabhishek
commited on
Commit
•
cdf78df
1
Parent(s):
4ea86da
Updated Readme
Browse files
README.md
CHANGED
@@ -26,14 +26,14 @@ This repo contains 8-bit quantized (using bitsandbytes) model of Meta's meta-lla
|
|
26 |
- Original model: [Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf)
|
27 |
|
28 |
|
29 |
-
### About
|
30 |
|
31 |
|
32 |
-
QLoRA: Efficient Finetuning of Quantized LLMs: [arXiv - QLoRA: Efficient Finetuning of Quantized LLMs](https://arxiv.org/abs/2305.14314)
|
33 |
|
34 |
-
Hugging Face Blog post on
|
35 |
|
36 |
-
bitsandbytes github repo: [bitsandbytes github repo](https://github.com/TimDettmers/bitsandbytes)
|
37 |
|
38 |
# How to Get Started with the Model
|
39 |
|
|
|
26 |
- Original model: [Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf)
|
27 |
|
28 |
|
29 |
+
### About 8 bit quantization using bitsandbytes
|
30 |
|
31 |
|
32 |
+
- QLoRA: Efficient Finetuning of Quantized LLMs: [arXiv - QLoRA: Efficient Finetuning of Quantized LLMs](https://arxiv.org/abs/2305.14314)
|
33 |
|
34 |
+
- Hugging Face Blog post on 8-bit quantization using bitsandbytes: [A Gentle Introduction to 8-bit Matrix Multiplication for transformers at scale using Hugging Face Transformers, Accelerate and bitsandbytes](https://huggingface.co/blog/hf-bitsandbytes-integration)
|
35 |
|
36 |
+
- bitsandbytes github repo: [bitsandbytes github repo](https://github.com/TimDettmers/bitsandbytes)
|
37 |
|
38 |
# How to Get Started with the Model
|
39 |
|