alokabhishek commited on
Commit
cdf78df
1 Parent(s): 4ea86da

Updated Readme

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -26,14 +26,14 @@ This repo contains 8-bit quantized (using bitsandbytes) model of Meta's meta-lla
26
  - Original model: [Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf)
27
 
28
 
29
- ### About 4 bit quantization using bitsandbytes
30
 
31
 
32
- QLoRA: Efficient Finetuning of Quantized LLMs: [arXiv - QLoRA: Efficient Finetuning of Quantized LLMs](https://arxiv.org/abs/2305.14314)
33
 
34
- Hugging Face Blog post on 4-bit quantization using bitsandbytes: [Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA](https://huggingface.co/blog/4bit-transformers-bitsandbytes)
35
 
36
- bitsandbytes github repo: [bitsandbytes github repo](https://github.com/TimDettmers/bitsandbytes)
37
 
38
  # How to Get Started with the Model
39
 
 
26
  - Original model: [Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf)
27
 
28
 
29
+ ### About 8 bit quantization using bitsandbytes
30
 
31
 
32
+ - QLoRA: Efficient Finetuning of Quantized LLMs: [arXiv - QLoRA: Efficient Finetuning of Quantized LLMs](https://arxiv.org/abs/2305.14314)
33
 
34
+ - Hugging Face Blog post on 8-bit quantization using bitsandbytes: [A Gentle Introduction to 8-bit Matrix Multiplication for transformers at scale using Hugging Face Transformers, Accelerate and bitsandbytes](https://huggingface.co/blog/hf-bitsandbytes-integration)
35
 
36
+ - bitsandbytes github repo: [bitsandbytes github repo](https://github.com/TimDettmers/bitsandbytes)
37
 
38
  # How to Get Started with the Model
39