Mrw33554432 commited on
Commit
bf9d895
1 Parent(s): 9d04c92

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -7,7 +7,9 @@ datasets:
7
 
8
  BitLinear-phi-1.5 is a model trained partially using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
9
 
10
- Our BitLinear layer will only apply 1-bit quantization to the weight, all other computations in the paper is discarded.
 
 
11
 
12
  The model structure is from [phi-1.5](https://huggingface.co/microsoft/phi-1_5), with all linear layers except lm_head replaced with our custom BitLinear layer.
13
 
 
7
 
8
  BitLinear-phi-1.5 is a model trained partially using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
9
 
10
+ ### Notice: Our BitLinear layer will only apply 1-bit quantization to the weight
11
+ ### Other components (RMSnorm, activation quant) in the paper is discarded.
12
+ Idea behind: The major contribution in their paper is introduced a valid binary weight quantization, we don't want to mix it with other components to make it difficult to evaluate the major part.
13
 
14
  The model structure is from [phi-1.5](https://huggingface.co/microsoft/phi-1_5), with all linear layers except lm_head replaced with our custom BitLinear layer.
15