Mrw33554432
commited on
Commit
•
bf9d895
1
Parent(s):
9d04c92
Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,9 @@ datasets:
|
|
7 |
|
8 |
BitLinear-phi-1.5 is a model trained partially using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
|
9 |
|
10 |
-
Our BitLinear layer will only apply 1-bit quantization to the weight
|
|
|
|
|
11 |
|
12 |
The model structure is from [phi-1.5](https://huggingface.co/microsoft/phi-1_5), with all linear layers except lm_head replaced with our custom BitLinear layer.
|
13 |
|
|
|
7 |
|
8 |
BitLinear-phi-1.5 is a model trained partially using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
|
9 |
|
10 |
+
### Notice: Our BitLinear layer will only apply 1-bit quantization to the weight
|
11 |
+
### Other components (RMSnorm, activation quant) in the paper is discarded.
|
12 |
+
Idea behind: The major contribution in their paper is introduced a valid binary weight quantization, we don't want to mix it with other components to make it difficult to evaluate the major part.
|
13 |
|
14 |
The model structure is from [phi-1.5](https://huggingface.co/microsoft/phi-1_5), with all linear layers except lm_head replaced with our custom BitLinear layer.
|
15 |
|