Upload folder using huggingface_hub
Browse files- .gitattributes +13 -1
- README.md +41 -0
- gemma-2b.Q2_K.gguf +3 -0
- gemma-2b.Q3_K_L.gguf +3 -0
- gemma-2b.Q3_K_M.gguf +3 -0
- gemma-2b.Q3_K_S.gguf +3 -0
- gemma-2b.Q4_0.gguf +3 -0
- gemma-2b.Q4_1.gguf +3 -0
- gemma-2b.Q4_K_M.gguf +3 -0
- gemma-2b.Q4_K_S.gguf +3 -0
- gemma-2b.Q5_1.gguf +3 -0
- gemma-2b.Q5_K_M.gguf +3 -0
- gemma-2b.Q5_K_S.gguf +3 -0
- gemma-2b.Q6_K.gguf +3 -0
- gemma-2b.Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -33,5 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
-
gemma-2b
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
gemma-2b.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
gemma-2b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
gemma-2b.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
gemma-2b.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
gemma-2b.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
gemma-2b.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
gemma-2b.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
gemma-2b.Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
gemma-2b.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
gemma-2b.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
gemma-2b.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
gemma-2b.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
gemma-2b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
gemma-2b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
49 |
+
gemma-2b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
tags: []
|
4 |
+
extra_gated_heading: "Access Gemma on Hugging Face"
|
5 |
+
extra_gated_prompt: "To access Gemma on Hugging Face, you’re required to review and agree to Google’s usage license. To do this, please ensure you’re logged-in to Hugging Face and click below. Requests are processed immediately."
|
6 |
+
extra_gated_button_content: "Acknowledge license"
|
7 |
+
license: other
|
8 |
+
license_name: gemma-terms-of-use
|
9 |
+
license_link: https://ai.google.dev/gemma/terms
|
10 |
+
---
|
11 |
+
|
12 |
+
# Gemma-2B GGUF
|
13 |
+
|
14 |
+
This is a quantized version of the [google/gemma-2b](https://huggingface.co/google/gemma-2b) model using [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
15 |
+
|
16 |
+
This model card corresponds to the 2B base version of the Gemma model. You can also visit the model card of the [7B base model](https://huggingface.co/google/gemma-7b), [7B instruct model](https://huggingface.co/google/gemma-7b-it), and [2B instruct model](https://huggingface.co/google/gemma-2b-it).
|
17 |
+
|
18 |
+
**Model Page**: [Gemma](https://ai.google.dev/gemma/docs)
|
19 |
+
|
20 |
+
**Terms of Use**: [Terms](https://www.kaggle.com/models/google/gemma/license/consent)
|
21 |
+
|
22 |
+
## ⚡ Quants
|
23 |
+
|
24 |
+
* `q2_k`: Uses Q4_K for the attention.vw and feed_forward.w2 tensors, Q2_K for the other tensors.
|
25 |
+
* `q3_k_l`: Uses Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K
|
26 |
+
* `q3_k_m`: Uses Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K
|
27 |
+
* `q3_k_s`: Uses Q3_K for all tensors
|
28 |
+
* `q4_0`: Original quant method, 4-bit.
|
29 |
+
* `q4_1`: Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.
|
30 |
+
* `q4_k_m`: Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q4_K
|
31 |
+
* `q4_k_s`: Uses Q4_K for all tensors
|
32 |
+
* `q5_0`: Higher accuracy, higher resource usage and slower inference.
|
33 |
+
* `q5_1`: Even higher accuracy, resource usage and slower inference.
|
34 |
+
* `q5_k_m`: Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q5_K
|
35 |
+
* `q5_k_s`: Uses Q5_K for all tensors
|
36 |
+
* `q6_k`: Uses Q8_K for all tensors
|
37 |
+
* `q8_0`: Almost indistinguishable from float16. High resource use and slow. Not recommended for most users.
|
38 |
+
|
39 |
+
## 💻 Usage
|
40 |
+
|
41 |
+
This model can be used with the latest version of llama.cpp and LM Studio >0.2.16.
|
gemma-2b.Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e987bf77ffff2d873e5b06310dfe8847f03178cdd27812c84f95d8c5c488c59
|
3 |
+
size 899875936
|
gemma-2b.Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f2a686bdf24e52055ee55b323ddd2a077d99bdb370db1d114d96415227e67a2
|
3 |
+
size 1260790880
|
gemma-2b.Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3711a880951eba56dc201e0160154cff115f9a96cf0b56a8041d7136daa55f2
|
3 |
+
size 1179001952
|
gemma-2b.Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a5ccf460c7491354923e8be50978b962e3c8e742df0d7e9466578dccf95bfdbd
|
3 |
+
size 1083180128
|
gemma-2b.Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4acaf847c0c4b51888f4ba9362e031216a07629c1cbf30d5db89d460719e810e
|
3 |
+
size 1416021088
|
gemma-2b.Q4_1.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:16a58d042d6eda5436f47c4ca125cba401c7b5ec67ec4194bbbfaa473fa76295
|
3 |
+
size 1572652128
|
gemma-2b.Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37d50c21ef7847926204ad9b3007127d9a2722188cfd240ce7f9f7f041aa71a5
|
3 |
+
size 1495094368
|
gemma-2b.Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:588b26a208ef99d7cefa7137ef074257e51250c7096cecad51349fdc941647ac
|
3 |
+
size 1424671840
|
gemma-2b.Q5_1.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97498ad926467c426d2ee87da99b54afd840e5dec6b65fa3b1c9de2f7f5d6b5a
|
3 |
+
size 1885914208
|
gemma-2b.Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:226f65d1a2463a35baad4c94af0d5deb4df29ea2a15c208de00be80989f508eb
|
3 |
+
size 1770017888
|
gemma-2b.Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee20c0c2046d06ab9c4447fa3abaa880bfbaa1816de483ded5f1a060edfbf9be
|
3 |
+
size 1729283168
|
gemma-2b.Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29f42096801c3c6b1537e60f9416abfd103d5c4a37312ac437170485d61a4375
|
3 |
+
size 2062124128
|
gemma-2b.Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0ef61ca7b06ea0afba21ee65922e066f7ae8e1a35bb151cf41846251bfb1a5a8
|
3 |
+
size 2669069408
|