File size: 1,199 Bytes
468b171
 
d3edbf8
468b171
d3edbf8
 
 
468b171
d3edbf8
 
 
 
 
 
 
468b171
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
E60/EQ60 means embeddings in ggml_type q6_0, thus only compatible with IK_Llama.CPP and Croco.CPP until this q6_0 quant eventually reaches mainline.
E50/EQ50 or no mention means embeddings in mainline ggml_types (iq4_xs or Q5_0, usually)

For the IQ3 145L quantized model, the peformances close to IQ4_XS, with a few gigabites torn down.
Best suited for 24-24-16GB GPU configs at 24-32k context with KV cache q6_0/q5_0
Or for more, of course.

For the IQ3_XSSL quantized model, the performances are probably akin to IQ3_XS.

These quants are made for my own use, and I decided to share them.
Nothing special about them, except that they suit my needs.

Basically, my quant strategies obey to a few rules diverging from mainline.
- I often dump attn_q by one degree of quant, like mainline does for iq3_xxs in iq2_s, as well as attn_output.
- I often up attn_k and attn_v by one degree of quant, for example. Mainline usually neglects too much those tensors in the GQA era.
- I bump the embeddings, because they do not offload on the GPU (aside for Bitner and Maybe Gemma).
- I sometimes bump a whole FFN_down by one degree of quant, or down some layers of FFN_up and FFN_gate by one degree of quant.