File size: 382 Bytes
7f90a80 98022bc d880956 41c1b9d |
1 2 3 4 5 6 7 8 9 10 11 12 |
---
language:
- en
base_model:
- deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
---
- Extracted a 64 Rank Lora from DeepSeek-R1-Distill-Qwen-32B
- Merged & Quantized into Q4_K_M
### Note: The model seems to be somewhat working with the R1's weird template too but it repeats random Chinese characters and the quality seems to be consistently worse.
### Maybe try using the R1 tokenizer. |