language: | |
- en | |
base_model: | |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-32B | |
- Extracted a 64 Rank Lora from DeepSeek-R1-Distill-Qwen-32B | |
- Merged & Quantized into Q4_K_M | |
### Note: The model seems to be somewhat working with the R1's weird template too but it repeats random Chinese characters and the quality seems to be consistently worse. | |
### Maybe try using the R1 tokenizer. |