sometimesanotion
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ pipeline_tag: text-generation
|
|
23 |
Lamarck 14B v0.7: A generalist merge focused on multi-step reasoning, prose, and multi-language ability. It is based on components that have punched above their weight in the 14 billion parameter class. It uses a custom toolchain to create and apply multiple sequences of complex merges:
|
24 |
|
25 |
- **Extracted LoRA adapters from special-purpose merges**
|
26 |
-
- **Custom base models and model_stocks of original models with LoRAs from from [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2) to minimize IFEVAL loss
|
27 |
- **Separate branches for aggressive breadcrumbs and conservative DELLA merges**
|
28 |
- **Highly targeted weight/density gradients for every 2-4 layers**
|
29 |
- **Finalization through SLERP merges recombining the separate branches as is most stable**
|
|
|
23 |
Lamarck 14B v0.7: A generalist merge focused on multi-step reasoning, prose, and multi-language ability. It is based on components that have punched above their weight in the 14 billion parameter class. It uses a custom toolchain to create and apply multiple sequences of complex merges:
|
24 |
|
25 |
- **Extracted LoRA adapters from special-purpose merges**
|
26 |
+
- **Custom base models and model_stocks of original models with LoRAs from from [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2) to minimize IFEVAL loss often seen in model_stock merges**
|
27 |
- **Separate branches for aggressive breadcrumbs and conservative DELLA merges**
|
28 |
- **Highly targeted weight/density gradients for every 2-4 layers**
|
29 |
- **Finalization through SLERP merges recombining the separate branches as is most stable**
|