sometimesanotion commited on
Commit
2a3aba7
·
verified ·
1 Parent(s): e9db55f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -23,7 +23,7 @@ pipeline_tag: text-generation
23
  Lamarck 14B v0.7: A generalist merge focused on multi-step reasoning, prose, and multi-language ability. It is based on components that have punched above their weight in the 14 billion parameter class. It uses a custom toolchain to create and apply multiple sequences of complex merges:
24
 
25
  - **Extracted LoRA adapters from special-purpose merges**
26
- - **Custom base models and model_stocks of original models with LoRAs from from [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2) to minimize IFEVAL loss from model_stocks**
27
  - **Separate branches for aggressive breadcrumbs and conservative DELLA merges**
28
  - **Highly targeted weight/density gradients for every 2-4 layers**
29
  - **Finalization through SLERP merges recombining the separate branches as is most stable**
 
23
  Lamarck 14B v0.7: A generalist merge focused on multi-step reasoning, prose, and multi-language ability. It is based on components that have punched above their weight in the 14 billion parameter class. It uses a custom toolchain to create and apply multiple sequences of complex merges:
24
 
25
  - **Extracted LoRA adapters from special-purpose merges**
26
+ - **Custom base models and model_stocks of original models with LoRAs from from [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2) to minimize IFEVAL loss often seen in model_stock merges**
27
  - **Separate branches for aggressive breadcrumbs and conservative DELLA merges**
28
  - **Highly targeted weight/density gradients for every 2-4 layers**
29
  - **Finalization through SLERP merges recombining the separate branches as is most stable**