sometimesanotion
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -27,11 +27,9 @@ Lamarck 14B v0.6: A generalist merge focused on multi-step reasoning, prose, an
|
|
27 |
|
28 |
![Lamarck.webp](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6/resolve/main/comparison.png)
|
29 |
|
30 |
-
Previous releases were based on a SLERP merge of model_stock+della branches focused on reasoning and prose. The prose branch got surprisingly good at reasoning, and the reasoning branch became a strong generalist in its own right. Some of you have already downloaded it as [sometimesanotion/Qwen2.5-14B-Vimarckoso-v3](https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3).
|
31 |
-
|
32 |
A notable contribution to the middle to upper layers of Lamarck v0.6 comes from [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B). It has a fascinating research paper: [DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought](https://huggingface.co/papers/2412.17498).
|
33 |
|
34 |
-
Lamarck 0.6
|
35 |
|
36 |
- **Extracted LoRA adapters from special-purpose merges**
|
37 |
- **Separate branches for breadcrumbs and DELLA merges**
|
|
|
27 |
|
28 |
![Lamarck.webp](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6/resolve/main/comparison.png)
|
29 |
|
|
|
|
|
30 |
A notable contribution to the middle to upper layers of Lamarck v0.6 comes from [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B). It has a fascinating research paper: [DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought](https://huggingface.co/papers/2412.17498).
|
31 |
|
32 |
+
Lamarck 0.6 uses a custom toolchain to create the merges which target specific layers:
|
33 |
|
34 |
- **Extracted LoRA adapters from special-purpose merges**
|
35 |
- **Separate branches for breadcrumbs and DELLA merges**
|