sometimesanotion
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -14,11 +14,13 @@ license: apache-2.0
|
|
14 |
language:
|
15 |
- en
|
16 |
pipeline_tag: text-generation
|
|
|
|
|
17 |
---
|
18 |
-
![Lamarck.webp](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7/resolve/main/
|
19 |
---
|
20 |
|
21 |
-
> [!TIP] This
|
22 |
|
23 |
Lamarck 14B v0.7: A generalist merge focused on multi-step reasoning, prose, and multi-language ability. It is based on components that have punched above their weight in the 14 billion parameter class. It uses a custom toolchain to create and apply multiple sequences of complex merges:
|
24 |
|
|
|
14 |
language:
|
15 |
- en
|
16 |
pipeline_tag: text-generation
|
17 |
+
metrics:
|
18 |
+
- accuracy
|
19 |
---
|
20 |
+
![Lamarck.webp](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7/resolve/main/LamarckShades.webp)
|
21 |
---
|
22 |
|
23 |
+
> [!TIP] This version of the model has broken the 41.0 average barrier for 14B parameter models, and as of this writing, ranks #7 among models under 70B parameters. Given the respectable performance in the 32B range, I think Lamarck deserves his shades. A little layer analysis in the 14B range goes a long, long way.
|
24 |
|
25 |
Lamarck 14B v0.7: A generalist merge focused on multi-step reasoning, prose, and multi-language ability. It is based on components that have punched above their weight in the 14 billion parameter class. It uses a custom toolchain to create and apply multiple sequences of complex merges:
|
26 |
|