I am a fan of @jpacifico models! 🔥
Maziyar Panahi PRO
MaziyarPanahi
AI & ML interests
Fine-Tuning, RLHF, Merging, Quantizations, Leaderboards
Recent Activity
replied to
sometimesanotion's
post
about 6 hours ago
I'd like to draw your attention to a Lamarck-based experiment which uses Arcee AI's newly published arcee_fusion merge method for three out of its four merges. Yes, just four. This is a simple one, and its recipe is fully open:
https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-Fusion
It unifies three branches, all of which feature models which bring Lamarck-14B-v0.7 and Qwenvergence-14B-v12-Prose together. One side features @jpacifico's http://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0.3 and the other features @suayptalha's http://huggingface.co/suayptalha/Lamarckvergence-14B paired with my models which were their merge ancestors.
A fusion merge - of a fusion merge and a SLERP of a fusion and older merge - should demonstrate the new merge method's behavior in interesting ways, especially in the first 1/4th of the model where the SLERP has less impact.
I welcome you to kick the tires and learn from it. It has prose quality near Qwenvergence v12's - as you'd expect.
Thank you, @mradermacher and @MaziyarPanahi, for the first-day quantizations! Your work helped get me started. https://huggingface.co/models?other=base_model:quantized:sometimesanotion/Lamarck-14B-v0.7-Fusion
replied to
sometimesanotion's
post
about 11 hours ago
I'd like to draw your attention to a Lamarck-based experiment which uses Arcee AI's newly published arcee_fusion merge method for three out of its four merges. Yes, just four. This is a simple one, and its recipe is fully open:
https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-Fusion
It unifies three branches, all of which feature models which bring Lamarck-14B-v0.7 and Qwenvergence-14B-v12-Prose together. One side features @jpacifico's http://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0.3 and the other features @suayptalha's http://huggingface.co/suayptalha/Lamarckvergence-14B paired with my models which were their merge ancestors.
A fusion merge - of a fusion merge and a SLERP of a fusion and older merge - should demonstrate the new merge method's behavior in interesting ways, especially in the first 1/4th of the model where the SLERP has less impact.
I welcome you to kick the tires and learn from it. It has prose quality near Qwenvergence v12's - as you'd expect.
Thank you, @mradermacher and @MaziyarPanahi, for the first-day quantizations! Your work helped get me started. https://huggingface.co/models?other=base_model:quantized:sometimesanotion/Lamarck-14B-v0.7-Fusion
reacted
to
sometimesanotion's
post
with 🔥
about 11 hours ago
I'd like to draw your attention to a Lamarck-based experiment which uses Arcee AI's newly published arcee_fusion merge method for three out of its four merges. Yes, just four. This is a simple one, and its recipe is fully open:
https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-Fusion
It unifies three branches, all of which feature models which bring Lamarck-14B-v0.7 and Qwenvergence-14B-v12-Prose together. One side features @jpacifico's http://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0.3 and the other features @suayptalha's http://huggingface.co/suayptalha/Lamarckvergence-14B paired with my models which were their merge ancestors.
A fusion merge - of a fusion merge and a SLERP of a fusion and older merge - should demonstrate the new merge method's behavior in interesting ways, especially in the first 1/4th of the model where the SLERP has less impact.
I welcome you to kick the tires and learn from it. It has prose quality near Qwenvergence v12's - as you'd expect.
Thank you, @mradermacher and @MaziyarPanahi, for the first-day quantizations! Your work helped get me started. https://huggingface.co/models?other=base_model:quantized:sometimesanotion/Lamarck-14B-v0.7-Fusion
Organizations
MaziyarPanahi's activity

replied to
sometimesanotion's
post
about 6 hours ago

replied to
sometimesanotion's
post
about 11 hours ago
Beautiful work! 🤩

reacted to
sometimesanotion's
post with 🔥➕🔥👍
about 11 hours ago
Post
2721
I'd like to draw your attention to a Lamarck-based experiment which uses Arcee AI's newly published arcee_fusion merge method for three out of its four merges. Yes, just four. This is a simple one, and its recipe is fully open:
sometimesanotion/Lamarck-14B-v0.7-Fusion
It unifies three branches, all of which feature models which bring Lamarck-14B-v0.7 and Qwenvergence-14B-v12-Prose together. One side features @jpacifico 's jpacifico/Chocolatine-2-14B-Instruct-v2.0.3 and the other features @suayptalha 's suayptalha/Lamarckvergence-14B paired with my models which were their merge ancestors.
A fusion merge - of a fusion merge and a SLERP of a fusion and older merge - should demonstrate the new merge method's behavior in interesting ways, especially in the first 1/4th of the model where the SLERP has less impact.
I welcome you to kick the tires and learn from it. It has prose quality near Qwenvergence v12's - as you'd expect.
Thank you, @mradermacher and @MaziyarPanahi , for the first-day quantizations! Your work helped get me started. https://huggingface.co/models?other=base_model:quantized:sometimesanotion/Lamarck-14B-v0.7-Fusion
sometimesanotion/Lamarck-14B-v0.7-Fusion
It unifies three branches, all of which feature models which bring Lamarck-14B-v0.7 and Qwenvergence-14B-v12-Prose together. One side features @jpacifico 's jpacifico/Chocolatine-2-14B-Instruct-v2.0.3 and the other features @suayptalha 's suayptalha/Lamarckvergence-14B paired with my models which were their merge ancestors.
A fusion merge - of a fusion merge and a SLERP of a fusion and older merge - should demonstrate the new merge method's behavior in interesting ways, especially in the first 1/4th of the model where the SLERP has less impact.
I welcome you to kick the tires and learn from it. It has prose quality near Qwenvergence v12's - as you'd expect.
Thank you, @mradermacher and @MaziyarPanahi , for the first-day quantizations! Your work helped get me started. https://huggingface.co/models?other=base_model:quantized:sometimesanotion/Lamarck-14B-v0.7-Fusion
Update tokenizer_config.json
1
#5 opened about 19 hours ago
by
Epiphany
Upload folder using huggingface_hub
2
#1 opened about 17 hours ago
by
MaziyarPanahi

Upload folder using huggingface_hub
2
#1 opened about 18 hours ago
by
MaziyarPanahi

Upload folder using huggingface_hub
2
#1 opened about 20 hours ago
by
MaziyarPanahi
