--- base_model: - sometimesanotion/Lamarck-14B-v0.6 - sthenno-com/miscii-14b-1225 - sometimesanotion/Qwentinuum-14B-v013 - Krystalan/DRT-o1-14B - sometimesanotion/Qwenvergence-14B-v3-Prose - arcee-ai/Virtuoso-Small - huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2 library_name: transformers tags: - mergekit - merge license: apache-2.0 language: - en metrics: - accuracy pipeline_tag: text-generation --- This isn't meant for end users. This merge is meant to expand the range of weights and scores possible to a standardized base, and may be showing signs of overfitting. I'm leaving it public because it's made new record scores for MUSR and MATH for 14B Qwen2.5 models, and that's worth studying. This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using sometimesanotion/Lamarck-14B-v0.7-Base-001 as a base. ### Models Merged The following models were included in the merge: * [sthenno-com/miscii-14b-1225](https://huggingface.co/sthenno-com/miscii-14b-1225) * [sometimesanotion/Qwentinuum-14B-v013](https://huggingface.co/sometimesanotion/Qwentinuum-14B-v013) * [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B) * [sometimesanotion/Qwenvergence-14B-v3-Prose](https://huggingface.co/sometimesanotion/Qwenvergence-14B-v3-Prose) * [arcee-ai/Virtuoso-Small](https://huggingface.co/arcee-ai/Virtuoso-Small) * [sometimesanotion/Lamarck-14B-v0.6](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6) * [sometimesanotion/Abliterate-Qwenvergence](https://huggingface.co/sometimesanotion/Abliterate-Qwenvergence) ### Configuration The following YAML configuration was used to produce this model (**note**: -qv suffixes indicate LoRA application from [Abliterate-Qwenvergence](https://huggingface.co/sometimesanotion/Abliterate-Qwenvergence) which is almost identical to [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2)) ```yaml name: Qwenvergence-14B-v9 merge_method: model_stock base_model: sometimesanotion/Lamarck-14B-v0.7-Base-001 tokenizer_source: sometimesanotion/Abliterate-Qwenvergence dtype: float32 out_dtype: bfloat16 parameters: int8_mask: true normalize: true rescale: false models: - model: sometimesanotion/Lamarck-14B-v0.6 - model: sometimesanotion/Qwenvergence-14B-v3-Prose-qv256 - model: Krystalan/DRT-o1-14B-qv128 - model: arcee-ai/Virtuoso-Small-qv64 - model: sometimesanotion/Lamarck-14B-v0.6 - model: sometimesanotion/Qwentinuum-14B-v013-qv512 - model: sthenno-com/miscii-14b-1225-qv64 - model: sometimesanotion/Qwenvergence-14B-v3-Prose-qv256 - model: sometimesanotion/Lamarck-14B-v0.6 ```