--- base_model: - Sao10K/MN-12B-Lyra-v1 - aetherwiing/MN-12B-Starcannon-v3 - BeaverAI/mistral-doryV2-12b - cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the linear [DARE](https://arxiv.org/abs/2311.03099) merge method using [aetherwiing/MN-12B-Starcannon-v3](https://huggingface.co/aetherwiing/MN-12B-Starcannon-v3) as a base. ### Models Merged The following models were included in the merge: * [Sao10K/MN-12B-Lyra-v1](https://huggingface.co/Sao10K/MN-12B-Lyra-v1) * [BeaverAI/mistral-doryV2-12b](https://huggingface.co/BeaverAI/mistral-doryV2-12b) * [cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b](https://huggingface.co/cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b parameters: weight: 0.25 density: 0.3 - model: BeaverAI/mistral-doryV2-12b parameters: weight: 0.25 density: 0.3 - model: aetherwiing/MN-12B-Starcannon-v3 parameters: weight: 0.25 density: 0.6 - model: Sao10K/MN-12B-Lyra-v1 parameters: weight: 0.25 density: 0.4 merge_method: dare_linear base_model: aetherwiing/MN-12B-Starcannon-v3 dtype: bfloat16 ```