|
---
|
|
base_model:
|
|
- anthracite-org/magnum-v3-9b-customgemma2
|
|
- nbeerbower/gemma2-gutenberg-9B
|
|
- grimjim/Magnolia-v1-Gemma2-8k-9B
|
|
- UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
|
|
- BeaverLegacy/Smegmma-Deluxe-9B-v1
|
|
- ifable/gemma-2-Ifable-9B
|
|
library_name: transformers
|
|
tags:
|
|
- mergekit
|
|
- merge
|
|
|
|
---
|
|
# temp
|
|
|
|
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
|
|
|
## Merge Details
|
|
### Merge Method
|
|
|
|
This model was merged using the SLERP method to create an intermediate model. The [Model Stock](https://arxiv.org/abs/2403.19522) merge method used the SLERP model as a base to mix in more models.
|
|
|
|
The idea was to make a nice and smart base model and add in a few pinches of spice.
|
|
|
|
For some reason it wouldn't let me use any other merge method- it gave me ModelReference errors about my
|
|
intermediary model for every method except Model Stock for some reason. I'll see if I can fix it and
|
|
upload my intended task-arithmetic version as a v2.
|
|
|
|
This is the only one of my like 700 merges that I think uses something novel/interesting
|
|
enough in its creation to merit an upload.
|
|
|
|
### Models Merged
|
|
|
|
The following models were included in the merge:
|
|
* [anthracite-org/magnum-v3-9b-customgemma2](https://huggingface.co/anthracite-org/magnum-v3-9b-customgemma2)
|
|
* [nbeerbower/gemma2-gutenberg-9B](https://huggingface.co/nbeerbower/gemma2-gutenberg-9B)
|
|
|
|
* [grimjim/Magnolia-v1-Gemma2-8k-9B](https://huggingface.co/grimjim/Magnolia-v1-Gemma2-8k-9B)
|
|
* [UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3](https://huggingface.co/UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3)
|
|
* [BeaverLegacy/Smegmma-Deluxe-9B-v1](https://huggingface.co/BeaverLegacy/Smegmma-Deluxe-9B-v1)
|
|
* [ifable/gemma-2-Ifable-9B](https://huggingface.co/ifable/gemma-2-Ifable-9B)
|
|
|
|
### Configuration
|
|
|
|
The following YAML configuration was used to produce this model:
|
|
|
|
```yaml
|
|
# THIS YAML CONFIGURATION WAS USED TO CREATE THE INTERMEDIARY MODEL.
|
|
# slices:
|
|
# - sources:
|
|
# - model: anthracite-org/magnum-v3-9b-customgemma2
|
|
# layer_range: [0, 42]
|
|
# - model: nbeerbower/gemma2-gutenberg-9B
|
|
# layer_range: [0, 42]
|
|
# merge_method: slerp
|
|
# base_model: nbeerbower/gemma2-gutenberg-9B
|
|
# parameters:
|
|
# t:
|
|
# - filter: self_attn
|
|
# value: [0.2, 0.5, 0.4, 0.7, 1]
|
|
# - filter: mlp
|
|
# value: [1, 0.5, 0.3, 0.4, 0.2]
|
|
# - value: 0.5
|
|
# dtype: float16
|
|
|
|
# THIS YAML CONFIGURATION WAS USED TO CREATE ASTER. The E: model is the intermediate
|
|
# model created in the previous config.
|
|
models:
|
|
- model: E:/models/mergekit/output/intermediate/
|
|
- model: BeaverLegacy/Smegmma-Deluxe-9B-v1
|
|
parameters:
|
|
weight: 0.3
|
|
- model: ifable/gemma-2-Ifable-9B
|
|
parameters:
|
|
weight: 0.3
|
|
- model: UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
|
|
parameters:
|
|
weight: 0.15
|
|
- model: grimjim/Magnolia-v1-Gemma2-8k-9B
|
|
parameters:
|
|
weight: 0.25
|
|
merge_method: model_stock
|
|
base_model: E:/models/mergekit/output/intermediate/
|
|
dtype: float16
|
|
```
|
|
|
|
Alright, now back to smashing models together and seeing what happens... |