models: - model: ank028/Llama-3.2-1B-Instruct-gsm8k parameters: density: 0.5 # density gradient weight: 1.0 - model: autoprogrammer/Llama-3.2-1B-Instruct-MGSM8K-sft1 parameters: density: 0.5 weight: 0.5 # weight gradient merge_method: ties base_model: meta-llama/Llama-3.2-1B-Instruct parameters: normalize: true int8_mask: false dtype: float16 name: Llama-3.2-1B-Instruct-gsm8k-MGSM8K-sft1-ties