flam-kit

Building off of the success of my last merge, I identified some areas of weakness that I perceived in the resulting model and this model is an attempt to address them.

I created a wild model with a della merge (schonsense/Flamlama_70B_della), using models that all had strengths I was hoping to incorporate into my existing model SLERP. To create this new model (schonsense/flam-kit), I SLERP merged my previous model (schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP) as a base and gently brought in the wild flavors of the della merged model. The intent being to keep the instruct following and proper model function without having to stomp on it with excessive sampling parameters, while changing the voice and capabilities of the model. I believe this model to finally be a success, after a number of failures. I find (schonsense/flam-kit) to be superior in most respects to (schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP), requiring only the most modest of sampling parameters to function well.

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the SLERP merge method.

Models Merged

The following models were included in the merge:

  • schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP
  • schonsense/Flamlama_70B_della

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP
  - model: schonsense/Flamlama_70B_della
merge_method: slerp
base_model: schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP
dtype: bfloat16
parameters:
  t: [
    0, 0.0, 0.0, 0.055, 0.109, 0.127, 0.145, 0.164, 0.182, 0.2, 
    0.218, 0.236, 0.255, 0.273, 0.291, 0.309, 0.327, 0.345, 0.364, 0.382, 0.4, 
    0.418, 0.436, 0.455, 0.473, 0.491, 0.509, 0.527, 0.545, 0.564, 0.582, 0.6, 
    0.588, 0.576, 0.564, 0.552, 0.54, 0.527, 0.515, 0.503, 0.491, 0.479, 0.467, 
    0.455, 0.442, 0.43, 0.418, 0.406, 0.394, 0.382, 0.369, 0.357, 0.345, 0.333, 
    0.321, 0.309, 0.297, 0.285, 0.273, 0.26, 0.248, 0.236, 0.224, 0.212, 0.2, 
    0.188, 0.176, 0.164, 0.151, 0.139, 0.127, 0.115, 0.103, 0.091, 0.079, 0.067, 
    0.055, 0, 0, 0
]

Downloads last month
43
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for schonsense/flam-kit