Base model: (Falcon3-10B-deepseekv3-distill)[Virtuoso_Lite]
Quants: IQ4 GGUF Here 4bpw exl2 Here
ST Presets [Updated] Here
Prompt format: ChatML
<|im_start|>system
{system_prompt}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
Models Merged:
The following YAML configuration was used to produce this model:
slices:
- sources:
- model: Nitral-Archive/Virtuoso-Lite-chatmlified-10B_r16-ep1
layer_range: [0, 40]
- model: Nitral-Archive/NightWing3-10B-v0.1
layer_range: [0, 40]
merge_method: slerp
base_model: Nitral-Archive/Virtuoso-Lite-chatmlified-10B_r16-ep1
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.420
dtype: bfloat16
Notes: The goal of this merge was to make use of both the falcon3-10B base model I trained earlier (nightwing3) and my more recent training run over Arcee's distillation of DeepSeekV3, which also falcon3-10B as a base (Virtuoso-Lite-chatmlified-10B_r16-ep1). Initially, I wasn't entirely satisfied with the results of either model on their own. However, with limited testing, this merged version appears to have smoothed out some of the rough edges present in the originals. Further evaluation is needed to fully assess its performance.
- Downloads last month
- 0
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for Nitral-AI/NightWing3_Virtuoso-10B-v0.2
Merge model
this model