Swallow-MoE-4x7B-lisa / mergekit_moe_config.yml
Aratako's picture
Upload mergekit_moe_config.yml
f810a80 verified
raw
history blame
425 Bytes
base_model: tokyotech-llm/Swallow-7b-hf
gate_mode: random
dtype: bfloat16
experts:
- source_model: nitky/Superswallow-7b-v0.3
positive_prompts: []
- source_model: nitky/Superswallow-7b-v0.2
positive_prompts: []
- source_model: nitky/Superswallow-7b-v0.1
positive_prompts: []
- source_model: tokyotech-llm/Swallow-7b-instruct-hf
positive_prompts: []
tokenizer_source: model:tokyotech-llm/Swallow-7b-hf