sophosympatheia's picture
Update README.md
fed9415 verified
|
raw
history blame
2.29 kB
metadata
base_model: []
library_name: transformers
tags:
  - mergekit
  - merge
MidnightMiqu

Overview

This is a 103B frankenmerge of sophosympatheia/Midnight-Miqu-70B-v1.0 with itself. Please see that model card for details and usage instructions. This model is based on Miqu so it's capable of 32K context.

Quantizations

  • Pending
  • If you don't see something you're looking for, try searching Hugging Face. There may be newer quants available than what I've documented here.

Licence and usage restrictions

152334H/miqu-1-70b-sf was based on a leaked version of one of Mistral's models. All miqu-derived models, including this merge, are only suitable for personal use. Mistral has been cool about it so far, but you should be aware that by downloading this merge you are assuming whatever legal risk is iherent in acquiring and using a model based on leaked weights. This merge comes with no warranties or guarantees of any kind, but you probably already knew that. I am not a lawyer and I do not profess to know what we have gotten ourselves into here. You should consult with a lawyer before using any Hugging Face model beyond private use... but definitely don't use this one for that!

Merge Details

Merge Method

This model was merged using the passthrough merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

slices:
  - sources:
      - model: /home/llm/mergequant/models/midnight-miqu-70b
        layer_range: [0, 40] # 40
  - sources:
      - model: /home/llm/mergequant/models/midnight-miqu-70b
        layer_range: [20, 60] # 40
  - sources:
      - model: /home/llm/mergequant/models/midnight-miqu-70b
        layer_range: [40, 80] # 40
merge_method: passthrough
dtype: float16