# sd-webui-model-converter Model convert extension , Used for [AUTOMATIC1111's stable diffusion webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) ![image](https://github.com/Akegarasu/sd-webui-model-converter/assets/36563862/3f160408-6816-4fb5-9b27-1217126f5a6e) ## Features - convert to precisions: fp32, fp16, bf16 - pruning model: no-ema, ema-only - checkpoint ext convert: ckpt, safetensors - convert/copy/delete any parts of model: unet, text encoder(clip), vae - Fix CLIP - Force CLIP position_id to int64 before convert ### Fix CLIP Sometimes, the CLIP position_id becomes incorrect due to model merging. For example, Anything-v3. This option will reset CLIP position to `torch.Tensor([list(range(77))]).to(torch.int64)` ### Force CLIP position_id to int64 before convert If you use this extension to convert a model to fp16, and the model has an incorrect CLIP, the precision of the CLIP position_id may decrease during the compression process, which might coincidentally correct the offset. ![image](https://github.com/Akegarasu/sd-webui-model-converter/assets/36563862/d057d530-4b00-4937-ab30-8b8bd50fbd93) If you do not wish to correct this CLIP offset coincidentally (because fixing it would alter the model, even though the correction is accurate, not everyone prefers the most correct, right? :P), you can use this option. It will force the CLIP position_id to be int64 and retain the incorrect CLIP