John6666 commited on
Commit
ab41240
·
verified ·
1 Parent(s): 6fd8887

Upload 4 files

Browse files
Files changed (4) hide show
  1. app.py +29 -0
  2. dc.py +26 -17
  3. env.py +22 -0
  4. llmdolphin.py +25 -0
app.py CHANGED
@@ -384,6 +384,35 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
384
  copy_prompt_btn.click(gradio_copy_prompt, inputs=[output_text], outputs=[prompt], show_api=False)
385
  copy_prompt_btn_pony.click(gradio_copy_prompt, inputs=[output_text_pony], outputs=[prompt], show_api=False)
386
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
387
  gr.LoginButton()
388
  gr.DuplicateButton(value="Duplicate Space for private use (This demo does not work on CPU. Requires GPU Space)")
389
 
 
384
  copy_prompt_btn.click(gradio_copy_prompt, inputs=[output_text], outputs=[prompt], show_api=False)
385
  copy_prompt_btn_pony.click(gradio_copy_prompt, inputs=[output_text_pony], outputs=[prompt], show_api=False)
386
 
387
+ with gr.Tab("PNG Info"):
388
+ def extract_exif_data(image):
389
+ if image is None: return ""
390
+
391
+ try:
392
+ metadata_keys = ['parameters', 'metadata', 'prompt', 'Comment']
393
+
394
+ for key in metadata_keys:
395
+ if key in image.info:
396
+ return image.info[key]
397
+
398
+ return str(image.info)
399
+
400
+ except Exception as e:
401
+ return f"Error extracting metadata: {str(e)}"
402
+
403
+ with gr.Row():
404
+ with gr.Column():
405
+ image_metadata = gr.Image(label="Image with metadata", type="pil", sources=["upload"])
406
+
407
+ with gr.Column():
408
+ result_metadata = gr.Textbox(label="Metadata", show_label=True, show_copy_button=True, interactive=False, container=True, max_lines=99)
409
+
410
+ image_metadata.change(
411
+ fn=extract_exif_data,
412
+ inputs=[image_metadata],
413
+ outputs=[result_metadata],
414
+ )
415
+
416
  gr.LoginButton()
417
  gr.DuplicateButton(value="Duplicate Space for private use (This demo does not work on CPU. Requires GPU Space)")
418
 
dc.py CHANGED
@@ -23,6 +23,9 @@ from stablepy import (
23
  )
24
  import time
25
  #import urllib.parse
 
 
 
26
  import gradio as gr
27
  import logging
28
  logging.getLogger("diffusers").setLevel(logging.ERROR)
@@ -52,14 +55,14 @@ PREPROCESSOR_CONTROLNET = {
52
  ],
53
  "scribble": [
54
  "HED",
55
- "Pidinet",
56
  "None",
57
  ],
58
  "softedge": [
59
- "Pidinet",
60
  "HED",
61
  "HED safe",
62
- "Pidinet safe",
63
  "None",
64
  ],
65
  "segmentation": [
@@ -295,18 +298,21 @@ def extract_parameters(input_string):
295
  input_string = input_string.replace("\n", "")
296
 
297
  if "Negative prompt:" not in input_string:
298
- print("Negative prompt not detected")
299
- parameters["prompt"] = input_string
300
- return parameters
 
 
 
301
 
302
  parm = input_string.split("Negative prompt:")
303
- parameters["prompt"] = parm[0]
304
  if "Steps:" not in parm[1]:
305
  print("Steps not detected")
306
- parameters["neg_prompt"] = parm[1]
307
  return parameters
308
  parm = parm[1].split("Steps:")
309
- parameters["neg_prompt"] = parm[0]
310
  input_string = "Steps:" + parm[1]
311
 
312
  # Extracting Steps
@@ -355,12 +361,13 @@ class GuiSD:
355
  retain_task_model_in_cache=False,
356
  device="cpu",
357
  )
358
- self.model.device = torch.device("cpu") #
 
359
 
360
  def infer_short(self, model, pipe_params, progress=gr.Progress(track_tqdm=True)):
361
- progress(0, desc="Start inference...")
362
  images, seed, image_list, metadata = model(**pipe_params)
363
- progress(1, desc="Inference completed.")
364
  if not isinstance(images, list): images = [images]
365
  images = save_images(images, metadata)
366
  img = []
@@ -500,7 +507,7 @@ class GuiSD:
500
  pag_scale,
501
  progress=gr.Progress(track_tqdm=True),
502
  ):
503
- progress(0, desc="Preparing inference...")
504
 
505
  vae_model = vae_model if vae_model != "None" else None
506
  loras_list = [lora1, lora2, lora3, lora4, lora5]
@@ -670,7 +677,7 @@ class GuiSD:
670
  self.model.pipe.transformer.to(self.model.device)
671
  print("transformer to cuda")
672
 
673
- progress(1, desc="Inference preparation completed. Starting inference...")
674
 
675
  info_state = "" # for yield version
676
  return self.infer_short(self.model, pipe_params, progress), info_state
@@ -805,9 +812,12 @@ def infer(prompt, negative_prompt, seed, randomize_seed, width, height, guidance
805
  lora3 = get_valid_lora_path(lora3)
806
  lora4 = get_valid_lora_path(lora4)
807
  lora5 = get_valid_lora_path(lora5)
808
- progress(1, desc="Preparation completed. Starting inference preparation...")
809
 
810
- sd_gen.load_new_model(model_name, vae, TASK_MODEL_LIST[0], progress)
 
 
 
811
  images, info = sd_gen_generate_pipeline(prompt, negative_prompt, 1, num_inference_steps,
812
  guidance_scale, True, generator, lora1, lora1_wt, lora2, lora2_wt, lora3, lora3_wt,
813
  lora4, lora4_wt, lora5, lora5_wt, sampler,
@@ -820,7 +830,6 @@ def infer(prompt, negative_prompt, seed, randomize_seed, width, height, guidance
820
  True, None, None, "plus_face", "original", 0.7, None, None, "base", "style", 0.7, 0.0,
821
  load_lora_cpu, verbose_info, gpu_duration
822
  )
823
-
824
  progress(1, desc="Inference completed.")
825
  output_image = images[0][0] if images else None
826
 
 
23
  )
24
  import time
25
  #import urllib.parse
26
+
27
+ print(os.getenv("SPACES_ZERO_GPU"))
28
+
29
  import gradio as gr
30
  import logging
31
  logging.getLogger("diffusers").setLevel(logging.ERROR)
 
55
  ],
56
  "scribble": [
57
  "HED",
58
+ "PidiNet",
59
  "None",
60
  ],
61
  "softedge": [
62
+ "PidiNet",
63
  "HED",
64
  "HED safe",
65
+ "PidiNet safe",
66
  "None",
67
  ],
68
  "segmentation": [
 
298
  input_string = input_string.replace("\n", "")
299
 
300
  if "Negative prompt:" not in input_string:
301
+ if "Steps:" in input_string:
302
+ input_string = input_string.replace("Steps:", "Negative prompt: Steps:")
303
+ else:
304
+ print("Invalid metadata")
305
+ parameters["prompt"] = input_string
306
+ return parameters
307
 
308
  parm = input_string.split("Negative prompt:")
309
+ parameters["prompt"] = parm[0].strip()
310
  if "Steps:" not in parm[1]:
311
  print("Steps not detected")
312
+ parameters["neg_prompt"] = parm[1].strip()
313
  return parameters
314
  parm = parm[1].split("Steps:")
315
+ parameters["neg_prompt"] = parm[0].strip()
316
  input_string = "Steps:" + parm[1]
317
 
318
  # Extracting Steps
 
361
  retain_task_model_in_cache=False,
362
  device="cpu",
363
  )
364
+ self.model.load_beta_styles()
365
+ #self.model.device = torch.device("cpu") #
366
 
367
  def infer_short(self, model, pipe_params, progress=gr.Progress(track_tqdm=True)):
368
+ #progress(0, desc="Start inference...")
369
  images, seed, image_list, metadata = model(**pipe_params)
370
+ #progress(1, desc="Inference completed.")
371
  if not isinstance(images, list): images = [images]
372
  images = save_images(images, metadata)
373
  img = []
 
507
  pag_scale,
508
  progress=gr.Progress(track_tqdm=True),
509
  ):
510
+ #progress(0, desc="Preparing inference...")
511
 
512
  vae_model = vae_model if vae_model != "None" else None
513
  loras_list = [lora1, lora2, lora3, lora4, lora5]
 
677
  self.model.pipe.transformer.to(self.model.device)
678
  print("transformer to cuda")
679
 
680
+ #progress(1, desc="Inference preparation completed. Starting inference...")
681
 
682
  info_state = "" # for yield version
683
  return self.infer_short(self.model, pipe_params, progress), info_state
 
812
  lora3 = get_valid_lora_path(lora3)
813
  lora4 = get_valid_lora_path(lora4)
814
  lora5 = get_valid_lora_path(lora5)
815
+ progress(1, desc="Preparation completed. Starting inference...")
816
 
817
+ progress(0, desc="Loading model...")
818
+ sd_gen.load_new_model(model_name, vae, TASK_MODEL_LIST[0])
819
+ progress(1, desc="Model loaded.")
820
+ progress(0, desc="Starting Inference...")
821
  images, info = sd_gen_generate_pipeline(prompt, negative_prompt, 1, num_inference_steps,
822
  guidance_scale, True, generator, lora1, lora1_wt, lora2, lora2_wt, lora3, lora3_wt,
823
  lora4, lora4_wt, lora5, lora5_wt, sampler,
 
830
  True, None, None, "plus_face", "original", 0.7, None, None, "base", "style", 0.7, 0.0,
831
  load_lora_cpu, verbose_info, gpu_duration
832
  )
 
833
  progress(1, desc="Inference completed.")
834
  output_image = images[0][0] if images else None
835
 
env.py CHANGED
@@ -39,6 +39,11 @@ load_diffusers_format_model = [
39
  'votepurchase/juggernautXL_hyper_8step_sfw',
40
  'votepurchase/ponyRealism_v21MainVAE',
41
  'stabilityai/stable-diffusion-xl-base-1.0',
 
 
 
 
 
42
  'cagliostrolab/animagine-xl-3.1',
43
  'John6666/epicrealism-xl-v8kiss-sdxl',
44
  'misri/epicrealismXL_v7FinalDestination',
@@ -54,15 +59,30 @@ load_diffusers_format_model = [
54
  'WhiteAiZ/autismmixSDXL_autismmixConfetti_diffusers',
55
  'kitty7779/ponyDiffusionV6XL',
56
  'GraydientPlatformAPI/aniverse-pony',
 
 
57
  'John6666/mistoon-anime-ponyalpha-sdxl',
 
 
58
  'John6666/ebara-mfcg-pony-mix-v12-sdxl',
59
  'John6666/t-ponynai3-v51-sdxl',
 
 
60
  'John6666/mala-anime-mix-nsfw-pony-xl-v5-sdxl',
61
  'John6666/wai-real-mix-v11-sdxl',
 
 
 
 
 
62
  'John6666/cyberrealistic-pony-v63-sdxl',
 
63
  'GraydientPlatformAPI/realcartoon-pony-diffusion',
64
  'John6666/nova-anime-xl-pony-v5-sdxl',
65
  'John6666/autismmix-sdxl-autismmix-pony-sdxl',
 
 
 
66
  'yodayo-ai/kivotos-xl-2.0',
67
  'yodayo-ai/holodayo-xl-2.1',
68
  'yodayo-ai/clandestine-xl-1.0',
@@ -81,6 +101,8 @@ load_diffusers_format_model = [
81
  'GraydientPlatformAPI/picx-real',
82
  'GraydientPlatformAPI/perfectworld6',
83
  'emilianJR/epiCRealism',
 
 
84
  'Meina/MeinaMix_V11',
85
  'Meina/MeinaUnreal_V5',
86
  'Meina/MeinaPastel_V7',
 
39
  'votepurchase/juggernautXL_hyper_8step_sfw',
40
  'votepurchase/ponyRealism_v21MainVAE',
41
  'stabilityai/stable-diffusion-xl-base-1.0',
42
+ 'John6666/blue-pencil-flux1-v021-fp8-flux',
43
+ 'John6666/wai-ani-flux-v10forfp8-fp8-flux',
44
+ 'John6666/xe-anime-flux-v04-fp8-flux',
45
+ 'John6666/lyh-anime-flux-v2a1-fp8-flux',
46
+ 'John6666/carnival-unchained-v10-fp8-flux',
47
  'cagliostrolab/animagine-xl-3.1',
48
  'John6666/epicrealism-xl-v8kiss-sdxl',
49
  'misri/epicrealismXL_v7FinalDestination',
 
59
  'WhiteAiZ/autismmixSDXL_autismmixConfetti_diffusers',
60
  'kitty7779/ponyDiffusionV6XL',
61
  'GraydientPlatformAPI/aniverse-pony',
62
+ 'John6666/ras-real-anime-screencap-v1-sdxl',
63
+ 'John6666/duchaiten-pony-xl-no-score-v60-sdxl',
64
  'John6666/mistoon-anime-ponyalpha-sdxl',
65
+ 'John6666/3x3x3mixxl-v2-sdxl',
66
+ 'John6666/3x3x3mixxl-3dv01-sdxl',
67
  'John6666/ebara-mfcg-pony-mix-v12-sdxl',
68
  'John6666/t-ponynai3-v51-sdxl',
69
+ 'John6666/t-ponynai3-v65-sdxl',
70
+ 'John6666/prefect-pony-xl-v3-sdxl',
71
  'John6666/mala-anime-mix-nsfw-pony-xl-v5-sdxl',
72
  'John6666/wai-real-mix-v11-sdxl',
73
+ 'John6666/wai-c-v6-sdxl',
74
+ 'John6666/iniverse-mix-xl-sfwnsfw-pony-guofeng-v43-sdxl',
75
+ 'John6666/photo-realistic-pony-v5-sdxl',
76
+ 'John6666/pony-realism-v21main-sdxl',
77
+ 'John6666/pony-realism-v22main-sdxl',
78
  'John6666/cyberrealistic-pony-v63-sdxl',
79
+ 'John6666/cyberrealistic-pony-v64-sdxl',
80
  'GraydientPlatformAPI/realcartoon-pony-diffusion',
81
  'John6666/nova-anime-xl-pony-v5-sdxl',
82
  'John6666/autismmix-sdxl-autismmix-pony-sdxl',
83
+ 'John6666/aimz-dream-real-pony-mix-v3-sdxl',
84
+ 'John6666/duchaiten-pony-real-v11fix-sdxl',
85
+ 'John6666/duchaiten-pony-real-v20-sdxl',
86
  'yodayo-ai/kivotos-xl-2.0',
87
  'yodayo-ai/holodayo-xl-2.1',
88
  'yodayo-ai/clandestine-xl-1.0',
 
101
  'GraydientPlatformAPI/picx-real',
102
  'GraydientPlatformAPI/perfectworld6',
103
  'emilianJR/epiCRealism',
104
+ 'votepurchase/counterfeitV30_v30',
105
+ 'votepurchase/ChilloutMix',
106
  'Meina/MeinaMix_V11',
107
  'Meina/MeinaUnreal_V5',
108
  'Meina/MeinaPastel_V7',
llmdolphin.py CHANGED
@@ -20,6 +20,7 @@ llm_models = {
20
  #"": ["", MessagesFormatterType.PHI_3],
21
  "mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
22
  "L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
 
23
  "MN-12B-Mag-Mell-Q4_K_M.gguf": ["inflatebot/MN-12B-Mag-Mell-R1-GGUF", MessagesFormatterType.MISTRAL],
24
  "Qwen2.5-14B-Instruct-Q4_K_M.gguf": ["bartowski/Qwen2.5-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
25
  "Instant-RP-Noodles-12B-v1.3.Q4_K_M.gguf": ["mradermacher/Instant-RP-Noodles-12B-v1.3-GGUF", MessagesFormatterType.MISTRAL],
@@ -62,6 +63,30 @@ llm_models = {
62
  "ChatWaifu_22B_v2.0_preview.Q4_K_S.gguf": ["mradermacher/ChatWaifu_22B_v2.0_preview-GGUF", MessagesFormatterType.MISTRAL],
63
  "ChatWaifu_v1.4.Q5_K_M.gguf": ["mradermacher/ChatWaifu_v1.4-GGUF", MessagesFormatterType.MISTRAL],
64
  "ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
65
  "eidolon-v1-14b-q4_k_m.gguf": ["Lambent/Eidolon-v1-14B-Q4_K_M-GGUF", MessagesFormatterType.OPEN_CHAT],
66
  "Mistral-Nemo-Gutenberg-Doppel-12B-v2.Q4_K_M.gguf": ["QuantFactory/Mistral-Nemo-Gutenberg-Doppel-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
67
  "MN-Dark-Planet-TITAN-12B-D_AU-Q4_k_s.gguf": ["DavidAU/MN-Dark-Planet-TITAN-12B-GGUF", MessagesFormatterType.CHATML],
 
20
  #"": ["", MessagesFormatterType.PHI_3],
21
  "mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
22
  "L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
23
+ "Mahou-1.5-mistral-nemo-12B.i1-Q4_K_M.gguf": ["mradermacher/Mahou-1.5-mistral-nemo-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
24
  "MN-12B-Mag-Mell-Q4_K_M.gguf": ["inflatebot/MN-12B-Mag-Mell-R1-GGUF", MessagesFormatterType.MISTRAL],
25
  "Qwen2.5-14B-Instruct-Q4_K_M.gguf": ["bartowski/Qwen2.5-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
26
  "Instant-RP-Noodles-12B-v1.3.Q4_K_M.gguf": ["mradermacher/Instant-RP-Noodles-12B-v1.3-GGUF", MessagesFormatterType.MISTRAL],
 
63
  "ChatWaifu_22B_v2.0_preview.Q4_K_S.gguf": ["mradermacher/ChatWaifu_22B_v2.0_preview-GGUF", MessagesFormatterType.MISTRAL],
64
  "ChatWaifu_v1.4.Q5_K_M.gguf": ["mradermacher/ChatWaifu_v1.4-GGUF", MessagesFormatterType.MISTRAL],
65
  "ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
66
+ "Humanish-Qwen2.5-7B-Instruct-Q5_K_M.gguf": ["bartowski/Humanish-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
67
+ "FactAlign-LLaMA-3-8B.i1-Q4_K_M.gguf": ["mradermacher/FactAlign-LLaMA-3-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
68
+ "Magot-v3-Gemma2-8k-9B.i1-Q4_K_M.gguf": ["mradermacher/Magot-v3-Gemma2-8k-9B-i1-GGUF", MessagesFormatterType.ALPACA],
69
+ "EVA-Qwen2.5-14B-v0.1-Q4_K_M.gguf": ["bartowski/EVA-Qwen2.5-14B-v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
70
+ "Flammades-Mistral-Nemo-12B.i1-Q4_K_M.gguf": ["mradermacher/Flammades-Mistral-Nemo-12B-i1-GGUF", MessagesFormatterType.CHATML],
71
+ "Humanish-LLama3-8B-Instruct-Q5_K_M.gguf": ["bartowski/Humanish-LLama3-8B-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
72
+ "Humanish-Mistral-Nemo-Instruct-2407-Q4_K_M.gguf": ["bartowski/Humanish-Mistral-Nemo-Instruct-2407-GGUF", MessagesFormatterType.MISTRAL],
73
+ "L3.1-Clouded-Uchtave-v0.1-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Clouded-Uchtave-v0.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
74
+ "Baldur-Q5_K_M.gguf": ["Delta-Vector/Baldur-8B-GGUF", MessagesFormatterType.LLAMA_3],
75
+ "llama-jpsft-q6_k.gguf": ["ai-net/Llama-JPSFT-GGUF", MessagesFormatterType.LLAMA_3],
76
+ "Gemma-2-Ataraxy-v3i-9B.Q5_K_M.gguf": ["mradermacher/Gemma-2-Ataraxy-v3i-9B-GGUF", MessagesFormatterType.ALPACA],
77
+ "EVA-Qwen2.5-7B-v0.1.i1-Q5_K_M.gguf": ["mradermacher/EVA-Qwen2.5-7B-v0.1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
78
+ "EZO-Llama-3.2-3B-Instruct-dpoE.Q4_K_S.gguf": ["mradermacher/EZO-Llama-3.2-3B-Instruct-dpoE-GGUF", MessagesFormatterType.LLAMA_3],
79
+ "Bezaliel_p1.1-12B.i1-Q4_K_M.gguf": ["mradermacher/Bezaliel_p1.1-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
80
+ "Ice0.7-29.09-RP.i1-Q5_K_M.gguf": ["mradermacher/Ice0.7-29.09-RP-i1-GGUF", MessagesFormatterType.MISTRAL],
81
+ "IceWhiskeyRP-7b.i1-Q5_K_M.gguf": ["mradermacher/IceWhiskeyRP-7b-i1-GGUF", MessagesFormatterType.ALPACA],
82
+ "L3.1-Purosani-2-8B.i1-Q4_K_M.gguf": ["mradermacher/L3.1-Purosani-2-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
83
+ "llama-3.2-3b-instruct-abliterated-q4_k_m.gguf": ["darkc0de/Llama-3.2-3B-Instruct-abliterated-Q4_K_M-GGUF", MessagesFormatterType.LLAMA_3],
84
+ "Daredevil-Aura-8B_uncensored_OAS_abliterated.i1-Q4_K_S.gguf": ["mradermacher/Daredevil-Aura-8B_uncensored_OAS_abliterated-i1-GGUF", MessagesFormatterType.LLAMA_3],
85
+ "Qwen-2.5-7b-TitanFusion-v3.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7b-TitanFusion-v3-GGUF", MessagesFormatterType.OPEN_CHAT],
86
+ "MixTAO-7Bx2-MoE-v8.1.Q4_K_M.gguf": ["mradermacher/MixTAO-7Bx2-MoE-v8.1-GGUF", MessagesFormatterType.ALPACA],
87
+ "Promissum_Mane-8B-LINEAR.i1-Q5_K_M.gguf": ["mradermacher/Promissum_Mane-8B-LINEAR-i1-GGUF", MessagesFormatterType.LLAMA_3],
88
+ "L3.1-Purosani-1.5-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Purosani-1.5-8B-GGUF", MessagesFormatterType.LLAMA_3],
89
+ "SzilviaB-Daredevil-LongWriter-8B_abliterated.i1-Q5_K_M.gguf": ["mradermacher/SzilviaB-Daredevil-LongWriter-8B_abliterated-i1-GGUF", MessagesFormatterType.LLAMA_3],
90
  "eidolon-v1-14b-q4_k_m.gguf": ["Lambent/Eidolon-v1-14B-Q4_K_M-GGUF", MessagesFormatterType.OPEN_CHAT],
91
  "Mistral-Nemo-Gutenberg-Doppel-12B-v2.Q4_K_M.gguf": ["QuantFactory/Mistral-Nemo-Gutenberg-Doppel-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
92
  "MN-Dark-Planet-TITAN-12B-D_AU-Q4_k_s.gguf": ["DavidAU/MN-Dark-Planet-TITAN-12B-GGUF", MessagesFormatterType.CHATML],