John6666 commited on
Commit
691c505
·
verified ·
1 Parent(s): 51f2f9a

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +13 -12
  2. modutils.py +9 -8
app.py CHANGED
@@ -344,8 +344,9 @@ class GuiSD:
344
  vae_model=None,
345
  type_model_precision=torch.float16,
346
  retain_task_model_in_cache=False,
347
- device="cpu",
348
  )
 
349
 
350
  def infer_short(self, model, pipe_params, progress=gr.Progress(track_tqdm=True)):
351
  progress(0, desc="Start inference...")
@@ -725,12 +726,12 @@ class GuiSD:
725
  sd_gen = GuiSD()
726
 
727
  #@spaces.GPU
728
- def sd_gen_load_new_model(*args, progress=gr.Progress(track_tqdm=True)):
729
- yield from sd_gen.load_new_model(*args, progress)
730
 
731
- @spaces.GPU
732
- def sd_gen_generate_pipeline(*args, progress=gr.Progress(track_tqdm=True)):
733
- yield from sd_gen.generate_pipeline(*args, progress)
734
 
735
  ## BEGIN MOD
736
  CSS ="""
@@ -1250,8 +1251,8 @@ with gr.Blocks(theme='NoCrypt/miku@>=1.2.2', elem_id="main", fill_width=True, cs
1250
  "votepurchase/animagine-xl-3.1",
1251
  ],
1252
  ],
1253
- #fn=sd_gen.generate_pipeline,
1254
- fn=sd_gen_generate_pipeline,
1255
  inputs=[
1256
  prompt_gui,
1257
  neg_prompt_gui,
@@ -1430,8 +1431,8 @@ with gr.Blocks(theme='NoCrypt/miku@>=1.2.2', elem_id="main", fill_width=True, cs
1430
  translate_prompt_button.click(translate_prompt, [series_dbt], [series_dbt], queue=False)
1431
 
1432
  generate_button.click(
1433
- #fn=sd_gen.load_new_model,
1434
- fn=sd_gen_load_new_model,
1435
  inputs=[
1436
  model_name_gui,
1437
  vae_model_gui,
@@ -1441,8 +1442,8 @@ with gr.Blocks(theme='NoCrypt/miku@>=1.2.2', elem_id="main", fill_width=True, cs
1441
  queue=True,
1442
  show_progress="minimal",
1443
  ).success(
1444
- #fn=sd_gen.generate_pipeline,
1445
- fn=sd_gen_generate_pipeline,
1446
  inputs=[
1447
  prompt_gui,
1448
  neg_prompt_gui,
 
344
  vae_model=None,
345
  type_model_precision=torch.float16,
346
  retain_task_model_in_cache=False,
347
+ #device="cpu",
348
  )
349
+ self.model.device = torch.device("cpu") #
350
 
351
  def infer_short(self, model, pipe_params, progress=gr.Progress(track_tqdm=True)):
352
  progress(0, desc="Start inference...")
 
726
  sd_gen = GuiSD()
727
 
728
  #@spaces.GPU
729
+ #def sd_gen_load_new_model(*args, progress=gr.Progress(track_tqdm=True)):
730
+ # yield from sd_gen.load_new_model(*args, progress)
731
 
732
+ #@spaces.GPU
733
+ #def sd_gen_generate_pipeline(*args, progress=gr.Progress(track_tqdm=True)):
734
+ # yield from sd_gen.generate_pipeline(*args, progress)
735
 
736
  ## BEGIN MOD
737
  CSS ="""
 
1251
  "votepurchase/animagine-xl-3.1",
1252
  ],
1253
  ],
1254
+ fn=sd_gen.generate_pipeline,
1255
+ #fn=sd_gen_generate_pipeline,
1256
  inputs=[
1257
  prompt_gui,
1258
  neg_prompt_gui,
 
1431
  translate_prompt_button.click(translate_prompt, [series_dbt], [series_dbt], queue=False)
1432
 
1433
  generate_button.click(
1434
+ fn=sd_gen.load_new_model,
1435
+ #fn=sd_gen_load_new_model,
1436
  inputs=[
1437
  model_name_gui,
1438
  vae_model_gui,
 
1442
  queue=True,
1443
  show_progress="minimal",
1444
  ).success(
1445
+ fn=sd_gen.generate_pipeline,
1446
+ #fn=sd_gen_generate_pipeline,
1447
  inputs=[
1448
  prompt_gui,
1449
  neg_prompt_gui,
modutils.py CHANGED
@@ -39,7 +39,6 @@ def get_local_model_list(dir_path):
39
 
40
  def download_things(directory, url, hf_token="", civitai_api_key=""):
41
  url = url.strip()
42
-
43
  if "drive.google.com" in url:
44
  original_dir = os.getcwd()
45
  os.chdir(directory)
@@ -187,10 +186,10 @@ def get_model_id_list():
187
  try:
188
  models_likes = []
189
  for author in HF_MODEL_USER_LIKES:
190
- models_likes.extend(api.list_models(author=author, cardData=True, sort="likes"))
191
  models_ex = []
192
  for author in HF_MODEL_USER_EX:
193
- models_ex = api.list_models(author=author, cardData=True, sort="last_modified")
194
  except Exception as e:
195
  print(f"Error: Failed to list {author}'s models.")
196
  print(e)
@@ -200,8 +199,8 @@ def get_model_id_list():
200
  anime_models = []
201
  real_models = []
202
  for model in models_ex:
203
- if not model.private:
204
- anime_models.append(model.id) if 'anime' in model.tags else real_models.append(model.id)
205
  model_ids.extend(anime_models)
206
  model_ids.extend(real_models)
207
  model_id_list = model_ids.copy()
@@ -252,6 +251,8 @@ def get_tupled_model_list(model_list):
252
  tags = model.tags
253
  info = []
254
  if not 'diffusers' in tags: continue
 
 
255
  if 'diffusers:StableDiffusionXLPipeline' in tags:
256
  info.append("SDXL")
257
  elif 'diffusers:StableDiffusionPipeline' in tags:
@@ -858,7 +859,7 @@ def find_similar_lora(q: str):
858
  def change_interface_mode(mode: str):
859
  if mode == "Fast":
860
  return gr.update(open=False), gr.update(visible=True), gr.update(open=False), gr.update(open=False),\
861
- gr.update(visible=True), gr.update(open=False), gr.update(visible=True), gr.update(open=True),\
862
  gr.update(visible=True), gr.update(value="Fast")
863
  elif mode == "Simple": # t2i mode
864
  return gr.update(open=True), gr.update(visible=True), gr.update(open=False), gr.update(open=False),\
@@ -866,11 +867,11 @@ def change_interface_mode(mode: str):
866
  gr.update(visible=False), gr.update(value="Standard")
867
  elif mode == "LoRA": # t2i LoRA mode
868
  return gr.update(open=True), gr.update(visible=True), gr.update(open=True), gr.update(open=False),\
869
- gr.update(visible=True), gr.update(open=True), gr.update(visible=True), gr.update(open=True),\
870
  gr.update(visible=False), gr.update(value="Standard")
871
  else: # Standard
872
  return gr.update(open=False), gr.update(visible=True), gr.update(open=False), gr.update(open=False),\
873
- gr.update(visible=True), gr.update(open=False), gr.update(visible=True), gr.update(open=True),\
874
  gr.update(visible=True), gr.update(value="Standard")
875
 
876
 
 
39
 
40
  def download_things(directory, url, hf_token="", civitai_api_key=""):
41
  url = url.strip()
 
42
  if "drive.google.com" in url:
43
  original_dir = os.getcwd()
44
  os.chdir(directory)
 
186
  try:
187
  models_likes = []
188
  for author in HF_MODEL_USER_LIKES:
189
+ models_likes.extend(api.list_models(author=author, task="text-to-image", cardData=True, sort="likes"))
190
  models_ex = []
191
  for author in HF_MODEL_USER_EX:
192
+ models_ex = api.list_models(author=author, task="text-to-image", cardData=True, sort="last_modified")
193
  except Exception as e:
194
  print(f"Error: Failed to list {author}'s models.")
195
  print(e)
 
199
  anime_models = []
200
  real_models = []
201
  for model in models_ex:
202
+ if not model.private and not model.gated and "diffusers:FluxPipeline" not in model.tags:
203
+ anime_models.append(model.id) if "anime" in model.tags else real_models.append(model.id)
204
  model_ids.extend(anime_models)
205
  model_ids.extend(real_models)
206
  model_id_list = model_ids.copy()
 
251
  tags = model.tags
252
  info = []
253
  if not 'diffusers' in tags: continue
254
+ if 'diffusers:FluxPipeline' in tags:
255
+ info.append("FLUX.1")
256
  if 'diffusers:StableDiffusionXLPipeline' in tags:
257
  info.append("SDXL")
258
  elif 'diffusers:StableDiffusionPipeline' in tags:
 
859
  def change_interface_mode(mode: str):
860
  if mode == "Fast":
861
  return gr.update(open=False), gr.update(visible=True), gr.update(open=False), gr.update(open=False),\
862
+ gr.update(visible=True), gr.update(open=False), gr.update(visible=True), gr.update(open=False),\
863
  gr.update(visible=True), gr.update(value="Fast")
864
  elif mode == "Simple": # t2i mode
865
  return gr.update(open=True), gr.update(visible=True), gr.update(open=False), gr.update(open=False),\
 
867
  gr.update(visible=False), gr.update(value="Standard")
868
  elif mode == "LoRA": # t2i LoRA mode
869
  return gr.update(open=True), gr.update(visible=True), gr.update(open=True), gr.update(open=False),\
870
+ gr.update(visible=True), gr.update(open=True), gr.update(visible=True), gr.update(open=False),\
871
  gr.update(visible=False), gr.update(value="Standard")
872
  else: # Standard
873
  return gr.update(open=False), gr.update(visible=True), gr.update(open=False), gr.update(open=False),\
874
+ gr.update(visible=True), gr.update(open=False), gr.update(visible=True), gr.update(open=False),\
875
  gr.update(visible=True), gr.update(value="Standard")
876
 
877