Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,28 +1,3 @@
|
|
| 1 |
-
task_stablepy = {
|
| 2 |
-
'txt2img': 'txt2img',
|
| 3 |
-
'img2img': 'img2img',
|
| 4 |
-
'inpaint': 'inpaint',
|
| 5 |
-
'sdxl_canny T2I Adapter': 'sdxl_canny',
|
| 6 |
-
'sdxl_sketch T2I Adapter': 'sdxl_sketch',
|
| 7 |
-
'sdxl_lineart T2I Adapter': 'sdxl_lineart',
|
| 8 |
-
'sdxl_depth-midas T2I Adapter': 'sdxl_depth-midas',
|
| 9 |
-
'sdxl_openpose T2I Adapter': 'sdxl_openpose',
|
| 10 |
-
'sd_openpose ControlNet': 'openpose',
|
| 11 |
-
'sd_canny ControlNet': 'canny',
|
| 12 |
-
'sd_mlsd ControlNet': 'mlsd',
|
| 13 |
-
'sd_scribble ControlNet': 'scribble',
|
| 14 |
-
'sd_softedge ControlNet': 'softedge',
|
| 15 |
-
'sd_segmentation ControlNet': 'segmentation',
|
| 16 |
-
'sd_depth ControlNet': 'depth',
|
| 17 |
-
'sd_normalbae ControlNet': 'normalbae',
|
| 18 |
-
'sd_lineart ControlNet': 'lineart',
|
| 19 |
-
'sd_lineart_anime ControlNet': 'lineart_anime',
|
| 20 |
-
'sd_shuffle ControlNet': 'shuffle',
|
| 21 |
-
'sd_ip2p ControlNet': 'ip2p',
|
| 22 |
-
}
|
| 23 |
-
|
| 24 |
-
task_model_list = list(task_stablepy.keys())
|
| 25 |
-
|
| 26 |
#######################
|
| 27 |
# UTILS
|
| 28 |
#######################
|
|
@@ -34,7 +9,22 @@ from stablepy.diffusers_vanilla.style_prompt_config import STYLE_NAMES
|
|
| 34 |
import torch
|
| 35 |
import re
|
| 36 |
import shutil
|
| 37 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
|
| 39 |
preprocessor_controlnet = {
|
| 40 |
"openpose": [
|
|
@@ -88,6 +78,33 @@ preprocessor_controlnet = {
|
|
| 88 |
]
|
| 89 |
}
|
| 90 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 91 |
|
| 92 |
def download_things(directory, url, hf_token="", civitai_api_key=""):
|
| 93 |
url = url.strip()
|
|
@@ -154,9 +171,9 @@ os.makedirs(directory_vaes, exist_ok=True)
|
|
| 154 |
# - **Download SD 1.5 Models**
|
| 155 |
download_model = "https://huggingface.co/frankjoshua/toonyou_beta6/resolve/main/toonyou_beta6.safetensors"
|
| 156 |
# - **Download VAEs**
|
| 157 |
-
download_vae = "https://huggingface.co/madebyollin/sdxl-vae-fp16-fix/resolve/main/sdxl.vae.safetensors?download=true, https://huggingface.co/nubby/blessed-sdxl-vae-fp16-fix/resolve/main/sdxl_vae-fp16fix-c-1.1-b-0.5.safetensors?download=true, https://huggingface.co/nubby/blessed-sdxl-vae-fp16-fix/resolve/main/sdxl_vae-fp16fix-blessed.safetensors?download=true, https://huggingface.co/digiplay/VAE/resolve/main/vividReal_v20.safetensors?download=true, https://huggingface.co/fp16-guy/anything_kl-f8-anime2_vae-ft-mse-840000-ema-pruned_blessed_clearvae_fp16_cleaned/resolve/main/
|
| 158 |
# - **Download LoRAs**
|
| 159 |
-
download_lora = "https://civitai.com/api/download/models/135867, https://civitai.com/api/download/models/135931, https://civitai.com/api/download/models/177492, https://civitai.com/api/download/models/145907, https://huggingface.co/Linaqruf/anime-detailer-xl-lora/resolve/main/anime-detailer-xl.safetensors?download=true, https://huggingface.co/Linaqruf/style-enhancer-xl-lora/resolve/main/style-enhancer-xl.safetensors?download=true, https://civitai.com/api/download/models/28609"
|
| 160 |
load_diffusers_format_model = [
|
| 161 |
'stabilityai/stable-diffusion-xl-base-1.0',
|
| 162 |
'misri/epicrealismXL_v7FinalDestination',
|
|
@@ -165,7 +182,6 @@ load_diffusers_format_model = [
|
|
| 165 |
'cagliostrolab/animagine-xl-3.1',
|
| 166 |
'misri/kohakuXLEpsilon_rev1',
|
| 167 |
'kitty7779/ponyDiffusionV6XL',
|
| 168 |
-
'runwayml/stable-diffusion-v1-5',
|
| 169 |
'digiplay/majicMIX_realistic_v6',
|
| 170 |
'digiplay/majicMIX_realistic_v7',
|
| 171 |
'digiplay/DreamShaper_8',
|
|
@@ -191,16 +207,9 @@ for url in [url.strip() for url in download_lora.split(',')]:
|
|
| 191 |
directory_embeds = 'embedings'
|
| 192 |
os.makedirs(directory_embeds, exist_ok=True)
|
| 193 |
download_embeds = [
|
| 194 |
-
'https://huggingface.co/datasets/Nerfgun3/bad_prompt/resolve/main/bad_prompt.pt',
|
| 195 |
'https://huggingface.co/datasets/Nerfgun3/bad_prompt/blob/main/bad_prompt_version2.pt',
|
| 196 |
-
'https://huggingface.co/embed/EasyNegative/resolve/main/EasyNegative.safetensors',
|
| 197 |
'https://huggingface.co/embed/negative/resolve/main/EasyNegativeV2.safetensors',
|
| 198 |
'https://huggingface.co/embed/negative/resolve/main/bad-hands-5.pt',
|
| 199 |
-
'https://huggingface.co/embed/negative/resolve/main/bad-artist.pt',
|
| 200 |
-
'https://huggingface.co/embed/negative/resolve/main/ng_deepnegative_v1_75t.pt',
|
| 201 |
-
'https://huggingface.co/embed/negative/resolve/main/bad-artist-anime.pt',
|
| 202 |
-
'https://huggingface.co/embed/negative/resolve/main/bad-image-v2-39000.pt',
|
| 203 |
-
'https://huggingface.co/embed/negative/resolve/main/verybadimagenegative_v1.3.pt',
|
| 204 |
]
|
| 205 |
|
| 206 |
for url_embed in download_embeds:
|
|
@@ -299,21 +308,51 @@ warnings.filterwarnings(action="ignore", category=FutureWarning, module="transfo
|
|
| 299 |
from stablepy import logger
|
| 300 |
logger.setLevel(logging.DEBUG)
|
| 301 |
|
| 302 |
-
|
| 303 |
class GuiSD:
|
| 304 |
-
def __init__(self):
|
| 305 |
self.model = None
|
| 306 |
-
|
| 307 |
-
|
| 308 |
-
|
| 309 |
-
|
| 310 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 311 |
|
| 312 |
@spaces.GPU(duration=120)
|
| 313 |
def infer(self, model, pipe_params):
|
| 314 |
images, image_list = model(**pipe_params)
|
| 315 |
return images
|
| 316 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 317 |
def generate_pipeline(
|
| 318 |
self,
|
| 319 |
prompt,
|
|
@@ -406,8 +445,23 @@ class GuiSD:
|
|
| 406 |
mask_dilation_b,
|
| 407 |
mask_blur_b,
|
| 408 |
mask_padding_b,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 409 |
):
|
| 410 |
|
|
|
|
|
|
|
| 411 |
vae_model = vae_model if vae_model != "None" else None
|
| 412 |
loras_list = [lora1, lora2, lora3, lora4, lora5]
|
| 413 |
|
|
@@ -427,23 +481,45 @@ class GuiSD:
|
|
| 427 |
|
| 428 |
for la in loras_list:
|
| 429 |
if la is not None and la != "None":
|
| 430 |
-
|
|
|
|
| 431 |
if (model_is_xl and lora_type) or (not model_is_xl and not lora_type):
|
| 432 |
gr.Info(f"The LoRA {la} is for { 'SD 1.5' if model_is_xl else 'SDXL' }, but you are using { model_type }.")
|
| 433 |
|
| 434 |
task = task_stablepy[task]
|
| 435 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 436 |
# First load
|
| 437 |
model_precision = torch.float16
|
| 438 |
if not self.model:
|
| 439 |
-
from
|
| 440 |
|
| 441 |
print("Loading model...")
|
| 442 |
-
self.model =
|
| 443 |
base_model_id=model_name,
|
| 444 |
task_name=task,
|
| 445 |
vae_model=vae_model if vae_model != "None" else None,
|
| 446 |
-
type_model_precision=model_precision
|
|
|
|
| 447 |
)
|
| 448 |
|
| 449 |
if task != "txt2img" and not image_control:
|
|
@@ -473,7 +549,8 @@ class GuiSD:
|
|
| 473 |
model_name,
|
| 474 |
task_name=task,
|
| 475 |
vae_model=vae_model if vae_model != "None" else None,
|
| 476 |
-
type_model_precision=model_precision
|
|
|
|
| 477 |
)
|
| 478 |
|
| 479 |
if textual_inversion and self.model.class_name == "StableDiffusionXLPipeline":
|
|
@@ -575,27 +652,30 @@ class GuiSD:
|
|
| 575 |
"hires_negative_prompt": hires_negative_prompt,
|
| 576 |
"hires_sampler": hires_sampler,
|
| 577 |
"hires_before_adetailer": hires_before_adetailer,
|
| 578 |
-
"hires_after_adetailer": hires_after_adetailer
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 579 |
}
|
| 580 |
|
| 581 |
# print(pipe_params)
|
| 582 |
|
| 583 |
-
|
| 584 |
-
|
| 585 |
-
|
| 586 |
-
|
| 587 |
-
|
| 588 |
-
|
| 589 |
-
or (adetailer_active_a and adetailer_active_b)
|
| 590 |
-
or (upscaler_model and upscaler_increases_size > 1.7)
|
| 591 |
-
or (steps > 75)
|
| 592 |
-
or (image_resolution > 1048)
|
| 593 |
-
):
|
| 594 |
-
print("Inference 2")
|
| 595 |
-
return self.infer(self.model, pipe_params)
|
| 596 |
|
| 597 |
-
print("Inference 1")
|
| 598 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 599 |
|
| 600 |
|
| 601 |
sd_gen = GuiSD()
|
|
@@ -605,11 +685,8 @@ CSS ="""
|
|
| 605 |
#component-0 { height: 100%; }
|
| 606 |
#gallery { flex-grow: 1; }
|
| 607 |
"""
|
| 608 |
-
|
| 609 |
-
|
| 610 |
-
sd_task = task_model_list[:3] + task_model_list[8:]
|
| 611 |
-
|
| 612 |
-
|
| 613 |
def update_task_options(model_name, task_name):
|
| 614 |
if model_name in model_list:
|
| 615 |
if "xl" in model_name.lower():
|
|
@@ -641,14 +718,20 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 641 |
model_name_gui = gr.Dropdown(label="Model", choices=model_list, value=model_list[0], allow_custom_value=True)
|
| 642 |
prompt_gui = gr.Textbox(lines=5, placeholder="Enter prompt", label="Prompt")
|
| 643 |
neg_prompt_gui = gr.Textbox(lines=3, placeholder="Enter Neg prompt", label="Negative prompt")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 644 |
generate_button = gr.Button(value="GENERATE", variant="primary")
|
| 645 |
-
|
| 646 |
model_name_gui.change(
|
| 647 |
update_task_options,
|
| 648 |
[model_name_gui, task_gui],
|
| 649 |
[task_gui],
|
| 650 |
)
|
| 651 |
|
|
|
|
|
|
|
| 652 |
result_images = gr.Gallery(
|
| 653 |
label="Generated images",
|
| 654 |
show_label=False,
|
|
@@ -662,20 +745,147 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 662 |
selected_index=50,
|
| 663 |
)
|
| 664 |
|
|
|
|
|
|
|
| 665 |
with gr.Column(scale=1):
|
| 666 |
steps_gui = gr.Slider(minimum=1, maximum=100, step=1, value=30, label="Steps")
|
| 667 |
cfg_gui = gr.Slider(minimum=0, maximum=30, step=0.5, value=7.5, label="CFG")
|
| 668 |
sampler_gui = gr.Dropdown(label="Sampler", choices=scheduler_names, value="Euler a")
|
| 669 |
img_width_gui = gr.Slider(minimum=64, maximum=4096, step=8, value=1024, label="Img Width")
|
| 670 |
img_height_gui = gr.Slider(minimum=64, maximum=4096, step=8, value=1024, label="Img Height")
|
| 671 |
-
clip_skip_gui = gr.Checkbox(value=True, label="Layer 2 Clip Skip")
|
| 672 |
-
free_u_gui = gr.Checkbox(value=True, label="FreeU")
|
| 673 |
seed_gui = gr.Number(minimum=-1, maximum=9999999999, value=-1, label="Seed")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 674 |
num_images_gui = gr.Slider(minimum=1, maximum=4, step=1, value=1, label="Images")
|
| 675 |
-
prompt_s_options = [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 676 |
prompt_syntax_gui = gr.Dropdown(label="Prompt Syntax", choices=prompt_s_options, value=prompt_s_options[0][1])
|
| 677 |
vae_model_gui = gr.Dropdown(label="VAE Model", choices=vae_model_list)
|
| 678 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 679 |
with gr.Accordion("ControlNet / Img2img / Inpaint", open=False, visible=True):
|
| 680 |
image_control = gr.Image(label="Image ControlNet/Inpaint/Img2img", type="filepath")
|
| 681 |
image_mask_gui = gr.Image(label="Image Mask", type="filepath")
|
|
@@ -685,7 +895,7 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 685 |
)
|
| 686 |
image_resolution_gui = gr.Slider(minimum=64, maximum=2048, step=64, value=1024, label="Image Resolution")
|
| 687 |
preprocessor_name_gui = gr.Dropdown(label="Preprocessor Name", choices=preprocessor_controlnet["canny"])
|
| 688 |
-
|
| 689 |
def change_preprocessor_choices(task):
|
| 690 |
task = task_stablepy[task]
|
| 691 |
if task in preprocessor_controlnet.keys():
|
|
@@ -693,7 +903,7 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 693 |
else:
|
| 694 |
choices_task = preprocessor_controlnet["canny"]
|
| 695 |
return gr.update(choices=choices_task, value=choices_task[0])
|
| 696 |
-
|
| 697 |
task_gui.change(
|
| 698 |
change_preprocessor_choices,
|
| 699 |
[task_gui],
|
|
@@ -713,25 +923,13 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 713 |
adapter_conditioning_scale_gui = gr.Slider(minimum=0, maximum=5., step=0.1, value=1, label="Adapter Conditioning Scale")
|
| 714 |
adapter_conditioning_factor_gui = gr.Slider(minimum=0, maximum=1., step=0.01, value=0.55, label="Adapter Conditioning Factor (%)")
|
| 715 |
|
| 716 |
-
with gr.Accordion("LoRA", open=False, visible=True):
|
| 717 |
-
lora1_gui = gr.Dropdown(label="Lora1", choices=lora_model_list)
|
| 718 |
-
lora_scale_1_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 1")
|
| 719 |
-
lora2_gui = gr.Dropdown(label="Lora2", choices=lora_model_list)
|
| 720 |
-
lora_scale_2_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 2")
|
| 721 |
-
lora3_gui = gr.Dropdown(label="Lora3", choices=lora_model_list)
|
| 722 |
-
lora_scale_3_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 3")
|
| 723 |
-
lora4_gui = gr.Dropdown(label="Lora4", choices=lora_model_list)
|
| 724 |
-
lora_scale_4_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 4")
|
| 725 |
-
lora5_gui = gr.Dropdown(label="Lora5", choices=lora_model_list)
|
| 726 |
-
lora_scale_5_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 5")
|
| 727 |
-
|
| 728 |
with gr.Accordion("Styles", open=False, visible=True):
|
| 729 |
-
|
| 730 |
try:
|
| 731 |
style_names_found = sd_gen.model.STYLE_NAMES
|
| 732 |
except:
|
| 733 |
style_names_found = STYLE_NAMES
|
| 734 |
-
|
| 735 |
style_prompt_gui = gr.Dropdown(
|
| 736 |
style_names_found,
|
| 737 |
multiselect=True,
|
|
@@ -746,42 +944,28 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 746 |
if not sd_gen.model:
|
| 747 |
gr.Info("First load the model")
|
| 748 |
return gr.update(value=None, choices=STYLE_NAMES)
|
| 749 |
-
|
| 750 |
sd_gen.model.load_style_file(json)
|
| 751 |
gr.Info(f"{len(sd_gen.model.STYLE_NAMES)} styles loaded")
|
| 752 |
return gr.update(value=None, choices=sd_gen.model.STYLE_NAMES)
|
| 753 |
|
| 754 |
style_button.click(load_json_style_file, [style_json_gui], [style_prompt_gui])
|
| 755 |
-
|
| 756 |
with gr.Accordion("Textual inversion", open=False, visible=False):
|
| 757 |
active_textual_inversion_gui = gr.Checkbox(value=False, label="Active Textual Inversion in prompt")
|
| 758 |
|
| 759 |
-
with gr.Accordion("Hires fix", open=False, visible=True):
|
| 760 |
-
|
| 761 |
-
upscaler_keys = list(upscaler_dict_gui.keys())
|
| 762 |
-
|
| 763 |
-
upscaler_model_path_gui = gr.Dropdown(label="Upscaler", choices=upscaler_keys, value=upscaler_keys[0])
|
| 764 |
-
upscaler_increases_size_gui = gr.Slider(minimum=1.1, maximum=6., step=0.1, value=1.5, label="Upscale by")
|
| 765 |
-
esrgan_tile_gui = gr.Slider(minimum=0, value=100, maximum=500, step=1, label="ESRGAN Tile")
|
| 766 |
-
esrgan_tile_overlap_gui = gr.Slider(minimum=1, maximum=200, step=1, value=10, label="ESRGAN Tile Overlap")
|
| 767 |
-
hires_steps_gui = gr.Slider(minimum=0, value=30, maximum=100, step=1, label="Hires Steps")
|
| 768 |
-
hires_denoising_strength_gui = gr.Slider(minimum=0.1, maximum=1.0, step=0.01, value=0.55, label="Hires Denoising Strength")
|
| 769 |
-
hires_sampler_gui = gr.Dropdown(label="Hires Sampler", choices=["Use same sampler"] + scheduler_names[:-1], value="Use same sampler")
|
| 770 |
-
hires_prompt_gui = gr.Textbox(label="Hires Prompt", placeholder="Main prompt will be use", lines=3)
|
| 771 |
-
hires_negative_prompt_gui = gr.Textbox(label="Hires Negative Prompt", placeholder="Main negative prompt will be use", lines=3)
|
| 772 |
-
|
| 773 |
with gr.Accordion("Detailfix", open=False, visible=True):
|
| 774 |
|
| 775 |
# Adetailer Inpaint Only
|
| 776 |
adetailer_inpaint_only_gui = gr.Checkbox(label="Inpaint only", value=True)
|
| 777 |
-
|
| 778 |
# Adetailer Verbose
|
| 779 |
adetailer_verbose_gui = gr.Checkbox(label="Verbose", value=False)
|
| 780 |
-
|
| 781 |
# Adetailer Sampler
|
| 782 |
adetailer_sampler_options = ["Use same sampler"] + scheduler_names[:-1]
|
| 783 |
adetailer_sampler_gui = gr.Dropdown(label="Adetailer sampler:", choices=adetailer_sampler_options, value="Use same sampler")
|
| 784 |
-
|
| 785 |
with gr.Accordion("Detailfix A", open=False, visible=True):
|
| 786 |
# Adetailer A
|
| 787 |
adetailer_active_a_gui = gr.Checkbox(label="Enable Adetailer A", value=False)
|
|
@@ -794,7 +978,7 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 794 |
mask_dilation_a_gui = gr.Number(label="Mask dilation:", value=4, minimum=1)
|
| 795 |
mask_blur_a_gui = gr.Number(label="Mask blur:", value=4, minimum=1)
|
| 796 |
mask_padding_a_gui = gr.Number(label="Mask padding:", value=32, minimum=1)
|
| 797 |
-
|
| 798 |
with gr.Accordion("Detailfix B", open=False, visible=True):
|
| 799 |
# Adetailer B
|
| 800 |
adetailer_active_b_gui = gr.Checkbox(label="Enable Adetailer B", value=False)
|
|
@@ -809,16 +993,17 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 809 |
mask_padding_b_gui = gr.Number(label="Mask padding:", value=32, minimum=1)
|
| 810 |
|
| 811 |
with gr.Accordion("Other settings", open=False, visible=True):
|
|
|
|
| 812 |
hires_before_adetailer_gui = gr.Checkbox(value=False, label="Hires Before Adetailer")
|
| 813 |
hires_after_adetailer_gui = gr.Checkbox(value=True, label="Hires After Adetailer")
|
| 814 |
generator_in_cpu_gui = gr.Checkbox(value=False, label="Generator in CPU")
|
| 815 |
|
| 816 |
with gr.Accordion("More settings", open=False, visible=False):
|
| 817 |
loop_generation_gui = gr.Slider(minimum=1, value=1, label="Loop Generation")
|
|
|
|
| 818 |
leave_progress_bar_gui = gr.Checkbox(value=True, label="Leave Progress Bar")
|
| 819 |
disable_progress_bar_gui = gr.Checkbox(value=False, label="Disable Progress Bar")
|
| 820 |
-
|
| 821 |
-
display_images_gui = gr.Checkbox(value=False, label="Display Images")
|
| 822 |
save_generated_images_gui = gr.Checkbox(value=False, label="Save Generated Images")
|
| 823 |
image_storage_location_gui = gr.Textbox(value="./images", label="Image Storage Location")
|
| 824 |
retain_compel_previous_load_gui = gr.Checkbox(value=False, label="Retain Compel Previous Load")
|
|
@@ -938,7 +1123,7 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 938 |
1024,
|
| 939 |
"misri/epicrealismXL_v7FinalDestination",
|
| 940 |
None, # vae
|
| 941 |
-
"
|
| 942 |
"image.webp", # img conttol
|
| 943 |
"Canny", # preprocessor
|
| 944 |
1024, # preproc resolution
|
|
@@ -1067,7 +1252,7 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 1067 |
512,
|
| 1068 |
"digiplay/majicMIX_realistic_v7",
|
| 1069 |
None, # vae
|
| 1070 |
-
"
|
| 1071 |
"image.webp", # img conttol
|
| 1072 |
"Canny", # preprocessor
|
| 1073 |
512, # preproc resolution
|
|
@@ -1176,7 +1361,7 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 1176 |
brush=gr.Brush(
|
| 1177 |
default_size="16", # or leave it as 'auto'
|
| 1178 |
color_mode="fixed", # 'fixed' hides the user swatches and colorpicker, 'defaults' shows it
|
| 1179 |
-
#default_color="black", # html names are supported
|
| 1180 |
colors=[
|
| 1181 |
"rgba(0, 0, 0, 1)", # rgb(a)
|
| 1182 |
"rgba(0, 0, 0, 0.1)",
|
|
@@ -1200,6 +1385,16 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 1200 |
btn_send.click(send_img, [img_source, img_result], [image_control, image_mask_gui])
|
| 1201 |
|
| 1202 |
generate_button.click(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1203 |
fn=sd_gen.generate_pipeline,
|
| 1204 |
inputs=[
|
| 1205 |
prompt_gui,
|
|
@@ -1292,9 +1487,21 @@ with gr.Blocks(theme="NoCrypt/miku", css=CSS) as app:
|
|
| 1292 |
mask_dilation_b_gui,
|
| 1293 |
mask_blur_b_gui,
|
| 1294 |
mask_padding_b_gui,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1295 |
],
|
| 1296 |
-
outputs=[result_images],
|
| 1297 |
queue=True,
|
|
|
|
| 1298 |
)
|
| 1299 |
|
| 1300 |
app.queue()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
#######################
|
| 2 |
# UTILS
|
| 3 |
#######################
|
|
|
|
| 9 |
import torch
|
| 10 |
import re
|
| 11 |
import shutil
|
| 12 |
+
import random
|
| 13 |
+
from stablepy import (
|
| 14 |
+
CONTROLNET_MODEL_IDS,
|
| 15 |
+
VALID_TASKS,
|
| 16 |
+
T2I_PREPROCESSOR_NAME,
|
| 17 |
+
FLASH_LORA,
|
| 18 |
+
SCHEDULER_CONFIG_MAP,
|
| 19 |
+
scheduler_names,
|
| 20 |
+
IP_ADAPTER_MODELS,
|
| 21 |
+
IP_ADAPTERS_SD,
|
| 22 |
+
IP_ADAPTERS_SDXL,
|
| 23 |
+
REPO_IMAGE_ENCODER,
|
| 24 |
+
ALL_PROMPT_WEIGHT_OPTIONS,
|
| 25 |
+
SD15_TASKS,
|
| 26 |
+
SDXL_TASKS,
|
| 27 |
+
)
|
| 28 |
|
| 29 |
preprocessor_controlnet = {
|
| 30 |
"openpose": [
|
|
|
|
| 78 |
]
|
| 79 |
}
|
| 80 |
|
| 81 |
+
task_stablepy = {
|
| 82 |
+
'txt2img': 'txt2img',
|
| 83 |
+
'img2img': 'img2img',
|
| 84 |
+
'inpaint': 'inpaint',
|
| 85 |
+
# 'canny T2I Adapter': 'sdxl_canny_t2i', # NO HAVE STEP CALLBACK PARAMETERS SO NOT WORKS WITH DIFFUSERS 0.29.0
|
| 86 |
+
# 'sketch T2I Adapter': 'sdxl_sketch_t2i',
|
| 87 |
+
# 'lineart T2I Adapter': 'sdxl_lineart_t2i',
|
| 88 |
+
# 'depth-midas T2I Adapter': 'sdxl_depth-midas_t2i',
|
| 89 |
+
# 'openpose T2I Adapter': 'sdxl_openpose_t2i',
|
| 90 |
+
'openpose ControlNet': 'openpose',
|
| 91 |
+
'canny ControlNet': 'canny',
|
| 92 |
+
'mlsd ControlNet': 'mlsd',
|
| 93 |
+
'scribble ControlNet': 'scribble',
|
| 94 |
+
'softedge ControlNet': 'softedge',
|
| 95 |
+
'segmentation ControlNet': 'segmentation',
|
| 96 |
+
'depth ControlNet': 'depth',
|
| 97 |
+
'normalbae ControlNet': 'normalbae',
|
| 98 |
+
'lineart ControlNet': 'lineart',
|
| 99 |
+
'lineart_anime ControlNet': 'lineart_anime',
|
| 100 |
+
'shuffle ControlNet': 'shuffle',
|
| 101 |
+
'ip2p ControlNet': 'ip2p',
|
| 102 |
+
'optical pattern ControlNet': 'pattern',
|
| 103 |
+
'tile realistic': 'sdxl_tile_realistic',
|
| 104 |
+
}
|
| 105 |
+
|
| 106 |
+
task_model_list = list(task_stablepy.keys())
|
| 107 |
+
|
| 108 |
|
| 109 |
def download_things(directory, url, hf_token="", civitai_api_key=""):
|
| 110 |
url = url.strip()
|
|
|
|
| 171 |
# - **Download SD 1.5 Models**
|
| 172 |
download_model = "https://huggingface.co/frankjoshua/toonyou_beta6/resolve/main/toonyou_beta6.safetensors"
|
| 173 |
# - **Download VAEs**
|
| 174 |
+
download_vae = "https://huggingface.co/madebyollin/sdxl-vae-fp16-fix/resolve/main/sdxl.vae.safetensors?download=true, https://huggingface.co/nubby/blessed-sdxl-vae-fp16-fix/resolve/main/sdxl_vae-fp16fix-c-1.1-b-0.5.safetensors?download=true, https://huggingface.co/nubby/blessed-sdxl-vae-fp16-fix/resolve/main/sdxl_vae-fp16fix-blessed.safetensors?download=true, https://huggingface.co/digiplay/VAE/resolve/main/vividReal_v20.safetensors?download=true, https://huggingface.co/fp16-guy/anything_kl-f8-anime2_vae-ft-mse-840000-ema-pruned_blessed_clearvae_fp16_cleaned/resolve/main/vae-ft-mse-840000-ema-pruned_fp16.safetensors?download=true"
|
| 175 |
# - **Download LoRAs**
|
| 176 |
+
download_lora = "https://civitai.com/api/download/models/135867, https://civitai.com/api/download/models/135931, https://civitai.com/api/download/models/177492, https://civitai.com/api/download/models/145907, https://huggingface.co/Linaqruf/anime-detailer-xl-lora/resolve/main/anime-detailer-xl.safetensors?download=true, https://huggingface.co/Linaqruf/style-enhancer-xl-lora/resolve/main/style-enhancer-xl.safetensors?download=true, https://civitai.com/api/download/models/28609, https://huggingface.co/ByteDance/Hyper-SD/resolve/main/Hyper-SD15-8steps-CFG-lora.safetensors?download=true, https://huggingface.co/ByteDance/Hyper-SD/resolve/main/Hyper-SDXL-8steps-CFG-lora.safetensors?download=true"
|
| 177 |
load_diffusers_format_model = [
|
| 178 |
'stabilityai/stable-diffusion-xl-base-1.0',
|
| 179 |
'misri/epicrealismXL_v7FinalDestination',
|
|
|
|
| 182 |
'cagliostrolab/animagine-xl-3.1',
|
| 183 |
'misri/kohakuXLEpsilon_rev1',
|
| 184 |
'kitty7779/ponyDiffusionV6XL',
|
|
|
|
| 185 |
'digiplay/majicMIX_realistic_v6',
|
| 186 |
'digiplay/majicMIX_realistic_v7',
|
| 187 |
'digiplay/DreamShaper_8',
|
|
|
|
| 207 |
directory_embeds = 'embedings'
|
| 208 |
os.makedirs(directory_embeds, exist_ok=True)
|
| 209 |
download_embeds = [
|
|
|
|
| 210 |
'https://huggingface.co/datasets/Nerfgun3/bad_prompt/blob/main/bad_prompt_version2.pt',
|
|
|
|
| 211 |
'https://huggingface.co/embed/negative/resolve/main/EasyNegativeV2.safetensors',
|
| 212 |
'https://huggingface.co/embed/negative/resolve/main/bad-hands-5.pt',
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 213 |
]
|
| 214 |
|
| 215 |
for url_embed in download_embeds:
|
|
|
|
| 308 |
from stablepy import logger
|
| 309 |
logger.setLevel(logging.DEBUG)
|
| 310 |
|
|
|
|
| 311 |
class GuiSD:
|
| 312 |
+
def __init__(self, stream=True):
|
| 313 |
self.model = None
|
| 314 |
+
|
| 315 |
+
print("Loading model...")
|
| 316 |
+
self.model = Model_Diffusers(
|
| 317 |
+
base_model_id="cagliostrolab/animagine-xl-3.1",
|
| 318 |
+
task_name="txt2img",
|
| 319 |
+
vae_model=None,
|
| 320 |
+
type_model_precision=torch.float16,
|
| 321 |
+
retain_task_model_in_cache=False,
|
| 322 |
+
)
|
| 323 |
|
| 324 |
@spaces.GPU(duration=120)
|
| 325 |
def infer(self, model, pipe_params):
|
| 326 |
images, image_list = model(**pipe_params)
|
| 327 |
return images
|
| 328 |
|
| 329 |
+
def load_new_model(self, model_name, vae_model, task, progress=gr.Progress(track_tqdm=True)):
|
| 330 |
+
|
| 331 |
+
yield f"Loading model: {model_name}"
|
| 332 |
+
|
| 333 |
+
vae_model = vae_model if vae_model != "None" else None
|
| 334 |
+
|
| 335 |
+
|
| 336 |
+
if model_name in model_list:
|
| 337 |
+
model_is_xl = "xl" in model_name.lower()
|
| 338 |
+
sdxl_in_vae = vae_model and "sdxl" in vae_model.lower()
|
| 339 |
+
model_type = "SDXL" if model_is_xl else "SD 1.5"
|
| 340 |
+
incompatible_vae = (model_is_xl and vae_model and not sdxl_in_vae) or (not model_is_xl and sdxl_in_vae)
|
| 341 |
+
|
| 342 |
+
if incompatible_vae:
|
| 343 |
+
vae_model = None
|
| 344 |
+
|
| 345 |
+
|
| 346 |
+
self.model.load_pipe(
|
| 347 |
+
model_name,
|
| 348 |
+
task_name=task_stablepy[task],
|
| 349 |
+
vae_model=vae_model if vae_model != "None" else None,
|
| 350 |
+
type_model_precision=torch.float16,
|
| 351 |
+
retain_task_model_in_cache=False,
|
| 352 |
+
)
|
| 353 |
+
yield f"Model loaded: {model_name} {vae_model if vae_model else ''}"
|
| 354 |
+
|
| 355 |
+
@spaces.GPU
|
| 356 |
def generate_pipeline(
|
| 357 |
self,
|
| 358 |
prompt,
|
|
|
|
| 445 |
mask_dilation_b,
|
| 446 |
mask_blur_b,
|
| 447 |
mask_padding_b,
|
| 448 |
+
retain_task_cache_gui,
|
| 449 |
+
image_ip1,
|
| 450 |
+
mask_ip1,
|
| 451 |
+
model_ip1,
|
| 452 |
+
mode_ip1,
|
| 453 |
+
scale_ip1,
|
| 454 |
+
image_ip2,
|
| 455 |
+
mask_ip2,
|
| 456 |
+
model_ip2,
|
| 457 |
+
mode_ip2,
|
| 458 |
+
scale_ip2,
|
| 459 |
+
# progress=gr.Progress(track_tqdm=True),
|
| 460 |
+
# progress=gr.Progress()
|
| 461 |
):
|
| 462 |
|
| 463 |
+
# progress(0.01, desc="Loading model...")
|
| 464 |
+
|
| 465 |
vae_model = vae_model if vae_model != "None" else None
|
| 466 |
loras_list = [lora1, lora2, lora3, lora4, lora5]
|
| 467 |
|
|
|
|
| 481 |
|
| 482 |
for la in loras_list:
|
| 483 |
if la is not None and la != "None":
|
| 484 |
+
print(la)
|
| 485 |
+
lora_type = ("animetarot" in la.lower() or "Hyper-SD15-8steps".lower() in la.lower())
|
| 486 |
if (model_is_xl and lora_type) or (not model_is_xl and not lora_type):
|
| 487 |
gr.Info(f"The LoRA {la} is for { 'SD 1.5' if model_is_xl else 'SDXL' }, but you are using { model_type }.")
|
| 488 |
|
| 489 |
task = task_stablepy[task]
|
| 490 |
|
| 491 |
+
params_ip_img = []
|
| 492 |
+
params_ip_msk = []
|
| 493 |
+
params_ip_model = []
|
| 494 |
+
params_ip_mode = []
|
| 495 |
+
params_ip_scale = []
|
| 496 |
+
|
| 497 |
+
all_adapters = [
|
| 498 |
+
(image_ip1, mask_ip1, model_ip1, mode_ip1, scale_ip1),
|
| 499 |
+
(image_ip2, mask_ip2, model_ip2, mode_ip2, scale_ip2),
|
| 500 |
+
]
|
| 501 |
+
|
| 502 |
+
for imgip, mskip, modelip, modeip, scaleip in all_adapters:
|
| 503 |
+
if imgip:
|
| 504 |
+
params_ip_img.append(imgip)
|
| 505 |
+
if mskip:
|
| 506 |
+
params_ip_msk.append(mskip)
|
| 507 |
+
params_ip_model.append(modelip)
|
| 508 |
+
params_ip_mode.append(modeip)
|
| 509 |
+
params_ip_scale.append(scaleip)
|
| 510 |
+
|
| 511 |
# First load
|
| 512 |
model_precision = torch.float16
|
| 513 |
if not self.model:
|
| 514 |
+
from modelstream import Model_Diffusers2
|
| 515 |
|
| 516 |
print("Loading model...")
|
| 517 |
+
self.model = Model_Diffusers2(
|
| 518 |
base_model_id=model_name,
|
| 519 |
task_name=task,
|
| 520 |
vae_model=vae_model if vae_model != "None" else None,
|
| 521 |
+
type_model_precision=model_precision,
|
| 522 |
+
retain_task_model_in_cache=retain_task_cache_gui,
|
| 523 |
)
|
| 524 |
|
| 525 |
if task != "txt2img" and not image_control:
|
|
|
|
| 549 |
model_name,
|
| 550 |
task_name=task,
|
| 551 |
vae_model=vae_model if vae_model != "None" else None,
|
| 552 |
+
type_model_precision=model_precision,
|
| 553 |
+
retain_task_model_in_cache=retain_task_cache_gui,
|
| 554 |
)
|
| 555 |
|
| 556 |
if textual_inversion and self.model.class_name == "StableDiffusionXLPipeline":
|
|
|
|
| 652 |
"hires_negative_prompt": hires_negative_prompt,
|
| 653 |
"hires_sampler": hires_sampler,
|
| 654 |
"hires_before_adetailer": hires_before_adetailer,
|
| 655 |
+
"hires_after_adetailer": hires_after_adetailer,
|
| 656 |
+
"ip_adapter_image": params_ip_img,
|
| 657 |
+
"ip_adapter_mask": params_ip_msk,
|
| 658 |
+
"ip_adapter_model": params_ip_model,
|
| 659 |
+
"ip_adapter_mode": params_ip_mode,
|
| 660 |
+
"ip_adapter_scale": params_ip_scale,
|
| 661 |
}
|
| 662 |
|
| 663 |
# print(pipe_params)
|
| 664 |
|
| 665 |
+
random_number = random.randint(1, 100)
|
| 666 |
+
if random_number < 25 and num_images < 3:
|
| 667 |
+
if not upscaler_model and steps < 45 and task in ["txt2img", "img2img"] and not adetailer_active_a and not adetailer_active_b:
|
| 668 |
+
num_images *=2
|
| 669 |
+
pipe_params["num_images"] = num_images
|
| 670 |
+
gr.Info("Num images x 2 🎉")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 671 |
|
| 672 |
+
# print("Inference 1")
|
| 673 |
+
# yield self.infer_short(self.model, pipe_params)
|
| 674 |
+
for img, seed, data in self.model(**pipe_params):
|
| 675 |
+
info_state = f"PROCESSING..."
|
| 676 |
+
if data:
|
| 677 |
+
info_state = f"COMPLETE: seeds={str(seed)}"
|
| 678 |
+
yield img, info_state
|
| 679 |
|
| 680 |
|
| 681 |
sd_gen = GuiSD()
|
|
|
|
| 685 |
#component-0 { height: 100%; }
|
| 686 |
#gallery { flex-grow: 1; }
|
| 687 |
"""
|
| 688 |
+
sdxl_task = [k for k, v in task_stablepy.items() if v in SDXL_TASKS ]
|
| 689 |
+
sd_task = [k for k, v in task_stablepy.items() if v in SD15_TASKS ]
|
|
|
|
|
|
|
|
|
|
| 690 |
def update_task_options(model_name, task_name):
|
| 691 |
if model_name in model_list:
|
| 692 |
if "xl" in model_name.lower():
|
|
|
|
| 718 |
model_name_gui = gr.Dropdown(label="Model", choices=model_list, value=model_list[0], allow_custom_value=True)
|
| 719 |
prompt_gui = gr.Textbox(lines=5, placeholder="Enter prompt", label="Prompt")
|
| 720 |
neg_prompt_gui = gr.Textbox(lines=3, placeholder="Enter Neg prompt", label="Negative prompt")
|
| 721 |
+
with gr.Row(equal_height=False):
|
| 722 |
+
set_params_gui = gr.Button(value="↙️")
|
| 723 |
+
clear_prompt_gui = gr.Button(value="🗑️")
|
| 724 |
+
set_random_seed = gr.Button(value="🎲")
|
| 725 |
generate_button = gr.Button(value="GENERATE", variant="primary")
|
| 726 |
+
|
| 727 |
model_name_gui.change(
|
| 728 |
update_task_options,
|
| 729 |
[model_name_gui, task_gui],
|
| 730 |
[task_gui],
|
| 731 |
)
|
| 732 |
|
| 733 |
+
load_model_gui = gr.HTML()
|
| 734 |
+
|
| 735 |
result_images = gr.Gallery(
|
| 736 |
label="Generated images",
|
| 737 |
show_label=False,
|
|
|
|
| 745 |
selected_index=50,
|
| 746 |
)
|
| 747 |
|
| 748 |
+
actual_task_info = gr.HTML()
|
| 749 |
+
|
| 750 |
with gr.Column(scale=1):
|
| 751 |
steps_gui = gr.Slider(minimum=1, maximum=100, step=1, value=30, label="Steps")
|
| 752 |
cfg_gui = gr.Slider(minimum=0, maximum=30, step=0.5, value=7.5, label="CFG")
|
| 753 |
sampler_gui = gr.Dropdown(label="Sampler", choices=scheduler_names, value="Euler a")
|
| 754 |
img_width_gui = gr.Slider(minimum=64, maximum=4096, step=8, value=1024, label="Img Width")
|
| 755 |
img_height_gui = gr.Slider(minimum=64, maximum=4096, step=8, value=1024, label="Img Height")
|
|
|
|
|
|
|
| 756 |
seed_gui = gr.Number(minimum=-1, maximum=9999999999, value=-1, label="Seed")
|
| 757 |
+
with gr.Row():
|
| 758 |
+
clip_skip_gui = gr.Checkbox(value=True, label="Layer 2 Clip Skip")
|
| 759 |
+
free_u_gui = gr.Checkbox(value=True, label="FreeU")
|
| 760 |
+
|
| 761 |
+
with gr.Row(equal_height=False):
|
| 762 |
+
|
| 763 |
+
|
| 764 |
+
|
| 765 |
+
def run_set_params_gui(base_prompt):
|
| 766 |
+
valid_receptors = { # default values
|
| 767 |
+
"prompt": gr.update(value=base_prompt),
|
| 768 |
+
"neg_prompt": gr.update(value=""),
|
| 769 |
+
"Steps": gr.update(value=30),
|
| 770 |
+
"width": gr.update(value=1024),
|
| 771 |
+
"height": gr.update(value=1024),
|
| 772 |
+
"Seed": gr.update(value=-1),
|
| 773 |
+
"Sampler": gr.update(value="Euler a"),
|
| 774 |
+
"scale": gr.update(value=7.5), # cfg
|
| 775 |
+
"skip": gr.update(value=True),
|
| 776 |
+
}
|
| 777 |
+
valid_keys = list(valid_receptors.keys())
|
| 778 |
+
|
| 779 |
+
parameters = extract_parameters(base_prompt)
|
| 780 |
+
for key, val in parameters.items():
|
| 781 |
+
# print(val)
|
| 782 |
+
if key in valid_keys:
|
| 783 |
+
if key == "Sampler":
|
| 784 |
+
if val not in scheduler_names:
|
| 785 |
+
continue
|
| 786 |
+
elif key == "skip":
|
| 787 |
+
if int(val) >= 2:
|
| 788 |
+
val = True
|
| 789 |
+
if key == "prompt":
|
| 790 |
+
if ">" in val and "<" in val:
|
| 791 |
+
val = re.sub(r'<[^>]+>', '', val)
|
| 792 |
+
print("Removed LoRA written in the prompt")
|
| 793 |
+
if key in ["prompt", "neg_prompt"]:
|
| 794 |
+
val = val.strip()
|
| 795 |
+
if key in ["Steps", "width", "height", "Seed"]:
|
| 796 |
+
val = int(val)
|
| 797 |
+
if key == "scale":
|
| 798 |
+
val = float(val)
|
| 799 |
+
if key == "Seed":
|
| 800 |
+
continue
|
| 801 |
+
valid_receptors[key] = gr.update(value=val)
|
| 802 |
+
# print(val, type(val))
|
| 803 |
+
# print(valid_receptors)
|
| 804 |
+
return [value for value in valid_receptors.values()]
|
| 805 |
+
|
| 806 |
+
set_params_gui.click(
|
| 807 |
+
run_set_params_gui, [prompt_gui],[
|
| 808 |
+
prompt_gui,
|
| 809 |
+
neg_prompt_gui,
|
| 810 |
+
steps_gui,
|
| 811 |
+
img_width_gui,
|
| 812 |
+
img_height_gui,
|
| 813 |
+
seed_gui,
|
| 814 |
+
sampler_gui,
|
| 815 |
+
cfg_gui,
|
| 816 |
+
clip_skip_gui,
|
| 817 |
+
],
|
| 818 |
+
)
|
| 819 |
+
|
| 820 |
+
|
| 821 |
+
def run_clear_prompt_gui():
|
| 822 |
+
return gr.update(value=""), gr.update(value="")
|
| 823 |
+
clear_prompt_gui.click(
|
| 824 |
+
run_clear_prompt_gui, [], [prompt_gui, neg_prompt_gui]
|
| 825 |
+
)
|
| 826 |
+
|
| 827 |
+
def run_set_random_seed():
|
| 828 |
+
return -1
|
| 829 |
+
set_random_seed.click(
|
| 830 |
+
run_set_random_seed, [], seed_gui
|
| 831 |
+
)
|
| 832 |
+
|
| 833 |
num_images_gui = gr.Slider(minimum=1, maximum=4, step=1, value=1, label="Images")
|
| 834 |
+
prompt_s_options = [
|
| 835 |
+
("Compel format: (word)weight", "Compel"),
|
| 836 |
+
("Classic format: (word:weight)", "Classic"),
|
| 837 |
+
("Classic-original format: (word:weight)", "Classic-original"),
|
| 838 |
+
("Classic-no_norm format: (word:weight)", "Classic-no_norm"),
|
| 839 |
+
("Classic-ignore", "Classic-ignore"),
|
| 840 |
+
("None", "None"),
|
| 841 |
+
]
|
| 842 |
prompt_syntax_gui = gr.Dropdown(label="Prompt Syntax", choices=prompt_s_options, value=prompt_s_options[0][1])
|
| 843 |
vae_model_gui = gr.Dropdown(label="VAE Model", choices=vae_model_list)
|
| 844 |
+
|
| 845 |
+
with gr.Accordion("Hires fix", open=False, visible=True):
|
| 846 |
+
|
| 847 |
+
upscaler_keys = list(upscaler_dict_gui.keys())
|
| 848 |
+
|
| 849 |
+
upscaler_model_path_gui = gr.Dropdown(label="Upscaler", choices=upscaler_keys, value=upscaler_keys[0])
|
| 850 |
+
upscaler_increases_size_gui = gr.Slider(minimum=1.1, maximum=6., step=0.1, value=1.4, label="Upscale by")
|
| 851 |
+
esrgan_tile_gui = gr.Slider(minimum=0, value=100, maximum=500, step=1, label="ESRGAN Tile")
|
| 852 |
+
esrgan_tile_overlap_gui = gr.Slider(minimum=1, maximum=200, step=1, value=10, label="ESRGAN Tile Overlap")
|
| 853 |
+
hires_steps_gui = gr.Slider(minimum=0, value=30, maximum=100, step=1, label="Hires Steps")
|
| 854 |
+
hires_denoising_strength_gui = gr.Slider(minimum=0.1, maximum=1.0, step=0.01, value=0.55, label="Hires Denoising Strength")
|
| 855 |
+
hires_sampler_gui = gr.Dropdown(label="Hires Sampler", choices=["Use same sampler"] + scheduler_names[:-1], value="Use same sampler")
|
| 856 |
+
hires_prompt_gui = gr.Textbox(label="Hires Prompt", placeholder="Main prompt will be use", lines=3)
|
| 857 |
+
hires_negative_prompt_gui = gr.Textbox(label="Hires Negative Prompt", placeholder="Main negative prompt will be use", lines=3)
|
| 858 |
+
|
| 859 |
+
with gr.Accordion("LoRA", open=False, visible=True):
|
| 860 |
+
lora1_gui = gr.Dropdown(label="Lora1", choices=lora_model_list)
|
| 861 |
+
lora_scale_1_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 1")
|
| 862 |
+
lora2_gui = gr.Dropdown(label="Lora2", choices=lora_model_list)
|
| 863 |
+
lora_scale_2_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 2")
|
| 864 |
+
lora3_gui = gr.Dropdown(label="Lora3", choices=lora_model_list)
|
| 865 |
+
lora_scale_3_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 3")
|
| 866 |
+
lora4_gui = gr.Dropdown(label="Lora4", choices=lora_model_list)
|
| 867 |
+
lora_scale_4_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 4")
|
| 868 |
+
lora5_gui = gr.Dropdown(label="Lora5", choices=lora_model_list)
|
| 869 |
+
lora_scale_5_gui = gr.Slider(minimum=-2, maximum=2, step=0.01, value=0.33, label="Lora Scale 5")
|
| 870 |
+
|
| 871 |
+
with gr.Accordion("IP-Adapter", open=False, visible=True):##############
|
| 872 |
+
|
| 873 |
+
IP_MODELS = sorted(list(set(IP_ADAPTERS_SD + IP_ADAPTERS_SDXL)))
|
| 874 |
+
MODE_IP_OPTIONS = ["original", "style", "layout", "style+layout"]
|
| 875 |
+
|
| 876 |
+
with gr.Accordion("IP-Adapter 1", open=False, visible=True):
|
| 877 |
+
image_ip1 = gr.Image(label="IP Image", type="filepath")
|
| 878 |
+
mask_ip1 = gr.Image(label="IP Mask", type="filepath")
|
| 879 |
+
model_ip1 = gr.Dropdown(value="plus_face", label="Model", choices=IP_MODELS)
|
| 880 |
+
mode_ip1 = gr.Dropdown(value="original", label="Mode", choices=MODE_IP_OPTIONS)
|
| 881 |
+
scale_ip1 = gr.Slider(minimum=0., maximum=2., step=0.01, value=0.7, label="Scale")
|
| 882 |
+
with gr.Accordion("IP-Adapter 2", open=False, visible=True):
|
| 883 |
+
image_ip2 = gr.Image(label="IP Image", type="filepath")
|
| 884 |
+
mask_ip2 = gr.Image(label="IP Mask (optional)", type="filepath")
|
| 885 |
+
model_ip2 = gr.Dropdown(value="base", label="Model", choices=IP_MODELS)
|
| 886 |
+
mode_ip2 = gr.Dropdown(value="style", label="Mode", choices=MODE_IP_OPTIONS)
|
| 887 |
+
scale_ip2 = gr.Slider(minimum=0., maximum=2., step=0.01, value=0.7, label="Scale")
|
| 888 |
+
|
| 889 |
with gr.Accordion("ControlNet / Img2img / Inpaint", open=False, visible=True):
|
| 890 |
image_control = gr.Image(label="Image ControlNet/Inpaint/Img2img", type="filepath")
|
| 891 |
image_mask_gui = gr.Image(label="Image Mask", type="filepath")
|
|
|
|
| 895 |
)
|
| 896 |
image_resolution_gui = gr.Slider(minimum=64, maximum=2048, step=64, value=1024, label="Image Resolution")
|
| 897 |
preprocessor_name_gui = gr.Dropdown(label="Preprocessor Name", choices=preprocessor_controlnet["canny"])
|
| 898 |
+
|
| 899 |
def change_preprocessor_choices(task):
|
| 900 |
task = task_stablepy[task]
|
| 901 |
if task in preprocessor_controlnet.keys():
|
|
|
|
| 903 |
else:
|
| 904 |
choices_task = preprocessor_controlnet["canny"]
|
| 905 |
return gr.update(choices=choices_task, value=choices_task[0])
|
| 906 |
+
|
| 907 |
task_gui.change(
|
| 908 |
change_preprocessor_choices,
|
| 909 |
[task_gui],
|
|
|
|
| 923 |
adapter_conditioning_scale_gui = gr.Slider(minimum=0, maximum=5., step=0.1, value=1, label="Adapter Conditioning Scale")
|
| 924 |
adapter_conditioning_factor_gui = gr.Slider(minimum=0, maximum=1., step=0.01, value=0.55, label="Adapter Conditioning Factor (%)")
|
| 925 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 926 |
with gr.Accordion("Styles", open=False, visible=True):
|
| 927 |
+
|
| 928 |
try:
|
| 929 |
style_names_found = sd_gen.model.STYLE_NAMES
|
| 930 |
except:
|
| 931 |
style_names_found = STYLE_NAMES
|
| 932 |
+
|
| 933 |
style_prompt_gui = gr.Dropdown(
|
| 934 |
style_names_found,
|
| 935 |
multiselect=True,
|
|
|
|
| 944 |
if not sd_gen.model:
|
| 945 |
gr.Info("First load the model")
|
| 946 |
return gr.update(value=None, choices=STYLE_NAMES)
|
| 947 |
+
|
| 948 |
sd_gen.model.load_style_file(json)
|
| 949 |
gr.Info(f"{len(sd_gen.model.STYLE_NAMES)} styles loaded")
|
| 950 |
return gr.update(value=None, choices=sd_gen.model.STYLE_NAMES)
|
| 951 |
|
| 952 |
style_button.click(load_json_style_file, [style_json_gui], [style_prompt_gui])
|
| 953 |
+
|
| 954 |
with gr.Accordion("Textual inversion", open=False, visible=False):
|
| 955 |
active_textual_inversion_gui = gr.Checkbox(value=False, label="Active Textual Inversion in prompt")
|
| 956 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 957 |
with gr.Accordion("Detailfix", open=False, visible=True):
|
| 958 |
|
| 959 |
# Adetailer Inpaint Only
|
| 960 |
adetailer_inpaint_only_gui = gr.Checkbox(label="Inpaint only", value=True)
|
| 961 |
+
|
| 962 |
# Adetailer Verbose
|
| 963 |
adetailer_verbose_gui = gr.Checkbox(label="Verbose", value=False)
|
| 964 |
+
|
| 965 |
# Adetailer Sampler
|
| 966 |
adetailer_sampler_options = ["Use same sampler"] + scheduler_names[:-1]
|
| 967 |
adetailer_sampler_gui = gr.Dropdown(label="Adetailer sampler:", choices=adetailer_sampler_options, value="Use same sampler")
|
| 968 |
+
|
| 969 |
with gr.Accordion("Detailfix A", open=False, visible=True):
|
| 970 |
# Adetailer A
|
| 971 |
adetailer_active_a_gui = gr.Checkbox(label="Enable Adetailer A", value=False)
|
|
|
|
| 978 |
mask_dilation_a_gui = gr.Number(label="Mask dilation:", value=4, minimum=1)
|
| 979 |
mask_blur_a_gui = gr.Number(label="Mask blur:", value=4, minimum=1)
|
| 980 |
mask_padding_a_gui = gr.Number(label="Mask padding:", value=32, minimum=1)
|
| 981 |
+
|
| 982 |
with gr.Accordion("Detailfix B", open=False, visible=True):
|
| 983 |
# Adetailer B
|
| 984 |
adetailer_active_b_gui = gr.Checkbox(label="Enable Adetailer B", value=False)
|
|
|
|
| 993 |
mask_padding_b_gui = gr.Number(label="Mask padding:", value=32, minimum=1)
|
| 994 |
|
| 995 |
with gr.Accordion("Other settings", open=False, visible=True):
|
| 996 |
+
image_previews_gui = gr.Checkbox(value=True, label="Image Previews")
|
| 997 |
hires_before_adetailer_gui = gr.Checkbox(value=False, label="Hires Before Adetailer")
|
| 998 |
hires_after_adetailer_gui = gr.Checkbox(value=True, label="Hires After Adetailer")
|
| 999 |
generator_in_cpu_gui = gr.Checkbox(value=False, label="Generator in CPU")
|
| 1000 |
|
| 1001 |
with gr.Accordion("More settings", open=False, visible=False):
|
| 1002 |
loop_generation_gui = gr.Slider(minimum=1, value=1, label="Loop Generation")
|
| 1003 |
+
retain_task_cache_gui = gr.Checkbox(value=False, label="Retain task model in cache")
|
| 1004 |
leave_progress_bar_gui = gr.Checkbox(value=True, label="Leave Progress Bar")
|
| 1005 |
disable_progress_bar_gui = gr.Checkbox(value=False, label="Disable Progress Bar")
|
| 1006 |
+
display_images_gui = gr.Checkbox(value=True, label="Display Images")
|
|
|
|
| 1007 |
save_generated_images_gui = gr.Checkbox(value=False, label="Save Generated Images")
|
| 1008 |
image_storage_location_gui = gr.Textbox(value="./images", label="Image Storage Location")
|
| 1009 |
retain_compel_previous_load_gui = gr.Checkbox(value=False, label="Retain Compel Previous Load")
|
|
|
|
| 1123 |
1024,
|
| 1124 |
"misri/epicrealismXL_v7FinalDestination",
|
| 1125 |
None, # vae
|
| 1126 |
+
"canny ControlNet",
|
| 1127 |
"image.webp", # img conttol
|
| 1128 |
"Canny", # preprocessor
|
| 1129 |
1024, # preproc resolution
|
|
|
|
| 1252 |
512,
|
| 1253 |
"digiplay/majicMIX_realistic_v7",
|
| 1254 |
None, # vae
|
| 1255 |
+
"openpose ControlNet",
|
| 1256 |
"image.webp", # img conttol
|
| 1257 |
"Canny", # preprocessor
|
| 1258 |
512, # preproc resolution
|
|
|
|
| 1361 |
brush=gr.Brush(
|
| 1362 |
default_size="16", # or leave it as 'auto'
|
| 1363 |
color_mode="fixed", # 'fixed' hides the user swatches and colorpicker, 'defaults' shows it
|
| 1364 |
+
# default_color="black", # html names are supported
|
| 1365 |
colors=[
|
| 1366 |
"rgba(0, 0, 0, 1)", # rgb(a)
|
| 1367 |
"rgba(0, 0, 0, 0.1)",
|
|
|
|
| 1385 |
btn_send.click(send_img, [img_source, img_result], [image_control, image_mask_gui])
|
| 1386 |
|
| 1387 |
generate_button.click(
|
| 1388 |
+
fn=sd_gen.load_new_model,
|
| 1389 |
+
inputs=[
|
| 1390 |
+
model_name_gui,
|
| 1391 |
+
vae_model_gui,
|
| 1392 |
+
task_gui
|
| 1393 |
+
],
|
| 1394 |
+
outputs=[load_model_gui],
|
| 1395 |
+
queue=True,
|
| 1396 |
+
show_progress="minimal",
|
| 1397 |
+
).success(
|
| 1398 |
fn=sd_gen.generate_pipeline,
|
| 1399 |
inputs=[
|
| 1400 |
prompt_gui,
|
|
|
|
| 1487 |
mask_dilation_b_gui,
|
| 1488 |
mask_blur_b_gui,
|
| 1489 |
mask_padding_b_gui,
|
| 1490 |
+
retain_task_cache_gui,
|
| 1491 |
+
image_ip1,
|
| 1492 |
+
mask_ip1,
|
| 1493 |
+
model_ip1,
|
| 1494 |
+
mode_ip1,
|
| 1495 |
+
scale_ip1,
|
| 1496 |
+
image_ip2,
|
| 1497 |
+
mask_ip2,
|
| 1498 |
+
model_ip2,
|
| 1499 |
+
mode_ip2,
|
| 1500 |
+
scale_ip2,
|
| 1501 |
],
|
| 1502 |
+
outputs=[result_images, actual_task_info],
|
| 1503 |
queue=True,
|
| 1504 |
+
show_progress="minimal",
|
| 1505 |
)
|
| 1506 |
|
| 1507 |
app.queue()
|