2022-09-12 10:13:03 -06:00
|
|
|
import math
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
import traceback
|
|
|
|
|
|
|
|
import modules.scripts as scripts
|
|
|
|
import gradio as gr
|
|
|
|
|
|
|
|
from modules.processing import Processed, process_images
|
|
|
|
from PIL import Image
|
|
|
|
from modules.shared import opts, cmd_opts, state
|
|
|
|
|
|
|
|
class Script(scripts.Script):
|
|
|
|
def title(self):
|
2022-09-17 02:34:33 -06:00
|
|
|
return "Prompts from file or textbox"
|
2022-09-12 10:13:03 -06:00
|
|
|
|
|
|
|
def ui(self, is_img2img):
|
2022-09-17 02:34:33 -06:00
|
|
|
# This checkbox would look nicer as two tabs, but there are two problems:
|
|
|
|
# 1) There is a bug in Gradio 3.3 that prevents visibility from working on Tabs
|
|
|
|
# 2) Even with Gradio 3.3.1, returning a control (like Tabs) that can't be used as input
|
|
|
|
# causes a AttributeError: 'Tabs' object has no attribute 'preprocess' assert,
|
|
|
|
# due to the way Script assumes all controls returned can be used as inputs.
|
|
|
|
# Therefore, there's no good way to use grouping components right now,
|
|
|
|
# so we will use a checkbox! :)
|
|
|
|
checkbox_txt = gr.Checkbox(label="Show Textbox", value=False)
|
2022-09-12 10:13:03 -06:00
|
|
|
file = gr.File(label="File with inputs", type='bytes')
|
2022-09-17 02:34:33 -06:00
|
|
|
prompt_txt = gr.TextArea(label="Prompts")
|
|
|
|
checkbox_txt.change(fn=lambda x: [gr.File.update(visible = not x), gr.TextArea.update(visible = x)], inputs=[checkbox_txt], outputs=[file, prompt_txt])
|
|
|
|
return [checkbox_txt, file, prompt_txt]
|
|
|
|
|
2022-10-01 08:43:24 -06:00
|
|
|
def process_string_tag(self, tag):
|
|
|
|
return tag[1:-2]
|
|
|
|
|
|
|
|
def process_int_tag(self, tag):
|
|
|
|
return int(tag)
|
|
|
|
|
|
|
|
def process_float_tag(self, tag):
|
|
|
|
return float(tag)
|
|
|
|
|
|
|
|
def process_boolean_tag(self, tag):
|
|
|
|
return True if (tag == "true") else False
|
|
|
|
|
|
|
|
prompt_tags = {
|
|
|
|
"sd_model": None,
|
|
|
|
"outpath_samples": process_string_tag,
|
|
|
|
"outpath_grids": process_string_tag,
|
|
|
|
"prompt_for_display": process_string_tag,
|
|
|
|
"prompt": process_string_tag,
|
|
|
|
"negative_prompt": process_string_tag,
|
|
|
|
"styles": process_string_tag,
|
|
|
|
"seed": process_int_tag,
|
|
|
|
"subseed_strength": process_float_tag,
|
|
|
|
"subseed": process_int_tag,
|
|
|
|
"seed_resize_from_h": process_int_tag,
|
|
|
|
"seed_resize_from_w": process_int_tag,
|
|
|
|
"sampler_index": process_int_tag,
|
|
|
|
"batch_size": process_int_tag,
|
|
|
|
"n_iter": process_int_tag,
|
|
|
|
"steps": process_int_tag,
|
|
|
|
"cfg_scale": process_float_tag,
|
|
|
|
"width": process_int_tag,
|
|
|
|
"height": process_int_tag,
|
|
|
|
"restore_faces": process_boolean_tag,
|
|
|
|
"tiling": process_boolean_tag,
|
|
|
|
"do_not_save_samples": process_boolean_tag,
|
|
|
|
"do_not_save_grid": process_boolean_tag
|
|
|
|
}
|
|
|
|
|
2022-10-07 18:46:39 -06:00
|
|
|
def on_show(self, checkbox_txt, file, prompt_txt):
|
|
|
|
return [ gr.Checkbox.update(visible = True), gr.File.update(visible = not checkbox_txt), gr.TextArea.update(visible = checkbox_txt) ]
|
|
|
|
|
2022-09-17 02:34:33 -06:00
|
|
|
def run(self, p, checkbox_txt, data: bytes, prompt_txt: str):
|
|
|
|
if (checkbox_txt):
|
|
|
|
lines = [x.strip() for x in prompt_txt.splitlines()]
|
|
|
|
else:
|
|
|
|
lines = [x.strip() for x in data.decode('utf8', errors='ignore').split("\n")]
|
2022-09-12 10:13:03 -06:00
|
|
|
lines = [x for x in lines if len(x) > 0]
|
|
|
|
|
2022-09-17 02:34:33 -06:00
|
|
|
img_count = len(lines) * p.n_iter
|
|
|
|
batch_count = math.ceil(img_count / p.batch_size)
|
|
|
|
loop_count = math.ceil(batch_count / p.n_iter)
|
2022-10-01 08:43:24 -06:00
|
|
|
# These numbers no longer accurately reflect the total images and number of batches
|
2022-09-17 02:34:33 -06:00
|
|
|
print(f"Will process {img_count} images in {batch_count} batches.")
|
2022-09-12 10:13:03 -06:00
|
|
|
|
|
|
|
p.do_not_save_grid = True
|
|
|
|
|
|
|
|
state.job_count = batch_count
|
|
|
|
|
|
|
|
images = []
|
2022-09-17 02:34:33 -06:00
|
|
|
for loop_no in range(loop_count):
|
|
|
|
state.job = f"{loop_no + 1} out of {loop_count}"
|
2022-10-01 08:43:24 -06:00
|
|
|
# The following line may need revising to remove batch_size references
|
|
|
|
current_line = lines[loop_no*p.batch_size:(loop_no+1)*p.batch_size] * p.n_iter
|
2022-10-01 08:57:29 -06:00
|
|
|
|
|
|
|
# If the current line has no tags, parse the whole line as a prompt, else parse each tag
|
2022-10-01 08:43:24 -06:00
|
|
|
if(current_line[0][:2] != "--"):
|
|
|
|
p.prompt = current_line
|
|
|
|
else:
|
|
|
|
tokenized_line = current_line[0].split("--")
|
|
|
|
|
|
|
|
for tag in tokenized_line:
|
|
|
|
tag_split = tag.split(" ", 1)
|
|
|
|
if(tag_split[0] != ''):
|
|
|
|
value_func = self.prompt_tags.get(tag_split[0], None)
|
|
|
|
if(value_func != None):
|
|
|
|
value = value_func(self, tag_split[1])
|
|
|
|
setattr(p, tag_split[0], value)
|
|
|
|
else:
|
|
|
|
print(f"Unknown option \"{tag_split}\"")
|
|
|
|
|
2022-09-12 10:13:03 -06:00
|
|
|
proc = process_images(p)
|
|
|
|
images += proc.images
|
|
|
|
|
|
|
|
return Processed(p, images, p.seed, "")
|