aboutsummaryrefslogtreecommitdiff
path: root/modules/ui.py
diff options
context:
space:
mode:
authorbrkirch <brkirch@users.noreply.github.com>2022-10-04 07:42:53 -0400
committerGitHub <noreply@github.com>2022-10-04 07:42:53 -0400
commite9e2a7ec9ac704f133f586eb34176e388c93c87c (patch)
tree969e0e595bd36987ae9de9ae302085ef555bba15 /modules/ui.py
parentdc9c5a97742e3a34d37da7108642d8adc0dc5858 (diff)
parentd5bba20a58f43a9f984bb67b4e17f48661f6b818 (diff)
Merge branch 'master' into cpu-cmdline-opt
Diffstat (limited to 'modules/ui.py')
-rw-r--r--modules/ui.py12
1 files changed, 10 insertions, 2 deletions
diff --git a/modules/ui.py b/modules/ui.py
index 55f7aa95..20dc8c37 100644
--- a/modules/ui.py
+++ b/modules/ui.py
@@ -386,14 +386,22 @@ def connect_reuse_seed(seed: gr.Number, reuse_seed: gr.Button, generation_info:
outputs=[seed, dummy_component]
)
+
def update_token_counter(text, steps):
- prompt_schedules = get_learned_conditioning_prompt_schedules([text], steps)
+ try:
+ prompt_schedules = get_learned_conditioning_prompt_schedules([text], steps)
+ except Exception:
+ # a parsing error can happen here during typing, and we don't want to bother the user with
+ # messages related to it in console
+ prompt_schedules = [[[steps, text]]]
+
flat_prompts = reduce(lambda list1, list2: list1+list2, prompt_schedules)
- prompts = [prompt_text for step,prompt_text in flat_prompts]
+ prompts = [prompt_text for step, prompt_text in flat_prompts]
tokens, token_count, max_length = max([model_hijack.tokenize(prompt) for prompt in prompts], key=lambda args: args[1])
style_class = ' class="red"' if (token_count > max_length) else ""
return f"<span {style_class}>{token_count}/{max_length}</span>"
+
def create_toprow(is_img2img):
id_part = "img2img" if is_img2img else "txt2img"