From 60e95f1d8c4e7296803389a46f542bd6e0a02770 Mon Sep 17 00:00:00 2001 From: AUTOMATIC <16777216c@gmail.com> Date: Tue, 23 Aug 2022 11:58:50 +0300 Subject: silence the warning from transformers add feature demonstrations to readme --- README.md | 48 +++++++++++++++++++++++++++++++++++++++++++++--- 1 file changed, 45 insertions(+), 3 deletions(-) (limited to 'README.md') diff --git a/README.md b/README.md index 41730012..52a4f1bf 100644 --- a/README.md +++ b/README.md @@ -4,8 +4,9 @@ A browser interface based on Gradio library for Stable Diffusion. Original script with Gradio UI was written by a kind anonymopus user. This is a modification. ![](screenshot.png) +## Installing and running -## Stable Diffusion +### Stable Diffusion This script assumes that you already have main Stable Diffusion sutff installed, assumed to be in directory `/sd`. If you don't have it installed, follow the guide: @@ -21,7 +22,7 @@ Particularly, following files must exist: - `/sd/ldm/util.py` - `/sd/k_diffusion/__init__.py` -## GFPGAN +### GFPGAN If you want to use GFPGAN to improve generated faces, you need to install it separately. Follow instructions from https://github.com/TencentARC/GFPGAN, but when cloning it, do so into Stable Diffusion main directory, `/sd`. @@ -37,7 +38,7 @@ The following files must exist: If the GFPGAN directory does not exist, you will not get the option to use GFPGAN in the UI. If it does exist, you will either be able to use it, or there will be a message in console with an error related to GFPGAN. -## Web UI +### Web UI Run the script as: @@ -56,3 +57,44 @@ Running on local URL: http://127.0.0.1:7860/ ``` Open the URL in browser, and you are good to go. + +## Features +The script creates a web UI for Stable Diffusion's txt2img and img2img scripts. Following are features added +that are not in original script. + +### GFPGAN +Lets you improve faces in pictures using the GFPGAN model. There is a checkbox in every tab to use GFPGAN at 100%, and +also a separate tab that just allows you to use GFPGAN on any picture, with a slider that controls how strongthe effect is. + +![](images/GFPGAN.png) + +### Sampling method selection +Pick out of three sampling methods for txt2img: DDIM, PLMS, k-diffusion: + +![](images/sampling.png) + +### Prompt matrix +Separate multiple prompts using the `|` character, and the system will produce an image for every combination of them. +For example, if you use `a house in a field of grass|at dawn|illustration` prompt, there are four combinations possible (first part of prompt is always kept): + +- `a house in a field of grass` +- `a house in a field of grass, at dawn` +- `a house in a field of grass, illustration` +- `a house in a field of grass, at dawn, illustration` + +Four images will be produced, in this order, all with same seed and each with corresponding prompt: + +![](images/prompt-matrix.png) + +### Flagging +Click the Flag button under the output section, and generated images will be saved to `log/images` directory, and generation parameters +will be appended to a csv file `log/log.csv` in the `/sd` directory. + +### Copy-paste generation parameters +A text output provides generation parameters in an easy to copy-paste form for easy sharing. + +![](images/kopipe.png) + +### Correct seeds for batches +If you use a seed of 1000 to generate two batches of two images each, four generated images will have seeds: `1000, 1001, 1002, 1003`. +Previous versions of the UI would produce `1000, x, 1001, x`, where x is an iamge that can't be generated by any seed. -- cgit v1.2.1