aboutsummaryrefslogtreecommitdiff
path: root/modules/shared.py
diff options
context:
space:
mode:
authorflamelaw <flamelaw.com3d2@gmail.com>2022-11-20 12:35:26 +0900
committerflamelaw <flamelaw.com3d2@gmail.com>2022-11-20 12:35:26 +0900
commitbd68e35de3b7cf7547ed97d8bdf60147402133cc (patch)
tree248d3810f1f06530f19c9db7e8b0b6d2e22f7225 /modules/shared.py
parent47a44c7e421b98ca07e92dbf88769b04c9e28f86 (diff)
Gradient accumulation, autocast fix, new latent sampling method, etc
Diffstat (limited to 'modules/shared.py')
-rw-r--r--modules/shared.py3
1 files changed, 1 insertions, 2 deletions
diff --git a/modules/shared.py b/modules/shared.py
index a4457305..3704ce23 100644
--- a/modules/shared.py
+++ b/modules/shared.py
@@ -322,8 +322,7 @@ options_templates.update(options_section(('system', "System"), {
options_templates.update(options_section(('training', "Training"), {
"unload_models_when_training": OptionInfo(False, "Move VAE and CLIP to RAM when training if possible. Saves VRAM."),
- "shuffle_tags": OptionInfo(False, "Shuffleing tags by ',' when create texts."),
- "tag_drop_out": OptionInfo(0, "Dropout tags when create texts", gr.Slider, {"minimum": 0, "maximum": 1, "step": 0.1}),
+ "pin_memory": OptionInfo(False, "Turn on pin_memory for DataLoader. Makes training slightly faster but can increase memory usage."),
"save_optimizer_state": OptionInfo(False, "Saves Optimizer state as separate *.optim file. Training can be resumed with HN itself and matching optim file."),
"dataset_filename_word_regex": OptionInfo("", "Filename word regex"),
"dataset_filename_join_string": OptionInfo(" ", "Filename join string"),