diff options
| author | Volpeon <git@volpeon.ink> | 2023-03-28 16:28:44 +0200 |
|---|---|---|
| committer | Volpeon <git@volpeon.ink> | 2023-03-28 16:28:44 +0200 |
| commit | b59611506d2eac1b280ada07db0bf9de15b41a26 (patch) | |
| tree | 3f4a1e49f0db72443e45bc64d2c5b9d99ce055fc | |
| parent | Support num_train_steps arg again (diff) | |
| download | textual-inversion-diff-b59611506d2eac1b280ada07db0bf9de15b41a26.tar.gz textual-inversion-diff-b59611506d2eac1b280ada07db0bf9de15b41a26.tar.bz2 textual-inversion-diff-b59611506d2eac1b280ada07db0bf9de15b41a26.zip | |
Fix
| -rw-r--r-- | train_dreambooth.py | 2 | ||||
| -rw-r--r-- | train_lora.py | 2 | ||||
| -rw-r--r-- | train_ti.py | 2 |
3 files changed, 3 insertions, 3 deletions
diff --git a/train_dreambooth.py b/train_dreambooth.py index acb8287..e3c8525 100644 --- a/train_dreambooth.py +++ b/train_dreambooth.py | |||
| @@ -598,7 +598,7 @@ def main(): | |||
| 598 | num_train_epochs = args.num_train_epochs | 598 | num_train_epochs = args.num_train_epochs |
| 599 | 599 | ||
| 600 | if num_train_epochs is None: | 600 | if num_train_epochs is None: |
| 601 | num_train_epochs = math.ceil(len(datamodule.train_dataset) / args.num_train_steps) | 601 | num_train_epochs = math.ceil(args.num_train_steps / len(datamodule.train_dataset)) |
| 602 | 602 | ||
| 603 | params_to_optimize = (unet.parameters(), ) | 603 | params_to_optimize = (unet.parameters(), ) |
| 604 | if args.train_text_encoder_epochs != 0: | 604 | if args.train_text_encoder_epochs != 0: |
diff --git a/train_lora.py b/train_lora.py index a9c6e52..6f8644b 100644 --- a/train_lora.py +++ b/train_lora.py | |||
| @@ -630,7 +630,7 @@ def main(): | |||
| 630 | num_train_epochs = args.num_train_epochs | 630 | num_train_epochs = args.num_train_epochs |
| 631 | 631 | ||
| 632 | if num_train_epochs is None: | 632 | if num_train_epochs is None: |
| 633 | num_train_epochs = math.ceil(len(datamodule.train_dataset) / args.num_train_steps) | 633 | num_train_epochs = math.ceil(args.num_train_steps / len(datamodule.train_dataset)) |
| 634 | 634 | ||
| 635 | optimizer = create_optimizer( | 635 | optimizer = create_optimizer( |
| 636 | itertools.chain( | 636 | itertools.chain( |
diff --git a/train_ti.py b/train_ti.py index 7bcc72f..9c4ad93 100644 --- a/train_ti.py +++ b/train_ti.py | |||
| @@ -761,7 +761,7 @@ def main(): | |||
| 761 | num_train_epochs = args.num_train_epochs | 761 | num_train_epochs = args.num_train_epochs |
| 762 | 762 | ||
| 763 | if num_train_epochs is None: | 763 | if num_train_epochs is None: |
| 764 | num_train_epochs = math.ceil(len(datamodule.train_dataset) / args.num_train_steps) | 764 | num_train_epochs = math.ceil(args.num_train_steps / len(datamodule.train_dataset)) |
| 765 | 765 | ||
| 766 | optimizer = create_optimizer( | 766 | optimizer = create_optimizer( |
| 767 | text_encoder.text_model.embeddings.temp_token_embedding.parameters(), | 767 | text_encoder.text_model.embeddings.temp_token_embedding.parameters(), |
