-
-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
v23.1.0 #2219
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
* 1st comit * BLIP2 captioning implementation
* Update logging for validate_requirements.py * dedupe return torch verison
Add support for ip_noise_gamma, ip_noise_gamma_random_strength and noise_offset_random_strength
…noise_gamma_random_strength
* ROCm installer * Update requirements_linux_rocm.txt
* chore(docker): enhance build process with apt caching - Cache `apt-get` update in Docker build to speed up subsequent builds - Avoid cleaning `apt-get` cache and removing /var/lib/apt/lists/* after each install to take advantage of caching - Add arguments for TARGETARCH and TARGETVARIANT - Utilize cache mounting for pillow-simd installation in x86 platform - Remove manual cleaning of `apt-get` and `/var/lib/apt/lists/*` after installing runtime dependencies Signed-off-by: 陳鈞 <[email protected]> * chore(docker): Fix bitsandbytes can't find cuda - Set work directory to /tmp - Defined several environment variables for CUDA version 12.1.1 - Install the CUDA keyring - Added a process for partial installation of CUDA to minimize Docker image size - Added /usr/local/cuda/lib and /usr/local/cuda/lib64 to PATH and LD_LIBRARY_PATH environment variables Signed-off-by: 陳鈞 <[email protected]> * chore(docker): add `--headless` in the Dockerfile - Add `--headless` option to the execution command of `kohya_gui.py` in the Dockerfile Signed-off-by: 陳鈞 <[email protected]> --------- Signed-off-by: 陳鈞 <[email protected]>
* Update IPEX to 2.1.20+xpu * Disable ipexrun by default
* Update zh-TW and fixed Lycoris Merge, additional_parameters priority modify. (#2182) * Update localizations for zh-TW * Fixed merge_lycoris_gui issue * Move additional_parameters to the very last of the run_cmd which can override the arguments. * Bump crate-ci/typos from 1.18.2 to 1.19.0 Bumps [crate-ci/typos](https://github.com/crate-ci/typos) from 1.18.2 to 1.19.0. - [Release notes](https://github.com/crate-ci/typos/releases) - [Changelog](https://github.com/crate-ci/typos/blob/master/CHANGELOG.md) - [Commits](crate-ci/typos@v1.18.2...v1.19.0) --- updated-dependencies: - dependency-name: crate-ci/typos dependency-type: direct:production update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] <[email protected]> --------- Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: bmaltais <[email protected]> Co-authored-by: Hina Chen <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
* Move accelerate launch parameters to own accordion
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Update sd-scripts to 0.8.7
The default value of
huber_schedule
in Scheduled Huber Loss is changed fromexponential
tosnr
, which is expected to give better results.Highlights
imagesize
is newly added, so if you cannot update the libraries immediately, please install withpip install imagesize==1.4.1
separately.bitsandbytes==0.43.0
,prodigyopt==1.0
,lion-pytorch==0.0.6
are included in the requirements.txt.bitsandbytes
no longer requires complex procedures as it now officially supports Windows..toml
). Thanks to bghira for raising the issue.--console_log_simple
option in the training script to disable rich logging.Training scripts
train_network.py
andsdxl_train_network.py
are modified to record some dataset settings in the metadata of the trained model (caption_prefix
,caption_suffix
,keep_tokens_separator
,secondary_separator
,enable_wildcard
).train_network.py
andsdxl_train_network.py
. The saving and loading of the state are faster, the file size is smaller, and the memory usage when loading is reduced.--noise_offset_random_strength
and--ip_noise_gamma_random_strength
are added to each training script. These options can be used to vary the noise offset and ip noise gamma in the range of 0 to the specified value. PR #1177 Thanks to KohakuBlueleaf!--save_state_on_train_end
are added to each training script. PR #1168 Thanks to gesen2egee!--sample_every_n_epochs
and--sample_every_n_steps
in each training script now display a warning and ignore them when a number less than or equal to0
is specified. Thanks to S-Del for raising the issue.Dataset settings
.toml
file for the dataset config is now read in UTF-8 encoding. PR #1167 Thanks to Horizon1704!secondary_separator
is added to specify the tag separator that is not the target of shuffling or dropping.secondary_separator=";;;"
. When you specifysecondary_separator
, the part is not shuffled or dropped.enable_wildcard
is added. When set totrue
, the wildcard notation{aaa|bbb|ccc}
can be used. The multi-line caption is also enabled.keep_tokens_separator
is updated to be used twice in the caption. When you specifykeep_tokens_separator="|||"
, the part divided by the second|||
is not shuffled or dropped and remains at the end.caption_prefix
andcaption_suffix
can be used together.caption_prefix
andcaption_suffix
are processed first, and thenenable_wildcard
,keep_tokens_separator
, shuffling and dropping, andsecondary_separator
are processed in order.Image tagging
tag_image_by_wd14_tagger.py
(--onnx
option only). PR #1192 Thanks to sdbds!pip install onnx==1.15.0 onnxruntime-gpu==1.17.1
etc. Please also check the comments inrequirements.txt
.--repo_id
intag_image_by_wd14_tagger.py
. This caches multiple repo_id models. Please delete unnecessary files under--model_dir
.tag_image_by_wd14_tagger.py
.--use_rating_tags
and--use_rating_tags_as_last_tag
--character_tags_first
--character_tag_expand
--always_first_tags
--tag_replacement
--beam_search
and a value of 2 or more for--num_beams
inmake_captions.py
.About Masked loss
The masked loss is supported in each training script. To enable the masked loss, specify the
--masked_loss
option.The feature is not fully tested, so there may be bugs. If you find any issues, please open an Issue.
ControlNet dataset is used to specify the mask. The mask images should be the RGB images. The pixel value 255 in R channel is treated as the mask (the loss is calculated only for the pixels with the mask), and 0 is treated as the non-mask. The pixel values 0-255 are converted to 0-1 (i.e., the pixel value 128 is treated as the half weight of the loss). See details for the dataset specification in the LLLite documentation.
About Scheduled Huber Loss
Scheduled Huber Loss has been introduced to each training scripts. This is a method to improve robustness against outliers or anomalies (data corruption) in the training data.
With the traditional MSE (L2) loss function, the impact of outliers could be significant, potentially leading to a degradation in the quality of generated images. On the other hand, while the Huber loss function can suppress the influence of outliers, it tends to compromise the reproduction of fine details in images.
To address this, the proposed method employs a clever application of the Huber loss function. By scheduling the use of Huber loss in the early stages of training (when noise is high) and MSE in the later stages, it strikes a balance between outlier robustness and fine detail reproduction.
Experimental results have confirmed that this method achieves higher accuracy on data containing outliers compared to pure Huber loss or MSE. The increase in computational cost is minimal.
The newly added arguments loss_type, huber_schedule, and huber_c allow for the selection of the loss function type (Huber, smooth L1, MSE), scheduling method (exponential, constant, SNR), and Huber's parameter. This enables optimization based on the characteristics of the dataset.
See PR #1228 for details.
loss_type
: Specify the loss function type. Choosehuber
for Huber loss,smooth_l1
for smooth L1 loss, andl2
for MSE loss. The default isl2
, which is the same as before.huber_schedule
: Specify the scheduling method. Chooseexponential
,constant
, orsnr
. The default issnr
.huber_c
: Specify the Huber's parameter. The default is0.1
.Please read Releases for recent updates.`
Added GUI support for the new parameters listed above.
Moved accelerate launch parameters to a new
Accelerate launch
accordion above theModel
accordion.Added support for
Debiased Estimation loss
to Dreambooth settings.Added support for "Dataset Preparation" defaults via the config.toml file.
Added a field to allow for the input of extra accelerate launch arguments.
Added new caption tool from https://github.com/kainatquaderee