Commit Graph

1825 Commits

Author SHA1 Message Date
AUTOMATIC1111
94450b8877
Merge pull request #5589 from MrCheeze/better-special-model-support
Better support for 2.0-inpainting and 2.0-depth special models
2022-12-24 09:53:44 +03:00
AUTOMATIC
9441c28c94 add an option for img2img background color 2022-12-24 09:46:35 +03:00
AUTOMATIC1111
b81fa1e7f1
Merge pull request #5644 from ThereforeGames/master
Improve img2img color correction by performing a luminosity blend
2022-12-24 09:17:40 +03:00
AUTOMATIC1111
684d7059bc
Merge pull request #5808 from stysmmaker/patch/fix-fnt-size
Prevent overlapping in X/Y plot by changing font size
2022-12-24 09:13:05 +03:00
AUTOMATIC1111
55f3ef876b
Merge pull request #5814 from timntorres/5802-save-upscaler-to-filename
Add option to save upscaler to filename suffix in extras tab.
2022-12-24 09:07:00 +03:00
AUTOMATIC1111
992a877a4a
Merge pull request #4684 from simcop2387/fix-extension-docker
Fix docker tmp/ and extensions/ handling for docker.
2022-12-24 09:06:10 +03:00
AUTOMATIC
399b229783 eliminate duplicated code
add an option to samplers for skipping next to last sigma
2022-12-24 09:03:45 +03:00
AUTOMATIC1111
5667ec4ca7
Merge pull request #5797 from mcmonkey4eva/dpm2-a-fix
Add a workaround patch for DPM2 a issue
2022-12-24 08:51:43 +03:00
AUTOMATIC1111
3bfc6c07ae
Merge pull request #5810 from brkirch/fix-training-mps
Training fixes for MPS
2022-12-24 08:34:46 +03:00
AUTOMATIC1111
f0dfed2a17
Merge pull request #5796 from brkirch/invoke-fix
Improve InvokeAI cross attention reliability and speed when using MPS for large images
2022-12-24 08:21:19 +03:00
AUTOMATIC
0c747d4013 add a comment for disable xformers hack 2022-12-24 07:57:56 +03:00
Akiba
13e0295ab6
fix: xformers use importlib 2022-12-24 11:17:21 +08:00
brkirch
35b1775b32 Use other MPS optimization for large q.shape[0] * q.shape[1]
Check if q.shape[0] * q.shape[1] is 2**18 or larger and use the lower memory usage MPS optimization if it is. This should prevent most crashes that were occurring at certain resolutions (e.g. 1024x1024, 2048x512, 512x2048).

Also included is a change to check slice_size and prevent it from being divisible by 4096 which also results in a crash. Otherwise a crash can occur at 1024x512 or 512x1024 resolution.
2022-12-20 21:30:00 -05:00
Philpax
22f1527fa7 feat(api): add override_settings_restore_afterwards 2022-12-20 20:36:49 +11:00
Alex "mcmonkey" Goodwin
7ba9bc2fdb fix dpm2 in img2img as well 2022-12-18 19:16:42 -08:00
Billy Cao
c02ef0f428 Fix PIL being imported before its installed (for new users only) 2022-12-18 20:51:59 +08:00
timntorres
6fd91c9179 Update OptionInfo to match preexisting option. 2022-12-17 08:59:02 -08:00
timntorres
a7a039d53a Add option to include upscaler name in filename. 2022-12-17 08:50:20 -08:00
timntorres
a26fe85056 Add upscaler name as a suffix. 2022-12-17 05:11:06 -08:00
brkirch
cca16373de Add attributes used by MPS 2022-12-17 04:23:08 -05:00
brkirch
16b4509fa6 Add numpy fix for MPS on PyTorch 1.12.1
When saving training results with torch.save(), an exception is thrown:
"RuntimeError: Can't call numpy() on Tensor that requires grad. Use tensor.detach().numpy() instead."

So for MPS, check if Tensor.requires_grad and detach() if necessary.
2022-12-17 04:22:58 -05:00
MMaker
b7c478c3eb
fix: Modify font size when unable to fit in plot
This prevents scenarios where text without line breaks will start overlapping with each other when generating X/Y plots. This is most evident when generating X/Y plots with checkpoints, as most don't contain spaces and sometimes include extra information such as the epoch, making it extra long.
2022-12-17 00:45:43 -05:00
Alex "mcmonkey" Goodwin
180fdf7809 apply to DPM2 (non-ancestral) as well 2022-12-16 08:42:00 -08:00
Alex "mcmonkey" Goodwin
8b0703b8fc Add a workaround patch for DPM2 a issue
DPM2 a and DPM2 a Karras samplers are both affected by an issue described by https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/3483 and can be resolved by a workaround suggested by the k-diffusion author at https://github.com/crowsonkb/k-diffusion/issues/43#issuecomment-1305195666
2022-12-16 08:18:29 -08:00
Akiba
35e1017e3e
fix: xformers 2022-12-16 20:43:09 +08:00
Jim Hays
c0355caefe Fix various typos 2022-12-14 21:01:32 -05:00
Yuval Aboulafia
957e15c464 Correct singleton comparisons 2022-12-14 20:59:33 +02:00
space-nuko
5f407ebd61 Fix comment 2022-12-13 14:32:26 -08:00
space-nuko
1fcb959514 Correctly restore default hypernetwork strength 2022-12-13 14:30:54 -08:00
space-nuko
9d5948e5f7 Correctly restore hypernetwork from hash 2022-12-13 14:25:16 -08:00
space-nuko
7077428209 Save hypernetwork hash in infotext 2022-12-13 13:05:40 -08:00
ThereforeGames
2e8b5418e3
Improve color correction with luminosity blend 2022-12-11 18:03:36 -05:00
Dean Hopkins
960293d6b2
API endpoint to refresh checkpoints
API endpoint to refresh checkpoints
2022-12-11 19:16:44 +00:00
MrCheeze
ec0a48826f unconditionally set use_ema=False if value not specified (True never worked, and all configs except v1-inpainting-inference.yaml already correctly set it to False) 2022-12-11 11:18:34 -05:00
Dean van Dugteren
59c6511494
fix: fallback model_checkpoint if it's empty
This fixes the following error when SD attempts to start with a deleted checkpoint:

```
Traceback (most recent call last):
  File "D:\Web\stable-diffusion-webui\launch.py", line 295, in <module>
    start()
  File "D:\Web\stable-diffusion-webui\launch.py", line 290, in start
    webui.webui()
  File "D:\Web\stable-diffusion-webui\webui.py", line 132, in webui
    initialize()
  File "D:\Web\stable-diffusion-webui\webui.py", line 62, in initialize
    modules.sd_models.load_model()
  File "D:\Web\stable-diffusion-webui\modules\sd_models.py", line 283, in load_model
    checkpoint_info = checkpoint_info or select_checkpoint()
  File "D:\Web\stable-diffusion-webui\modules\sd_models.py", line 117, in select_checkpoint
    checkpoint_info = checkpoints_list.get(model_checkpoint, None)
TypeError: unhashable type: 'list'
```
2022-12-11 17:08:51 +01:00
MrCheeze
bd81a09eac fix support for 2.0 inpainting model while maintaining support for 1.5 inpainting model 2022-12-10 11:29:26 -05:00
MrCheeze
a1c8ad8828 unload depth model if medvram/lowvram enabled 2022-12-10 11:02:47 -05:00
AUTOMATIC1111
94a35ca9d6
Merge pull request #5191 from aliencaocao/enable_checkpoint_switching_in_override_settings
Support changing checkpoint and vae through override_settings
2022-12-10 16:29:40 +03:00
AUTOMATIC
713c48ddd7 add an 'installed' tag to extensions 2022-12-10 15:05:22 +03:00
AUTOMATIC
991e2dcee9 remove NSFW filter and its dependency; if you still want it, find it in the extensions section 2022-12-10 14:54:16 +03:00
AUTOMATIC
d06592267c use less javascript for this non-js-only implementation of the clear prompt button. 2022-12-10 13:46:18 +03:00
AUTOMATIC1111
2028aa06c0
Merge pull request #3198 from papuSpartan/master
Add Clear Prompt button to roll_col
2022-12-10 13:35:51 +03:00
AUTOMATIC1111
89237852f4
Merge pull request #5119 from 0xb8/master
Atomically rename saved image to avoid race condition with other processes
2022-12-10 13:26:07 +03:00
AUTOMATIC1111
cce306cb67
Merge pull request #5441 from timntorres/add-5433-avoid-sending-size-option
Add option to avoid sending size between interfaces.
2022-12-10 11:07:16 +03:00
papuSpartan
6387043fd2
Merge branch 'AUTOMATIC1111:master' into master 2022-12-10 00:02:39 -08:00
AUTOMATIC1111
ec5e072124
Merge pull request #4841 from R-N/vae-fix-none
Fix None option of VAE selector
2022-12-10 09:58:20 +03:00
AUTOMATIC
bab91b1279 add Noise multiplier option to infotext 2022-12-10 09:51:26 +03:00
AUTOMATIC1111
8ee1acc1e4
Merge pull request #5373 from mezotaken/master
add noise strength parameter similar to NAI
2022-12-10 09:36:24 +03:00
AUTOMATIC1111
e5e557fa5d
Merge pull request #5404 from szhublox/merger-ram-usage
Merger ram usage
2022-12-10 09:33:39 +03:00
AUTOMATIC
505ec7e4d9 cleanup some unneeded imports for hijack files 2022-12-10 09:17:39 +03:00
AUTOMATIC
7dbfd8a7d8 do not replace entire unet for the resolution hack 2022-12-10 09:14:45 +03:00
AUTOMATIC1111
2641d1b83b
Merge pull request #4978 from aliencaocao/support_any_resolution
Patch UNet Forward to support resolutions that are not multiples of 64
2022-12-10 08:45:41 +03:00
AUTOMATIC1111
4d5fe3bfc0
Merge pull request #5555 from ywx9/master
Bug fix (a few lines in modules/api/api.py)
2022-12-10 08:27:44 +03:00
AUTOMATIC1111
a42a8e9112
Merge pull request #5547 from Ju1-js/master
Make "# settings changed" grammatically correct
2022-12-10 08:20:22 +03:00
ywx9
9539c2045a Bug fix 2022-12-09 23:03:06 +09:00
Ju1-js
ce04ba71b8 Make # settings changed message grammatically correct
Make the ": " in the settings changed message not show if 0 settings were changed.
"0 settings changed: ." -> "0 settings changed."
2022-12-08 22:47:45 -08:00
Jay Smith
1ed4f0e228 Depth2img model support 2022-12-08 20:50:08 -06:00
Andrew Ryan
358a8628f6 Add latent upscale option to img2img
Recently, the option to do latent upscale was added to txt2img highres
fix. This feature runs by scaling the latent sample of the image, and
then running a second pass of img2img.

But, in this edition of highres fix, the image and parameters cannot be
changed between the first pass and second pass. We might want to do a
fixup in img2img before doing the second pass, or might want to run the
second pass at a different resolution.

This change adds the option for img2img to perform its upscale in latent
space, rather than image space, giving very similar results to highres
fix with latent upscale.  The result is not exactly the same because
there is an additional latent -> decoder -> image -> encoder -> latent
that won't happen in highres fix, but this conversion has relatively
small losses
2022-12-08 07:09:09 +00:00
zhaohu xing
965fc5ac5a delete a file
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-12-06 16:15:15 +08:00
zhaohu xing
5dcc22606d add hash and fix undo hijack bug
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-12-06 16:04:50 +08:00
Zac Liu
a25dfebeed
Merge pull request #3 from 920232796/master
fix device support for mps
update the support for SD2.0
2022-12-06 09:17:57 +08:00
Zac Liu
3ebf977a6e
Merge branch 'AUTOMATIC1111:master' into master 2022-12-06 09:16:15 +08:00
zhaohu xing
4929503258 fix bugs
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-12-06 09:03:55 +08:00
timntorres
7057c72ae3 Add opt. to avoid sending size between interfaces. 2022-12-05 03:41:36 -08:00
Mackerel
681c450ecd extras.py: use as little RAM as possible, misc fixes
maximum of 2 models loaded at once. delete unneeded model before next
step. fix 'teritary' -> 'tertiary'. gracefully fail when "add
difference" is selected without a tertiary model
2022-12-04 10:31:06 -05:00
AUTOMATIC
44c46f0ed3 make it possible to merge inpainting model with non-inpainting one 2022-12-04 12:30:44 +03:00
AUTOMATIC
8504db5170 fix #4459 breaking inpainting when the option is not specified. 2022-12-04 01:04:24 +03:00
AUTOMATIC
60bd4d52a6 fix incorrect file extension filter for deepdanbooru models 2022-12-03 18:46:09 +03:00
AUTOMATIC
4b0dc206ed use modelloader for #4956 2022-12-03 18:45:51 +03:00
AUTOMATIC1111
2a649154ec
Merge pull request #4956 from TiagoSantos81/offline_BLIP
[CLIP interrogator] use local file, if available
2022-12-03 18:17:56 +03:00
AUTOMATIC
0d21624cee move #5216 to the extension 2022-12-03 18:16:19 +03:00
AUTOMATIC
89e1df013b Merge remote-tracking branch 'wywywywy/autoencoder-hijack' 2022-12-03 18:08:10 +03:00
AUTOMATIC
b6e5edd746 add built-in extension system
add support for adding upscalers in extensions
move LDSR, ScuNET and SwinIR to built-in extensions
2022-12-03 18:06:33 +03:00
Vladimir Repin
cf3e844d1d add noise strength parameter similar to NAI 2022-12-03 18:05:47 +03:00
AUTOMATIC
46b0d230e7 add comment for #4407 and remove seemingly unnecessary cudnn.enabled 2022-12-03 16:01:23 +03:00
AUTOMATIC
2651267e3a fix #4407 breaking UI entirely for card other than ones related to the PR 2022-12-03 15:57:52 +03:00
AUTOMATIC1111
681c0003df
Merge pull request #4407 from yoinked-h/patch-1
Fix issue with 16xx cards
2022-12-03 10:30:34 +03:00
AUTOMATIC1111
d2e5b4edfa
Merge pull request #5251 from adieyal/bug/negative-prompt-infotext
Fixed incorrect negative prompt text in infotext
2022-12-03 10:21:43 +03:00
AUTOMATIC1111
c9a2cfdf2a
Merge branch 'master' into racecond_fix 2022-12-03 10:19:51 +03:00
AUTOMATIC1111
5cd5a672f7
Merge pull request #4459 from kavorite/color-sketch-inpainting
add `--gradio-inpaint-tool` and option to specify `color-sketch`
2022-12-03 10:06:27 +03:00
AUTOMATIC1111
a2feaa95fc
Merge pull request #5194 from brkirch/autocast-and-mps-randn-fixes
Use devices.autocast() and fix MPS randn issues
2022-12-03 09:58:08 +03:00
AUTOMATIC
c7af672186 more simple config option name plus mouseover hint for clip skip 2022-12-03 09:41:39 +03:00
AUTOMATIC1111
c67d8bca4f
Merge pull request #5304 from space-nuko/fix/clip-skip-application
Fix clip skip of 1 not being restored from prompts
2022-12-03 09:37:10 +03:00
AUTOMATIC1111
28c79b8f05
Merge pull request #5328 from jcowens/fix-typo
fix typo
2022-12-03 09:20:39 +03:00
AUTOMATIC1111
eb0b8f92bc
Merge pull request #5331 from smirkingface/openaimodel_fix
Fixed AttributeError where openaimodel is not found
2022-12-03 09:18:36 +03:00
AUTOMATIC1111
bab6ea6b22
Merge pull request #5340 from PhytoEpidemic/master
Fix divide by 0 error
2022-12-03 09:17:54 +03:00
AUTOMATIC
b2f17dd367 prevent include_init_images from being passed to StableDiffusionProcessingImg2Img in API #4989 2022-12-03 09:15:24 +03:00
AUTOMATIC1111
ae81b377d4
Merge pull request #5165 from klimaleksus/fix-sequential-vae
Make VAE step sequential to prevent VRAM spikes, will fix #3059, #2082, #2561, #3462
2022-12-03 08:29:56 +03:00
PhytoEpidemic
119a945ef7
Fix divide by 0 error
Fix of the edge case 0 weight that occasionally will pop up in some specific situations. This was crashing the script.
2022-12-02 12:16:29 -06:00
SmirkingFace
da698ca92e Fixed AttributeError where openaimodel is not found 2022-12-02 13:47:02 +01:00
jcowens
99b19b1a8f fix typo 2022-12-02 02:53:26 -08:00
SmirkingFace
e461477869 Fixed safe.py for pytorch 1.13 ckpt files 2022-12-02 11:12:13 +01:00
zhaohu xing
9c86fb8cac fix bug
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-12-02 16:08:46 +08:00
space-nuko
be2e6de94a Fix clip skip of 1 not being restored from prompts 2022-12-01 11:34:16 -08:00
brkirch
0fddb4a1c0 Rework MPS randn fix, add randn_like fix
torch.manual_seed() already sets a CPU generator, so there is no reason to create a CPU generator manually. torch.randn_like also needs a MPS fix for k-diffusion, but a torch hijack with randn_like already exists so it can also be used for that.
2022-11-30 10:33:42 -05:00
brkirch
4d5f1691dd Use devices.autocast instead of torch.autocast 2022-11-30 10:33:42 -05:00
brkirch
21effd629d Add workaround for using MPS with torchsde 2022-11-30 10:33:39 -05:00
Adi Eyal
a44994e2c9 Fixed incorrect negative prompt text in infotext
Previously only the first negative prompt in all_negative_prompts was
being used for infotext. This fixes that by selecting the index-th
negative prompt
2022-11-30 15:23:53 +02:00
Billy Cao
3a724e91a2 Change to steps of 8 2022-11-30 20:52:32 +08:00
zhaohu xing
52cc83d36b fix bugs
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-11-30 14:56:12 +08:00