Commit Graph

85 Commits

Author SHA1 Message Date
AUTOMATIC
b6e5edd746 add built-in extension system
add support for adding upscalers in extensions
move LDSR, ScuNET and SwinIR to built-in extensions
2022-12-03 18:06:33 +03:00
AUTOMATIC
46b0d230e7 add comment for #4407 and remove seemingly unnecessary cudnn.enabled 2022-12-03 16:01:23 +03:00
AUTOMATIC
2651267e3a fix #4407 breaking UI entirely for card other than ones related to the PR 2022-12-03 15:57:52 +03:00
AUTOMATIC1111
681c0003df
Merge pull request #4407 from yoinked-h/patch-1
Fix issue with 16xx cards
2022-12-03 10:30:34 +03:00
brkirch
0fddb4a1c0 Rework MPS randn fix, add randn_like fix
torch.manual_seed() already sets a CPU generator, so there is no reason to create a CPU generator manually. torch.randn_like also needs a MPS fix for k-diffusion, but a torch hijack with randn_like already exists so it can also be used for that.
2022-11-30 10:33:42 -05:00
AUTOMATIC1111
cc90dcc933
Merge pull request #4918 from brkirch/pytorch-fixes
Fixes for PyTorch 1.12.1 when using MPS
2022-11-27 13:47:01 +03:00
AUTOMATIC
5b2c316890 eliminate duplicated code from #5095 2022-11-27 13:08:54 +03:00
Matthew McGoogan
c67c40f983 torch.cuda.empty_cache() defaults to cuda:0 device unless explicitly set otherwise first. Updating torch_gc() to use the device set by --device-id if specified to avoid OOM edge cases on multi-GPU systems. 2022-11-26 23:25:16 +00:00
brkirch
e247b7400a Add fixes for PyTorch 1.12.1
Fix typo "MasOS" -> "macOS"

If MPS is available and PyTorch is an earlier version than 1.13:
* Monkey patch torch.Tensor.to to ensure all tensors sent to MPS are contiguous
* Monkey patch torch.nn.functional.layer_norm to ensure input tensor is contiguous (required for this program to work with MPS on unmodified PyTorch 1.12.1)
2022-11-21 02:07:19 -05:00
brkirch
abfa22c16f Revert "MPS Upscalers Fix"
This reverts commit 768b95394a.
2022-11-17 00:08:21 -05:00
AUTOMATIC
0ab0a50f9a change formatting to match the main program in devices.py 2022-11-12 10:00:49 +03:00
源文雨
1130d5df66
Update devices.py 2022-11-12 11:09:28 +08:00
源文雨
76ab31e188 Fix wrong mps selection below MasOS 12.3 2022-11-12 11:02:40 +08:00
pepe10-gpu
62e9fec3df
actual better fix
thanks C43H66N12O12S2
2022-11-08 15:19:09 -08:00
pepe10-gpu
29eff4a194
terrible hack 2022-11-07 18:06:48 -08:00
pepe10-gpu
cd6c55c1ab
16xx card fix
cudnn
2022-11-06 17:05:51 -08:00
brkirch
faed465a0b MPS Upscalers Fix
Get ESRGAN, SCUNet, and SwinIR working correctly on MPS by ensuring memory is contiguous for tensor views before sending to MPS device.
2022-10-25 09:42:53 +03:00
brkirch
4c24347e45 Remove BSRGAN from --use-cpu, add SwinIR 2022-10-25 09:42:53 +03:00
AUTOMATIC
50b5504401 remove parsing command line from devices.py 2022-10-22 14:04:14 +03:00
Extraltodeus
57eb54b838
implement CUDA device selection by ID 2022-10-22 00:11:07 +02:00
brkirch
fdef8253a4 Add 'interrogate' and 'all' choices to --use-cpu
* Add 'interrogate' and 'all' choices to --use-cpu
* Change type for --use-cpu argument to str.lower, so that choices are case insensitive
2022-10-14 16:31:39 +03:00
AUTOMATIC
7349088d32 --no-half-vae 2022-10-10 16:16:29 +03:00
brkirch
e9e2a7ec9a
Merge branch 'master' into cpu-cmdline-opt 2022-10-04 07:42:53 -04:00
AUTOMATIC
6c6ae28bf5 send all three of GFPGAN's and codeformer's models to CPU memory instead of just one for #1283 2022-10-04 12:32:22 +03:00
brkirch
27ddc24fde Add BSRGAN to --add-cpu 2022-10-04 05:18:17 -04:00
brkirch
eeab7aedf5 Add --use-cpu command line option
Remove MPS detection to use CPU for GFPGAN / CodeFormer and add a --use-cpu command line option.
2022-10-04 04:24:35 -04:00
brkirch
b88e4ea7d6
Merge branch 'master' into master 2022-10-04 01:04:19 -04:00
AUTOMATIC
820f1dc96b initial support for training textual inversion 2022-10-02 15:03:39 +03:00
brkirch
bdaa36c844 When device is MPS, use CPU for GFPGAN instead
GFPGAN will not work if the device is MPS, so default to CPU instead.
2022-09-30 23:53:25 -04:00
AUTOMATIC
9d40212485 first attempt to produce crrect seeds in batch 2022-09-13 21:49:58 +03:00
AUTOMATIC
c7e0e28ccd changes for #294 2022-09-12 20:09:32 +03:00
AUTOMATIC
b70b51cc72 Allow TF32 in CUDA for increased performance #279 2022-09-12 16:34:13 +03:00
AUTOMATIC
8fb9c57ed6 add half() supporrt for CLIP interrogation 2022-09-11 23:24:24 +03:00
AUTOMATIC
f194457229 CLIP interrogator 2022-09-11 18:48:36 +03:00
Abdullah Barhoum
b5d1af11b7 Modular device management 2022-09-11 09:49:43 +03:00