ce0aeb109e
Remove print.
2023-02-11 03:41:40 -05:00
5489d5af04
Add uni_pc sampler to KSampler* nodes.
2023-02-11 03:34:09 -05:00
1a4edd19cd
Fix overflow issue with inplace softmax.
2023-02-10 11:47:41 -05:00
509c7dfc6d
Use real softmax in split op to fix issue with some images.
2023-02-10 03:13:49 -05:00
7e1e193f39
Automatically enable lowvram mode if vram is less than 4GB.
...
Use: --normalvram to disable it.
2023-02-10 00:47:56 -05:00
e9d3ac2ba0
Typo.
2023-02-09 16:31:54 -05:00
9ae4e1b0ff
Add a list of features to the Readme.
2023-02-09 16:27:52 -05:00
324273fff2
Fix embedding not working when on new line.
2023-02-09 14:12:02 -05:00
1f6a467e92
Update ldm dir with latest upstream stable diffusion changes.
2023-02-09 13:47:36 -05:00
642516a3a6
create output dir if none is present
2023-02-09 12:49:31 -05:00
773cdabfce
Same thing but for the other places where it's used.
2023-02-09 12:43:29 -05:00
df40d4f3bf
torch.cuda.OutOfMemoryError is not present on older pytorch versions.
2023-02-09 12:33:27 -05:00
1d9ec62cfb
Use absolute output directory path.
2023-02-09 09:59:43 -05:00
05d571fe7f
Merge branch 'master' of https://github.com/bazettfraga/ComfyUI into merge_pr2
2023-02-09 00:44:38 -05:00
e8c499ddd4
Split optimization for VAE attention block.
2023-02-08 22:04:20 -05:00
5b4e312749
Use inplace operations for less OOM issues.
2023-02-08 22:04:13 -05:00
e58887dfa7
forgot windows does double backslashes for paths due to its use as escape char.
2023-02-09 01:30:06 +01:00
81082045c2
add recursive_search, swap relevant os.listdirs
2023-02-09 01:22:33 +01:00
3fd87cbd21
Slightly smarter batching behaviour.
...
Try to keep batch sizes more consistent which seems to improve things on
AMD GPUs.
2023-02-08 17:28:43 -05:00
bbdcf0b737
Use relative imports for k_diffusion.
2023-02-08 16:51:19 -05:00
3e22815a9a
Fix k_diffusion not getting imported from the folder.
2023-02-08 16:29:22 -05:00
708138c77d
Remove print.
2023-02-08 14:51:18 -05:00
047775615b
Lower the chances of an OOM.
2023-02-08 14:24:27 -05:00
853e96ada3
Increase it/s by batching together some stuff sent to unet.
2023-02-08 14:24:00 -05:00
c92633eaa2
Auto calculate amount of memory to use for --lowvram
2023-02-08 11:42:37 -05:00
534736b924
Add some low vram modes: --lowvram and --novram
2023-02-08 11:37:10 -05:00
a84cd0d1ad
Don't unload/reload model from CPU uselessly.
2023-02-08 03:40:43 -05:00
e3e65947f2
Add a --help to main.py
2023-02-07 22:13:42 -05:00
1f18221e17
Add --port to set custom port.
2023-02-07 21:57:17 -05:00
6e40393b6b
Fix delete sometimes not properly refreshing queue state.
2023-02-07 00:07:31 -05:00
d71d0c88e5
Add some simple queue management to the GUI.
2023-02-06 23:40:38 -05:00
b1a7c9ebf6
Embeddings/textual inversion support for SD2.x
2023-02-05 15:49:03 -05:00
1de5aa6a59
Add a CLIPLoader node to load standalone clip weights.
...
Put them in models/clip
2023-02-05 15:20:18 -05:00
56d802e1f3
Use transformers CLIP instead of open_clip for SD2.x
...
This should make things a bit cleaner.
2023-02-05 14:36:28 -05:00
bf9ccffb17
Small fix for SD2.x loras.
2023-02-05 11:38:25 -05:00
678105fade
SD2.x CLIP support for Loras.
2023-02-05 01:54:09 -05:00
3f3d77a324
Fix image node always executing instead of only when the image changed.
2023-02-04 16:08:29 -05:00
4225d1cb9f
Add a basic ImageScale node.
...
It's pretty much the same as the LatentUpscale node for now but for images
in pixel space.
2023-02-04 16:01:01 -05:00
bff0e11941
Add a LatentCrop node.
2023-02-04 15:21:46 -05:00
43c795f462
Add a --listen argument to listen on 0.0.0.0
2023-02-04 12:01:53 -05:00
41a7532c15
A bit bigger.
2023-02-03 13:56:00 -05:00
7bc3f91bd6
Add some instructions how to use the venv from another SD install.
2023-02-03 13:54:45 -05:00
149a4de3f2
Fix potential issue if exception happens when patching model.
2023-02-03 03:55:50 -05:00
ef90e9c376
Add a LoraLoader node to apply loras to models and clip.
...
The models are modified in place before being used and unpatched after.
I think this is better than monkeypatching since it might make it easier
to use faster non pytorch unet inference in the future.
2023-02-03 02:46:24 -05:00
96664f5d5e
Web interface bug fix for multiple inputs from the same node.
2023-02-03 00:39:28 -05:00
1d84a44b08
Fix some small annoyances with the UI.
2023-02-02 14:36:11 -05:00
e65a20e62a
Add a button to queue prompts to the front of the queue.
2023-02-01 22:34:59 -05:00
4b08314257
Add more features to the backend queue code.
...
The queue can now be queried, entries can be deleted and prompts easily
queued to the front of the queue.
Just need to expose it in the UI next.
2023-02-01 22:33:10 -05:00
9d611a90e8
Small web interface fixes.
2023-01-31 03:37:34 -05:00
fef41d0a72
Add LatentComposite node.
...
This can be used to "paste" one latent image on top of the other.
2023-01-31 03:35:03 -05:00