Commit Graph

2501 Commits

Author SHA1 Message Date
06eb9fb426 feat: add support for HunYuanDit ControlNet (#4245)
* add support for HunYuanDit ControlNet

* fix hunyuandit controlnet

* fix typo in hunyuandit controlnet

* fix typo in hunyuandit controlnet

* fix code format style

* add control_weight support for HunyuanDit Controlnet

* use control_weights in HunyuanDit Controlnet

* fix typo
2024-08-09 02:59:24 -04:00
413322645e Raw torch is faster than einops? 2024-08-08 22:09:29 -04:00
11200de970 Cleaner code. 2024-08-08 20:07:09 -04:00
037c38eb0f Try to improve inference speed on some machines. 2024-08-08 17:29:27 -04:00
1e11d2d1f5 Better prints. 2024-08-08 17:29:27 -04:00
65ea6be38f PullRequest CI Run: use pull_request_target to allow the CI Dashboard to work (#4277)
'_target' allows secrets to pass through, and we're just using the secret that allows uploading to the dashboard and are manually vetting PRs before running this workflow anyway
2024-08-08 17:20:48 -04:00
5df6f57b5d minor fix on copypasta action name (#4276)
my bad sorry
2024-08-08 16:30:59 -04:00
6588bfdef9 add GitHub workflow for CI tests of PRs (#4275)
When the 'Run-CI-Test' label is added to a PR, it will be tested by the CI, on a small matrix of stable versions.
2024-08-08 16:24:49 -04:00
50ed2879ef Add full CI test matrix GitHub Workflow (#4274)
automatically runs a matrix of full GPU-enabled tests on all new commits to the ComfyUI master branch
2024-08-08 15:40:07 -04:00
66d4233210 Fix. 2024-08-08 15:16:51 -04:00
591010b7ef Support diffusers text attention flux loras. 2024-08-08 14:45:52 -04:00
08f92d55e9 Partial model shift support. 2024-08-08 14:45:06 -04:00
8115d8cce9 Add Flux fp16 support hack. 2024-08-07 15:08:39 -04:00
6969fc9ba4 Make supported_dtypes a priority list. 2024-08-07 15:00:06 -04:00
cb7c4b4be3 Workaround for lora OOM on lowvram mode. 2024-08-07 14:30:54 -04:00
1208863eca Fix "Comfy" lora keys.
They are in this format now:
diffusion_model.full.model.key.name.lora_up.weight
2024-08-07 13:49:31 -04:00
e1c528196e Fix bundled embed. 2024-08-07 13:30:45 -04:00
17030fd4c0 Support for "Comfy" lora format.
The keys are just: model.full.model.key.name.lora_up.weight

It is supported by all comfyui supported models.

Now people can just convert loras to this format instead of having to ask
for me to implement them.
2024-08-07 13:18:32 -04:00
c19dcd362f Controlnet code refactor. 2024-08-07 12:59:28 -04:00
1c08bf35b4 Support format for embeddings bundled in loras. 2024-08-07 03:45:25 -04:00
2a02546e20 Add type hints to folder_paths.py (#4191)
* add type hints to folder_paths.py

* replace deprecated standard collections type hints

* fix type error when using Python 3.8
2024-08-06 21:59:34 -04:00
b334605a66 Fix OOMs happening in some cases.
A cloned model patcher sometimes reported a model was loaded on a device
when it wasn't.
v0.0.4
2024-08-06 13:36:04 -04:00
de17a9755e Unload all models if there's an OOM error. 2024-08-06 03:30:28 -04:00
c14ac98fed Unload models and load them back in lowvram mode no free vram. 2024-08-06 03:22:39 -04:00
2894511893 Clone taesd with depth of 1 to reduce download size. (#4232) 2024-08-06 01:46:09 -04:00
f3bc40223a Add format metadata to CLIP save to make compatible with diffusers safetensors loading (#4233) 2024-08-06 01:45:24 -04:00
841e74ac40 Change browser test CI python to 3.8 (#4234) 2024-08-06 01:27:28 -04:00
2d75df45e6 Flux tweak memory usage. 2024-08-05 21:58:28 -04:00
1abc9c8703 Stable release uses cached dependencies (#4231)
* Release stable based on existing tag.

* Update default cuda to 12.1.
2024-08-05 20:07:16 -04:00
8edbcf5209 Improve performance on some lowend GPUs. 2024-08-05 16:24:04 -04:00
e545a636ba This probably doesn't work anymore. 2024-08-05 12:31:42 -04:00
33e5203a2a Don't cache index.html (#4211) 2024-08-05 12:25:28 -04:00
a178e25912 Fix Flux FP64 math on XPU (#4210) 2024-08-05 01:26:20 -04:00
78e133d041 Support simple diffusers Flux loras. 2024-08-04 22:05:48 -04:00
7afa985fba Correct spelling 'token_weight_pars_t5' to 'token_weight_pairs_t5' (#4200) 2024-08-04 17:10:02 -04:00
ddb6a9f47c Set the step in EmptySD3LatentImage to 16.
These models work better when the res is a multiple of 16.
2024-08-04 15:59:02 -04:00
3b71f84b50 ONNX tracing fixes. 2024-08-04 15:45:43 -04:00
0a6b008117 Fix issue with some custom nodes. 2024-08-04 10:03:33 -04:00
56f3c660bf ModelSamplingFlux now takes a resolution and adjusts the shift with it.
If you want to sample Flux dev exactly how the reference code does use
the same resolution as your image in this node.
2024-08-04 04:06:00 -04:00
f7a5107784 Fix crash. 2024-08-03 16:55:38 -04:00
91be9c2867 Tweak lowvram memory formula. 2024-08-03 16:44:50 -04:00
03c5018c98 Lower lowvram memory to 1/3 of free memory. 2024-08-03 15:14:07 -04:00
2ba5cc8b86 Fix some issues. 2024-08-03 15:06:40 -04:00
1e68002b87 Cap lowvram to half of free memory. 2024-08-03 14:50:20 -04:00
ba9095e5bd Automatically use fp8 for diffusion model weights if:
Checkpoint contains weights in fp8.

There isn't enough memory to load the diffusion model in GPU vram.
2024-08-03 13:45:19 -04:00
f123328b82 Load T5 in fp8 if it's in fp8 in the Flux checkpoint. 2024-08-03 12:39:33 -04:00
63a7e8edba More aggressive batch splitting. 2024-08-03 11:53:30 -04:00
0eea47d580 Add ModelSamplingFlux to experiment with the shift value.
Default shift on Flux Schnell is 0.0
2024-08-03 03:54:38 -04:00
7cd0cdfce6 Add advanced model merge node for Flux model. 2024-08-02 23:20:53 -04:00
ea03c9dcd2 Better per model memory usage estimations. 2024-08-02 18:09:24 -04:00