-
Notifications
You must be signed in to change notification settings - Fork 6.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Above 32 GB RAM usage, when loading Flux models in checkpoint version. #4239
Comments
Same happening here. The model I'm using is only 17.2 GB, but it tries to fill up all my RAM before it even tries to use the GPU. I'm so tired of requirements increasingly exponentially in AI. Feels like it's designed to be used online only so you're a slave to their GPU clusters. |
It's likely doing some kind of casting up to float32 or 16 and then back down to fp8, even if you're using an fp8 version of the model. It might not be the transformer though, maybe it's doing it for the t5 or something. I haven't actually checked to verify though. |
Here is a summary of my observations, in case it helps. When I use the fp16 models ( and t5 also in fp16 ):
With the Comfy-org Flux checkpoint:
|
Here's some observations from other users, with more RAM. |
Yeah I think I was on to something about it upcasting: ComfyUI/comfy/supported_models.py Line 631 in 1c08bf3
|
Even if fp8 is not possible, just supporting / upcasting to fp16 would be a good improvement. The fp16 model is 23.8 GB |
I got this problem to it blow my ram and swap even if I don't type python main.py --use-split-cross-attention its crash whole ubuntu os. If I run I Stuck at 32 gb ram load 4 gb frozen swap and stuck at .vae and can't gen. |
Expected Behavior
Keep RAM usage below the limit, to avoid wearing down my SSD.
Actual Behavior
I have 8GB VRAM and 32 GB RAM
I'm on Windows 10
With the full size fp16 models, my RAM usage goes above the limit, when the models need to be loaded.
It works, but the available SSD space goes down.
This is normal, I guess, considering the model sizes.
But it also happens with the ( fp8 ) Comfy-org checkpoint models ( 17.2 GB )
Steps to Reproduce
I used the default workflow.
Mode details in this discussion, with task manager images: #4226
The text was updated successfully, but these errors were encountered: