-
Notifications
You must be signed in to change notification settings - Fork 2.6k
Update supported cuda slot input. #8107
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update supported cuda slot input. #8107
Conversation
Great thanks. And just to confirm, does it work and let invoke run on the GPU you wanted it to? |
Sadly no, it's running on slot 1, I'm having a hard time tracing where the call to allocate is... |
I was wrong (to many generations running at the same time from my llm and comfyui confused me so I killed them all to confirm) with
|
That looks good but I'm sus on the VRAM usage in slot 0. Maybe there's something in Invoke that is not respecting the GPU selection. Could you please restart the system and try a couple different models - maybe add some controlnets or ip adapters in - and confirm we only see VRAM on slot 3? Edit: If it's not working, it'll be something we need to address separately from this PR - we can get this PR merged now. |
Gallery with a SDXL + lora + refiner + vae: Saw no usage outside of cuda:3 Canvas with a soft edge control + SDXL generation: saw no usage outside of cuda:3 |
d5f4722
to
37b98ee
Compare
Summary
Bring in support for defining a cuda gpu slot higher then 1.
Related Issues / Discussions
Closes #8102
QA Instructions
Ran the device pytest.
Merge Plan
Checklist
What's New
copy (if doing a release after this PR)