-
Notifications
You must be signed in to change notification settings - Fork 890
Issue with running Flux.1.dev on iGPU #2926
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Instead of a screenshot could you show the used code as text, please? That would make reproduction much easier. Can you provide more details about your environment, OperatingSystem, version information? |
Thank you for your reply. The code is shown below.
command: python3 test.py "model directory" "cyberpunk cityscape like Tokyo and New York with tall buildings at dusk, golden hour, cinematic lighting" |
I just tried to run the Jupyter notebook under but it fails for me with
(being logged into HuggingFace, accepted license agreement, access-token provided) EDIT: According to this issue #2792 it was working in March this year. @eaidova was the model recently moved, renamed, removed, do you know? Or do my HuggingFace credentials (I'm located in Europe/Germany) not allow to access the model? |
@brmarkus dev model was never uploaed to huggingface hub under openvino account, it has license agreement limitations for that unfortunately, we have only schnell |
I was using |
Ok, thank you - now I initiated download, conversion and compression using the Jupyter-Notebook for the model "black-forest-labs/FLUX.1-schnell". Then I will try to reproduce inference using CPU and GPU from my MS-Win11-Pro, 64GB system memory, Intel Core Ultra 7 155H, using the query "cyberpunk cityscape like Tokyo and New York with tall buildings at dusk, golden hour, cinematic lighting". |
With the default parameters:
With the default checkbox "Use compressed models" activated. Using the CPU |
Using the GPU, same parameters, same checkboxes, using "black-forest-labs/FLUX.1-schnell": Progress-bar: Task-Manager showing GPU-utilization: @stsxxx your code uses "num_inference_steps = 50", while the Juypter-Notebook uses "Number of steps: 4" only. Is there a bigger difference between "FLUX.1-schnell" and "FLUX.1-dev"? |
I think the term "Non-Commercial Use Only" wouldn't allow me to use "FLUX.1-dev"... @stsxxx do you see similar values in your environment when using "FLUX.1-schnell" instead, to compare and check regarding your initial question "how to check its status and ensure it's being utilized properly in my setup"? |
@brmarkus Thank you for your reply. Also, if possible, could you try running Stable Diffusion XL in FP16 with image size 1024x1024? Since I was able to run it successfully, we could use it as a comparison to check whether my GPU is being utilized correctly. It’s late night on my end, so I’ll run FLUX.1-schnell tomorrow. Thank you again for your help! |
@brmarkus |
Would you have a chance to run the Juypiter notebook https://github.com/openvinotoolkit/openvino_notebooks/blob/e5a8aa127c9464a356a6767d2fb62b88ed21be3c/notebooks/flux.1-image-generation/flux.1-image-generation.ipynb On my CPU (and MS-Win11-Pro, 64GB RAM), I got "4/4 [01:44<00:00, 22.83s/it]" - less than 2 minutes for 4 iterations. |
I'll give it a try. Does this mean that other model formats aren't supported here? Even in the provided notebook, it mentions that you can use FLUX.1-dev by simply switching. I'm using the weights from your model hub, but it's not working—so I believe there may be another issue at play. |
OpenVINO supports different formats (like ONNX and others), but there is also an optimized format called IntermediateRepresentation "IR". |
Describe the bug
I'm trying to run the Flux.1.dev model using fp16 on the integrated GPU of an Intel® Core™ Ultra 7 165H. I attempted to generate an image with 50 inference steps, but one single step takes an extremely long time (forever) to complete.
Additionally, I ran Stable Diffusion XL in fp16 under the same setup (50 steps, 1024×1024), and it completed in about 3 minutes per image, which is significantly faster than Flux.1.dev but still slow.
Expected behavior
I want to verify whether my Intel integrated GPU (iGPU) is correctly activated and being used during inference. Could you guide me on how to check its status and ensure it's being utilized properly in my setup? Or is the performance I'm seeing simply expected for this hardware?
Screenshots

The code i am using:
**Installation instructions

I am not using the notebook.
Models have all been converted using
optimum-cli export openvino
GPU and NPU can be detected:
** Environment information **
openai==1.77.0
opencv-python== 4.11.0.86
openvino==2025.1.0
openvino-genai==2025.1.0.0
optimum==1.25.0.dev0
optimum-intel==1.23.0.dev0+590692f
The text was updated successfully, but these errors were encountered: