Skip to content

Unable to infer ltx-video on iGPU #2830

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
rainhenry opened this issue Mar 19, 2025 · 4 comments
Open

Unable to infer ltx-video on iGPU #2830

rainhenry opened this issue Mar 19, 2025 · 4 comments
Assignees
Labels
bug Something isn't working category: GPU OpenVINO GPU plugin PSE Escalate to PSE for further investigate

Comments

@rainhenry
Copy link

The path of the file in question:
notebooks/ltx-video/ltx-video.ipynb

I modified some of the code as follows:
##compiled_transformer = core.compile_model(TRANSFORMER_OV_PATH, device.value) ## Original code
compiled_transformer = core.compile_model(TRANSFORMER_OV_PATH, "GPU") ## I modified
compiled_vae = core.compile_model(VAE_DECODER_PATH, device.value)
compiled_text_encoder = core.compile_model(TEXT_ENCODER_PATH, device.value)

My idea is to let the transformer process perform inference on the iGPU, because this part is the most time-consuming part of the entire inference process. The value of device.value in the original code is "AUTO". By observing the load of CPU and iGPU, it is not difficult to find that when it is set to "AUTO", it works on the CPU instead of iGPU. So I manually changed this value to "GPU". After I modified it, the whole inference process seemed to be smooth, and no errors occurred. However, the final generated video file was a pair of chaotic and blurred noise, not a normal video. If I change this value back to CPU, then the generated video file is correct.

So I don't know what the problem is? Is it that I modified it in the wrong way? Or is there some bug in the code?

According to my previous experience of deploying some large models for generating videos to iGPU, when exporting from the original model to IR format, the exported IR model is strongly related to the resolution and frame rate information that needs to be output. I didn't see it in this demo, and I'm not sure if it is related to this reason.

My hardware device is Core Ultra7 155H

@YuChern-Intel
Copy link

YuChern-Intel commented Mar 24, 2025

I noticed that the tutorial only shows "CPU" and "AUTO" as the available devices. In this case, selecting "AUTO" will be the same as selecting "CPU".

There is an added quantization steps in the latest branch compared to 2025.0 which helps in making model inference faster.

@rainhenry
Copy link
Author

I noticed that the tutorial only shows "CPU" and "AUTO" as the available devices. In this case, selecting "AUTO" will be the same as selecting "CPU".

There is an added quantization steps in the latest branch compared to 2025.0 which helps in making model inference faster.

You are right. The results of selecting AUTO and CPU are the same. But I want to run the inference process on iGPU. How should I modify it? Because the inference speed of CPU is too slow.

@YuChern-Intel YuChern-Intel added the PSE Escalate to PSE for further investigate label Mar 26, 2025
@avitial avitial self-assigned this Mar 31, 2025
@openvino-dev-samples
Copy link
Collaborator

hi @aleksandr-mokrov, as the author of this notebook, do you know this issue on GPU ? thanks

@aleksandr-mokrov
Copy link
Contributor

I've reproduced the problem. Looks like we have an issue on GPU plugin side. I will let you know when the problem is fixed.

@avitial avitial added bug Something isn't working category: GPU OpenVINO GPU plugin labels May 7, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working category: GPU OpenVINO GPU plugin PSE Escalate to PSE for further investigate
Projects
None yet
Development

No branches or pull requests

6 participants