-
Notifications
You must be signed in to change notification settings - Fork 890
Issue WIth Running Phi3 vision on NPU #2858
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
EDIT: OK, I see, looks like you are using "HWINfo64". Do you use this notebook https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/phi-3-vision/phi-3-vision.ipynb ? Which of the two models have you selected in the drop-down menu? Have you enabled the checkbox "Compress model" to apply model compression? Which version of the NPU driver do you have installed? Have you rebooted your system afterwards? I just cloned the current version of this repo, created a new virtual environment and started the Jupyter notebook - downloading and compression will take while; using the first model from dropdown "microsoft/Phi-3.5-vision-instruct" (default value) and kept the checkbox for model-compression (default value). Keep monitoring the system memory consumption while compression and conversion. Are you sure it finished successfully? |
Hi @brmarkus
|
Ok, thank you.
Let me now try "microsoft/Phi-3-vision-128k-instruct" as well - will take a while. However, when using NPU, in the step "pipe = ov_genai.VLMPipeline(model_dir, device.value)" I now get an exception:
This is also different than your shown exception. My environment:
Someone from OpenVINO-Notebook team need to have a closer look. UPDATE:
|
Uh oh!
There was an error while loading. Please reload this page.
I installed the NPU drivers as per the instructions, created the OpenVINO version of Phi 3 Vision model, set the device to NPU, and attempted to run the model. Initially, in Task Manager, I observed the model utilizing the NPU, reaching 100% usage. However, it immediately threw the following error:
RuntimeError:
Check 'prompt_ids.get_size() >= tokenized_history.size()' failed at C:\Jenkins\workspace\private-ci\ie\build-windows-vs2022\b\repos\openvino.genai\src\cpp\src\visual_language\pipeline.cpp:201:
Prompt IDs size is less than tokenized history size.
Expected behavior
Run Phi3 vision on NPU
Laptop Specs
Processor :Intel(R) Core(TM) Ultra 7 165U 2.10 GHz
RAM : 16GB
NPU : Intel(R)AI Boost
Screenshots
The text was updated successfully, but these errors were encountered: