-
Notifications
You must be signed in to change notification settings - Fork 308
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to run phi3 on NPU via OnnxRuntime+DirectML #679
Comments
Hi @Gusha-nye, To run Phi3 on an NPU via OnnxRuntime and DirectML, follow these steps:
Regarding your second question, the hardware configuration you provided should be sufficient. Just ensure that your NPU is properly configured. |
Thank you for your answer |
Hi, bro @ashumish-QCOM |
Hi @Gusha-nye, To load the model onto the NPU using the C# API, you can refer to the DirectMLNpuInference sample on Microsoft Learn. This sample demonstrates how to perform ONNX Runtime inference via DirectML on an NPU, including selecting an NPU device, creating an ONNX Runtime session, executing the model on the NPU, and retrieving the results. Additionally, the ONNX Runtime C# API documentation provides detailed guidance on using the C# API. While it might not explicitly mention setting up the NPU, the DirectML sample should help you understand the process. |
Okay. Thank you very much. |
1.Now I successfully run Phi3 on GPU via ORT+DML, but I want it to run on NPU, how should I do it and what are the steps required?
2. Does he have hardware configuration requirements for the machine?
3. The current computer configuration is:
Processor: Intel(R) Core(TM) Ultra 7 165H 1.40GHz
GPU: Intel(R) Arc(TM) Graphics
NPU: Intel(R) AI Boost
Installed RAM: 32.0GB
System Type:64-bit operating system, x64-based processor
Windows Edition: Windows 11 pro
version: 24H2
But I also have a Copilot+ PC with an Arm-based Snapdragon Elite X chip built by Qualcomm.
The text was updated successfully, but these errors were encountered: