-
Notifications
You must be signed in to change notification settings - Fork 186
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to export model by using --device npu --provider QNNExecutionProvider #1595
Comments
6 tasks
Thanks for reporting the bug! I created #1600 to fix this. Please note that the |
jambayk
added a commit
that referenced
this issue
Feb 10, 2025
…1600) ## Describe your changes - `onnxruntime.tools.onnx_model_utils.fix_output_shapes` cannot handle large models (#1595), so we use the ort shape infer helper and handle the logic ourselves. This also means it can now handle models with contrib operators too. - allow passing 0 as dim_value. this case is possible when creating a prompt processing model from a dynamic shaped llm where we want to make the past kv cache empty. ## Checklist before requesting a review - [ ] Add unit tests for this change. - [ ] Make sure all tests can pass. - [ ] Update documents if necessary. - [ ] Lint and apply fixes to your code by running `lintrunner -a` - [ ] Is this a user-facing change? If yes, give a description of this change to be included in the release notes. - [ ] Is this PR including examples changes? If yes, please remember to update [example documentation](https://github.com/microsoft/Olive/blob/main/docs/source/examples.md) in a follow-up PR. ## (Optional) Issue link
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Describe the bug
Hi
I am trying to quantize and export a fine-tuned microsoft/phi-3.5-mini-instruct model.
I was able to export it by using
--device cpu --provider CPUExecutionProvider
.However when I try to export it by using
--device npu --provider QNNExecutionProvider
I got the following error.To Reproduce
Steps to reproduce the behavior.
Expected behavior
Be able to export onnx model by using
--device npu --provider QNNExecutionProvider
Olive config
Olive logs
Other information
Additional context
I found a similar issue here, which is also an ONNX conversion issue when the file size exceeds 2GB.
#1165
The text was updated successfully, but these errors were encountered: