-
Notifications
You must be signed in to change notification settings - Fork 74.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multiplication following a int8 transposed convolutions isn't constant folded #57680
Comments
Hi @lgeiger ! |
@mohantym Thanks for the fast reply. However, I am able to reproduce this issue in 2.10, in fact this is the version which I am also using locally. Checkout this notebook. With per-tensor fake quantization I am getting the following TFLite graph: I get the same result also when re-running your linked notebook. |
I double checked by re-running your colab notebook and the issue persists even on nightly |
Issue exists in TF 2.11 and TF Nightly 2.12.0-dev20230111 as well. Please find the gist here. Thank you. |
Hi, Thank you for opening this issue. Since this issue has been open for a long time, the code/debug information for this issue may not be relevant with the current state of the code base. The TFLite team is constantly improving the framework by fixing bugs and adding new features. We suggest you try the latest TensorFlow version with the latest compatible hardware configuration which could potentially resolve the issue. If you are still facing the issue, please create a new GitHub issue with your latest findings, with all the debugging information which could help us investigate. Please follow the release notes to stay up to date with the latest developments which are happening in the TFLite space. |
@pkgoogle Thanks for the response. I double checked and the issue still persists in the latest TF nightly. See this notebook. |
Hi @lgeiger, Thanks for raising this issue. Are you aware of AI-Edge-Torch? As we believe this issue is better supported by and more relevant to AI-Edge-Torch we are moving your issue there. Please follow progress here. Let us know if you have any questions. Thanks. |
When converting a model that used int8 quantization aware training, conversion of transposed convolutions followed by a scalar multiplication fails.
The converter isn't able to correctly constant fold the per-tensor fake quantized weights and the scalar multiplication which is a common patter when using transposed convolutions followed by batch normalisation layers. This is a follow-up issue to #53766
1. System information
2. Code
A minimal reproduction of the issue is available in this notebook. Re-run the notebook to show netron visualisations showing the conversion problem.
The text was updated successfully, but these errors were encountered: