-
-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
INT 8 quantization is not implemented when selected to export TensorFlow Lite #748
Comments
👋 Hello @franciscocostela, thank you for raising an issue about Ultralytics HUB 🚀! Please visit our HUB Docs to learn more:
If this is a 🐛 Bug Report, please provide screenshots and steps to reproduce your problem to help us get started working on a fix. If this is a ❓ Question, please provide as much information as possible, including dataset, model, environment details etc. so that we might provide the most helpful response. We try to respond to all issues as promptly as possible. Thank you for your patience! |
@franciscocostela Hello! |
Hi Sergiu, Yes - It is about 4x smaller. These are the sizes of the files: I am trying to run the TFLite file through a conversion pipeline to deploy it into a camera but it fails with an error message about the file not being quantized. When I inspect it with Netron, I see that the quantization bias is FLOAT32. INT8 is used in some of the convolution layers but not all of them (see screenshot). This seems to trigger the error message using the conversion pipeline. |
@franciscocostela Based on the file size, the quantization is applied. |
👋 Hello there! We wanted to give you a friendly reminder that this issue has not had any recent activity and may be closed soon, but don't worry - you can always reopen it if needed. If you still have any questions or concerns, please feel free to let us know how we can help. For additional resources and information, please see the links below:
Feel free to inform us of any other issues you discover or feature requests that come to mind in the future. Pull Requests (PRs) are also always welcomed! Thank you for your contributions to YOLO 🚀 and Vision AI ⭐ |
Search before asking
HUB Component
Export
Bug
I trained an Object Detection Model using both YoloV5n and YoloV8n. In the Deploy tab, I selected TensorFlow Lite - advanced and selected 'Int 8 Quantization'. Then I clicked Export and downloaded the model when the button Download became available.
![Screenshot 2024-06-06 at 9 25 05 AM](https://private-user-images.githubusercontent.com/68574696/344306326-9c8839e2-095f-4130-801f-9cc30ee78910.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk1ODExMTcsIm5iZiI6MTczOTU4MDgxNywicGF0aCI6Ii82ODU3NDY5Ni8zNDQzMDYzMjYtOWM4ODM5ZTItMDk1Zi00MTMwLTgwMWYtOWNjMzBlZTc4OTEwLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTUlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjE1VDAwNTMzN1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWIwYzY4NmNhMjFjZTRmMzZjZTdlYTMxOTUyY2M5YmE4M2JmYzAwNzJkODMyNDUyMGEyN2Y0ZTBiMjhhOGZmNTMmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.GtXaE76YCwunhfxGvHiwN2OrUDYrVMfHxKf4lE1ZdLE)
![image](https://private-user-images.githubusercontent.com/68574696/344307229-aa52f73a-698b-4445-b0c4-859bf7cd5c34.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk1ODExMTcsIm5iZiI6MTczOTU4MDgxNywicGF0aCI6Ii82ODU3NDY5Ni8zNDQzMDcyMjktYWE1MmY3M2EtNjk4Yi00NDQ1LWIwYzQtODU5YmY3Y2Q1YzM0LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTUlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjE1VDAwNTMzN1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWYzYjA2MjZjZWU4NDhkMjMxY2E2YjI4YmE4MjExMGU5NjM2MDJjNjM1ZWFjM2U5ZDUwZTBmOGEzODYxYzBlMDAmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.w5UIe9PRYXIkx57-ldfqab9CozPoi7mI1g9pu4n44Aw)
However, when I inspect the file, it looks like the quantization is never applied. This happens for both YoloV5n and YoloV8n
Is there anything that I am not doing correctly or is this really a bug?
Environment
No response
Minimal Reproducible Example
No response
Additional
No response
The text was updated successfully, but these errors were encountered: