-
-
Notifications
You must be signed in to change notification settings - Fork 223
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
is it possible to quantify the AnimeGANv3 model from FP32 to INT8 #58
Comments
Quantization is possible, but I haven't tried it yet. |
I just want to know the feasibility, and try to do that |
As shown above, I built a script to convert tensorflow's pb model to a quantized tflite model. The quantization formats include INT8 and float16. You can deploy them on mobile devices that support tflite, such as Android phones. The comparison results before and after quantification are as follows: It can be seen that after quantization, the output visual effect of the AnimeGANv3 model still maintains a high quality, and the model file is also reduced a lot. |
awesome, thanks a lot |
with porting tflite lib to rk3588, at least it can work by cpu inference |
is it possible to quantify the AnimeGANv3 model from FP32 to INT8?
The text was updated successfully, but these errors were encountered: