-
Notifications
You must be signed in to change notification settings - Fork 681
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
In jetson nano, inference time of fp16 and fp_32 is almost same? #470
Comments
Hi xxy90, Thanks for reaching out! Do you mind sharing which model architecture you're referring to? The relative performance of FP32 vs. FP16 may depend on model architecture. I think the scaling also might not be linear with bit depth, because of various overhead when using reduced precision. Best, |
Hi John, |
The model architecture is new anchor-free object-detection Nanodet, whose backbone is shuffleNetv2 |
I also meet the same problem while using YOLOX-Nano(ref https://github.com/Megvii-BaseDetection/YOLOX) on jetson nano. |
No description provided.
The text was updated successfully, but these errors were encountered: