-
Notifications
You must be signed in to change notification settings - Fork 57
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
why the model size all so big? #22
Comments
Hi, FasterNet focuses on improving the accuracy-latency tradeoff but not on reducing the model size (slightly bigger than ShuffleNetV2 and MobileNetV2), as shown in the following table:
|
@JierunChen hello, may I ask which framework used to get these values? |
We use FVcore for #params and FLOPs calculation. Code for measuring throughput/latency is provided in utils.py. Note that when comparing the model size by storage, ensure all models compared have the same precision, e.g., 32 bits, 16 bits, or 8 bits. |
@JierunChen thanks, seems mostly measuring on torch, have u tried export to onnx and measuring on onnxruntime? Since the speed actually highly related how optimized in forward frameworks. |
@jinfagang No deployment engines are involved when measuring the latency/throughput. We haven't tried exporting to onnx. You are welcome to do so and try various deployment engines, e.g. OnnxRuntime, TensorRT, NCNN, etc. |
compared with mbv2 and shufflenetv2 fasternet has a bigger size, this caused limitations deploy on edge devices.
The text was updated successfully, but these errors were encountered: