-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the width of the model #8
Comments
Thank you for your question! Our aim is to maximize the parallelism of the network and reduce the number of layers as much as possible to achieve better speed on the GPU. The current number of channels represents a good trade-off. Reducing the channels may cause a significant decrease in accuracy with relatively little gain in speed, hence our choice of these specific channel numbers. Of course, you're welcome to try different channel numbers and train the network yourself to explore further. |
Thank you for your reply! as you say “maximize the parallelism of the network”, so if the optimal channels of layer in model is related to the hardware performance, whether the optimal number of channels is different on different hardware? Especially the edge end Gpus? |
Absolutely, the optimal number of channels is indeed influenced by the specific hardware in use. Particularly for edge-end GPUs, which may have less parallel processing capability compared to more robust systems, a smaller number of channels could be more suitable. It's always crucial to take into account the specific hardware characteristics when seeking to optimize for the best performance. |
I see. Thank you very much for your reply~ |
Your work is exciting! but I have a question about the width of the model,Have you tried to reduce the width of the model, why is the width of the smallest Vanillanet_5 need [1284, 2564, 5124, 10244]? thank you~
The text was updated successfully, but these errors were encountered: