Model Rubik's Cube: Twisting Resolution, Depth and Width for TinyNets
–Neural Information Processing Systems
To obtain excellent deep neural architectures, a series of techniques are carefully designed in EfficientNets. The giant formula for simultaneously enlarging the resolution, depth and width provides us a Rubik's cube for neural networks. So that we can find networks with high efficiency and excellent performance by twisting the three dimensions. This paper aims to explore the twisting rules for obtaining deep neural networks with minimum model sizes and computational costs. Different from the network enlarging, we observe that resolution and depth are more important than width for tiny networks.
Neural Information Processing Systems
Oct-11-2024, 14:24:32 GMT