Dilated Convolution with Learnable Spacings: beyond bilinear interpolation
Khalfaoui-Hassani, Ismail, Pellegrini, Thomas, Masquelier, Timothée
–arXiv.org Artificial Intelligence
Dilated Convolution with Learnable Spacings (DCLS) is a recently proposed variation of the dilated convolution in which the spacings between the non-zero elements in the kernel, or equivalently their positions, are learnable. Non-integer positions are handled via interpolation. Thanks to this trick, positions have well-defined gradients. The original DCLS used bilinear interpolation, and thus only considered the four nearest pixels. Yet here we show that longer range interpolations, and in particular a Gaussian interpolation, allow improving performance on ImageNet1k classification on two state-of-the-art convolutional architectures (ConvNeXt and Conv\-Former), without increasing the number of parameters. The method code is based on PyTorch and is available at https://github.com/K-H-Ismail/Dilated-Convolution-with-Learnable-Spacings-PyTorch
arXiv.org Artificial Intelligence
Sep-22-2023
- Country:
- Europe > France
- Occitanie > Haute-Garonne > Toulouse (0.06)
- North America > United States
- Hawaii > Honolulu County > Honolulu (0.04)
- Europe > France
- Genre:
- Research Report (0.64)
- Technology: