40/2-A01 (CERN)



Show room on map

The effectiveness of convolutional neural networks stems in large part from their ability to exploit the translation invariance that is inherent in many learning prob- lems. Recently, it was shown that CNNs can exploit other sources of invariance, such as rotation invariance, by using group convolutions instead of planar con- volutions. However, for reasons of performance and ease of implementation, it has been necessary to limit the group convolution to transformations that can be applied to the filters without interpolation. Thus, for images with square pixels, only integer translations, rotations by multiples of 90 degrees, and reflections are admissible.

Whereas the square tiling provides a 4-fold rotational symmetry, a hexagonal tiling of the plane has a 6-fold rotational symmetry. In this paper we show how one can efficiently implement planar convolution and group convolution over hexagonal lattices, by re-using existing highly optimized convolution routines. We find that, due to the reduced anisotropy of hexagonal filters, planar HexaConv provides bet- ter accuracy than planar convolution with square filters, given a fixed parameter budget. Furthermore, we find that the increased degree of symmetry of the hexag- onal grid increases the effectiveness of group convolutions, by allowing for more parameter sharing. We show that our method significantly outperforms conven- tional CNNs on the AID aerial scene classification dataset, even outperforming ImageNet pretrained models. 

    • 11:15 AM 11:45 AM
      paper discussion 30m
      Speaker: Shah Rukh Qasim (SEECS - School of Electrical Engine ering and Computer Science )