Speakers
Description
Recognizing symmetries in data allows for significant boosts in neural network training. In many cases, however, the underlying symmetry is present only in an idealized dataset, and is broken in the training data, due to effects such as arbitrary and/or non-uniform detector bin edges. Standard approaches, such as data augmentation or equivariant networks fail to represent the nature of the full, broken symmetry. We introduce a novel data-augmentation scheme, which augments the training set with transformed pre-detector examples which respect the true underlying symmetry and avoid artifacts. In addition, we encourage the network to treat the augmented copies identically, allowing it to learn the broken symmetry. While the technique can be extended to other symmetries, we demonstrate its application on rotational symmetry in particle physics calorimeter images. We find that standard neural networks converge to a solution more quickly than networks trained without data augmentation, and that networks modified to encourage similar internal treatment of augmentations of the same input converge even faster.