November 29, 2021 to December 3, 2021
Virtual and IBS Science Culture Center, Daejeon, South Korea
Asia/Seoul timezone

A Comprehensive Comparison of GNN Architectures for Jet Tagging

contribution ID 774
Not scheduled
Windmill (Gather.Town)



Poster Track 2: Data Analysis - Algorithms and Tools Posters: Windmill


Daniel Thomas Murnane (Lawrence Berkeley National Lab. (US))


There has been significant interest and development in the use of graph neural networks (GNNs) for jet tagging applications. These generally provide better accuracy than CNN and energy flow algorithms by exploiting a range of GNN mechanisms, such as dynamic graph construction, equivariance, attention, and large parameterizations. In this work, we present the first apples-to-apples exploration of these mechanisms for jet tagging applications. In particular, we focus on concrete use-cases in a detector environment, and introduce metrics to quantify performance gain vs model size. With this full exploration of the hyperparameter space, we are then in a position to make recommendations for given hardware and latency constraints. Additionally, we investigate specific claims that purport certain GNN mechanisms to be responsible for better performance, and show that in fact state-of-the-art performance can be attained with a subset of these mechanisms and with significantly reduced architectural complexity.


This is the first full comparison and sensitivity study of GNN architectures and hyperparameters for jet tagging, including new innovations in Lorentz equivariance

Speaker time zone Compatible with America

Primary authors

Daniel Thomas Murnane (Lawrence Berkeley National Lab. (US)) Jason Wong (University of California, Berkeley) Savannah Jennifer Thais (Princeton University (US))

Presentation materials