PHYSTAT - Statistics meets ML

Europe/London
Lecture Theatre 2, Blackett Laboratory (Imperial College London)

Lecture Theatre 2, Blackett Laboratory

Imperial College London

Louis Lyons (Imperial College (GB)), Olaf Behnke (Deutsches Elektronen-Synchrotron (DE)), Nicholas Wardle (Imperial College (GB)), Lydia Brenner (Nikhef National institute for subatomic physics (NL)), Sara Algeri (University of Minnesota)
Description

PHYSTAT Workshop on "Statistics meets ML" in Particle Physics & Astrophysics

In recent years, ML has become more and more integrated into many stages of our analyses.  In Particle Physics, it includes data collection and processing (triggering, tracking, etc), classification of different particle types, unfolding, parameter determination, anomaly detection, and even end-to-end processing. In astronomy its use is increasingly widespread in areas such as classification of objects, distance determination, and regression problems. It is used theoretically to enhance simulations, and in emulation of theoretical predictions that are expensive to compute. It is also increasingly used in simulation-based inference, both in finding informative summary statistics, and in variational methods of inference.
 
PHYSTAT’s “Statistics meets Machine Learning” aims to address some of the statistical issues that arise in these applications in Particle Physics and in Astronomy, with participation also of Statistics and Machine Learning experts. These issues are particularly important as the ML approaches tend to outperform traditional ones in terms of precision; the question is whether they are also more accurate. A problem is that in general it is hard to understand the procedure the ML method is adopting to achieve its result.

Some of the topics to be discussed:

  • Uncertainty quantification
  • Interpretable AI
  • Training samples & mismodelling
  • Generative methods
  • Statistical checks

There will be plenty of time for discussion.

The meeting will be hybrid i.e. particpitation either in person or remotely. In both cases, registration is necessary as the zoom link to the meeting will be sent a few days before the meeting only to people who have registered.
 
In-person registration closed on 9th August.  There was a limit of 110 for the number of in-person participants. Registration for remote participation remains open. Submisson of Abstracts for potential contributed talks/posters is also now closed.

For general enquiries please contact: fundamental-physics-admin@imperial.ac.uk. For questions regarding the content of the workshop, please contact Louis Lyons via: louis.lyons@physics.ox.ac.uk.

Local Organising Committee:
Matteo Agostini
Sunita Aubeeluck
Robert Bainbridge (chair)
Jonathon Langford
Louis Lyons (co-chair)
Benedikt Maier
Lucas Makinen
Marvin Pfaff
Nicholas Wardle
Daniel Winterbottom

Scientific Committee:
Pierre Baldi (UC Irvine, ML)
Olaf Behnke (DESY, PP)
Lydia Brenner (NIKHEF, PP)
Patrick Dunne (Imperial, PP)
Alan Heavens (Imperial, AC)
Lukas Heinrich (TUM, PP)
Mikael Kuusela (CMU, Statistics)
Ann Lee (CMU, Statistics)
Louis Lyons (Imperial, PP)
Ben Nachman (LBNL, PP)
Anna Sciafe (Jodrell Bank, AC)
Jesse Thaler (MIT, PP)
Roberto Trotta (Trieste, AC)
Nicholas Wardle (Imperial, PP)

AC: Astro/Cosmo
PP: Particle Physics

Poster PDF


PHYSTAT

The PHYSTAT series of Workshops started in 2000. They were the first meetings devoted solely to the statistical issues that occur in analyses in Particle Physics and neighbouring fields.The homepage of PHYSTAT with a list of all Workshops, Seminars and Informal Reviews is at https://phystat.github.io/Website/ .

Participants
  • Abanti Ranadhir Sahasransu
  • Abhijit Garai
  • Aditya Tandon
  • Agnieszka Oblakowska-Mucha
  • Ahmed Abdelmotteleb
  • Aishik Ghosh
  • Akram Khan
  • Alan Heavens
  • Alessio Spurio Mancini
  • Alex Hergenhan
  • Alexander Held
  • Ali Asghar
  • Alicja Polanska
  • Aliya Nigamova
  • Alvaro Tolosa-Delgado
  • Aman Desai
  • Amit Kumar
  • An Son
  • Andre Sznajder
  • Anibal Varela
  • Arda Batın Tank
  • Arda Eriş
  • Arsenii Gavrikov
  • Artur, Artemij Monsch
  • Arul Prakash
  • Arya Thampi
  • Ata Sattari
  • Atousa Kalantari
  • Aurora Singstad Grefsrud
  • Aurélien Dersy
  • Austin Murray
  • Badamsambuu Jigmeddorj
  • Baha Konukoğulları
  • Ben Nachman
  • Benjamin Boyd
  • Benjamin Quilain
  • Benjamin Wandelt
  • Biao Wang
  • Biswaranjan Behera
  • Borja Rodriguez Mateos
  • Brendan D'Aquino
  • Brianna Isola
  • Bruno Moraes
  • Burkan Bereketoğlu
  • Cal Hewitt
  • Carlos Javier Solano Salinas
  • Carlos Romero Mejia
  • Carolina Cuesta Lazaro
  • Caroline Samantha Heneka
  • Carsten Burgard
  • Carys Garvey
  • Caue Evangelista De Sousa
  • Cesare Tiziano Cazzaniga
  • Chandramauli Agrawal
  • Chayan Majumdar
  • chen lei
  • Cheng Fei Phung
  • Christos Leonidopoulos
  • Corentin Santos
  • Corey Lehmann
  • Cristiano Alpigiani
  • Daanish Sonawala
  • Dakshesh Kololgi
  • Dalmo Da Silva Dalto
  • Daniel Eduardo Conde Villatoro
  • Daniel Mortlock
  • David Colling
  • David van Dyk
  • David Yallup
  • Davide Valsecchi
  • Deaglan Bartlett
  • Dean Hansen
  • Denis Carabadjac
  • Dennis Noll
  • Dexing Miao
  • Diana Patricia Mendez Mendez
  • Dinupa Nawarathne
  • Dominik Fuchs
  • Duc Truyen LE
  • Eda Erdogan
  • Edgar Chucuan
  • Edgar Fernando Carrera Jarrin
  • Eduardo Ploerer
  • Efstathios Logothetis
  • Ekin Akman
  • Ekrem Oğuzhan Angüner
  • Eleni Skorda
  • Eleni Tsaprazi
  • Eleni Tsaprazi
  • Elliot Leader
  • Elrick Eduardo Haces Gil Sotelo
  • Elvira Rossi
  • Elías Abisaí Avendaño Lara
  • Emanuel Pfeffer
  • Emanuele Villa
  • Enrico Lupi
  • Enzo Canonero
  • Fatma Boran
  • Forhad Hossain
  • Francisco Matorras
  • Franz Glessgen
  • Gaia Grosso
  • Gavin Davies
  • Gehad Eldibany
  • Geting Qin
  • Gilberto Rodriguez Prado
  • Gilles Louppe
  • Giuseppe costantino
  • Giuseppe Costantino
  • Glen Cowan
  • Gregor Kasieczka
  • Haoyu Sun
  • Harisree Krishnamoorthy
  • Harry Desmond
  • HARVEY NEWMAN
  • Hassan Radwan
  • Havana Ide
  • Haydee Hernandez
  • Heather Battey
  • Henry Thomas Wallace
  • Heshy Roskes
  • Hirak Kumar Koley
  • Hiranya Peiris
  • Holden Truman
  • Ievgen Lavrukhin
  • Igor Volobouev
  • Ilias Tsaklidis
  • Irene Andreou
  • Ishwar Singh
  • Ivana Hristova
  • Jack Harrison
  • Jack Y. Araz
  • Jacob Smith
  • Jaime Leon Holgado
  • James Frost
  • James Pearson
  • Jan Conrad
  • Javier Jimenez Pena
  • Jawaher Altork
  • Jennifer Ngadiuba
  • Jesse Thaler
  • Jessie Micallef
  • Jieun Yoo
  • Jigarkumar Patel
  • Jinglu Wang
  • Jiri Kvita
  • Joe Egan
  • Joe McLaughlin
  • John Strologas
  • Jonas Spinner
  • Jonathan Pampel
  • Jonathan Pritchard
  • Jonathon Mark Langford
  • Josh Villarreal
  • Judita Mamuzic
  • Kai Hong Law
  • Kai Lehman
  • Kai Mulcock
  • Karolina Wresilo
  • Keng Lin
  • Kevin Greif
  • Kieren Joseph
  • Kiyam Lin
  • Klitos Savva
  • Konstantin Karchev
  • Kyungmin Park
  • Labh Singh
  • Laura Aguilar
  • Lilah Mercadante
  • Lily Zhang
  • Linh Nguyễn Vũ
  • Liv Helen Vage
  • Lorenzo Perisse
  • Louis Lyons
  • lubna ibrahim
  • Luca Tabarroni
  • Lucas Johnny Monte Tamayo
  • Lucas Kang
  • Lucas Karwatzki
  • Lucas Makinen
  • Luisa Lucie-Smith
  • Luiz Regis Emediato
  • Lydia Brenner
  • Lázaro Maximiliano Salas Valtierra
  • Lénea Luís
  • Maitreyee Moudgalya
  • Mapse Barroso Ferreira Filho
  • Marcelo Gameiro Munhoz
  • Marco Letizia
  • Maria Gabriela Ferreira Siqueira Amaral Gomes
  • Mariano Dominguez
  • Marienza Caldarola
  • Martin Makler
  • Martin Vala
  • Martín de los Rios
  • Marvin Pfaff
  • Massimiliano Galli
  • Matin Torkian
  • Matteo Agostini
  • Matthew Mould
  • Mauro Donega
  • Maxim Virta
  • Maximilian Autenrieth
  • Maximilian Dax
  • Md Sarfaraz
  • Meet Vyas
  • Meliha Karataş
  • Metin Yalvac
  • Michael Kagan
  • Michael Kent Wilkinson
  • Miguel Alfonso Martinez Barragan
  • Mikael Kuusela
  • Milagros Colazo
  • Mohamed Aly
  • Mohamed Ismail
  • Mohammad Abrar Wadud
  • Monika Machalová
  • Muhammad Farooq
  • Nadia Cooper
  • Natalia Zubova
  • Navneeth C
  • Nazanin Shiri
  • Niall Jeffrey
  • Nico Härringer
  • Nicola Fulvio Calabria
  • Nilay Bostan
  • Noah Clarke Hall
  • Noam Levi
  • Norman Martinez Figueroa
  • Octavio Pacheco Vazquez
  • Olaf Behnke
  • Oliver Rieger
  • Orhan Cakir
  • Paridhi Jain
  • Paris Sphicas
  • Patricia Rebello Teles
  • Paul Felix Kruper
  • Pedro Allendes S.
  • Pegah Golestaneh
  • Pegah Golestaneh
  • Petr Baron
  • Phelipe Antonie Darc de matos
  • Pierre Jouvelot
  • Pietro Vischia
  • Prachi Sharma
  • Prijith Babu Pradeep
  • Priya Yadav
  • Priyankush Deka
  • Pueh Leng Tan
  • Purvasha Chakravarti
  • Pushpalatha Bhat
  • Quratulain Zahoor
  • Rahmat Rahmat
  • RAHUL SHAW
  • Rahul Srinivasan
  • Ram Lal Awasthi
  • Ramon Winterhalder
  • Raymond Wynne
  • Rethabile Kgolobe
  • Rob Bainbridge
  • Robert Cousins
  • Robert Schoefbeck
  • Rodrigo Gamboa Goñi
  • Roger Wolf
  • Rohan Rajagopalan
  • Rui Zhang
  • Rui Zou
  • Ryan Nichol
  • Saad MOHIUDDIN
  • Safiye Kamile Eker
  • SAI KRISHNA PODEM
  • Sameer Patil
  • Samia Tanveer
  • Samrit Pramanik
  • Samuel Naugle
  • Samuel Sebastian
  • Samuele Grossi
  • Sanae Ezzarqtouni
  • Sandeep Pradhan
  • Sanmay Ganguly
  • Sebastian Schmitt
  • Seo Won Yi
  • Sergei Gleyzer
  • Sergei Zakharov
  • Shalini Epari
  • Shashank Samavedula
  • Shiqian Hu
  • Shivam Raj
  • Shreya Sharma
  • Siddharth Chaini
  • Silvio Motta de Jesus da Silva
  • Somayeh Khakpash
  • Sophia Dorothy Scarano
  • Spyros Nessis
  • Stephen Pate
  • Stephen Watts
  • Suber Corley
  • Sunny Seo
  • Syaefudin Jaelani
  • Süleyman Emir Yılmaz
  • TARIK AKAN
  • Tatiana Acero Cuellar
  • Thiago Badaró Saramela
  • Thiago De Andrade Rangel Monteiro
  • Thomas Junk
  • Tilman Plehn
  • Tim Bolton
  • Tobias Golling
  • Tobias Kortus
  • Toby Dixon
  • Tom Loredo
  • Toni Mlinarevic
  • Vasilis Belis
  • Volker Andreas Austrup
  • Waleed Esmail
  • Wi Han Ng
  • Will Handley
  • Wilson David Buitrago Ceballos
  • Wilson David Buitrago Ceballos
  • Wojciech Jan Krzemien
  • Won-Kee Park
  • Wouter Verkerke
  • Xiangxuan Zheng
  • Young-Lo Kim
  • ZEYNEP ECE OKATAN
  • Zubair Bhatti
  • +206
    • 13:15 13:45
      Registration
    • 13:45 15:45
      Introductory talks. These optional talks are aimed at people who feel that they would like more background introductory material, before the next 3 days' taks.

      (Optional)

      Convener: Benedikt Maier (Imperial College (GB))
    • 15:45 16:15
      Break: Coffee/Registration
    • 16:15 17:35
      Introductory talks. These optional talks are aimed at people who feel that they would like more background introductory material, before the next 3 days' taks.

      (Optional)

      Convener: Matteo Agostini
    • 08:30 09:00
      Registration
    • 09:00 10:30
      Talks
      Convener: Wouter Verkerke (Nikhef National institute for subatomic physics (NL))
      • 09:00
      • 09:15
        ML in Particle Physics 45m

        Machine learning and AI have quickly turned into indispensable tools for modern particle physics. They both greatly amplify the power of existing techniques - such as supercharging supervised classification - and enable qualitatively new ways of extracting information - such as anomaly detection and likelihood-free inference. Accordingly, the underlying statistical machinery needs to be understood in greater detail or in some cases newly developed.

        After briefly introducing the environment of collider based particle physics, this talk will review key developments in machine learning applied to data analysis with a special eye on statistical challenges.

        Speaker: Gregor Kasieczka (Hamburg University (DE))
      • 10:00
        Lorentz-Equivariant Geometric Algebra Transformers for High-Energy Physics 30m

        Extracting scientific understanding from particle-physics experiments requires solving diverse learning problems with high precision and good data efficiency. We propose the Lorentz Geometric Algebra Transformer (L-GATr), a new multi-purpose architecture for high-energy physics. L-GATr represents high-energy data in a geometric algebra over four-dimensional space-time and is equivariant under Lorentz transformations, the symmetry group of relativistic kinematics. At the same time, the architecture is a Transformer, which makes it versatile and scalable to large systems. L-GATr is first demonstrated on regression and classification tasks from particle physics. We then construct the first Lorentz-equivariant generative model: a continuous normalizing flow based on an L-GATr network, trained with Riemannian flow matching. Across our experiments, L-GATr is on par with or outperforms strong domain-specific baselines.

        Speaker: Jonas Spinner
    • 10:30 11:00
      Break: Coffee/Registration
    • 11:00 12:30
      Talks
      Convener: Alan Heavens
      • 11:00
        Cosmology and machine learning 45m

        Cosmologists strive to uncover the mysteries of the origin, composition, evolution, and fate of the cosmos from all the information the sky has to offer: the cosmic microwave background, galaxy surveys, exploding stars, and reverberations of space-time caused by colliding black holes and neutron stars. I will discuss new ways to connect cosmological theory and simulation with these data sets. Novel cosmological tests promise insights to classical cosmological questions; and Artificial Intelligence (AI) and Machine Learning are revolutionizing our ability to confront computational models with data, enabling end-to-end, quantitative Bayesian reasoning for problems that were previously deemed intractable. Very recently, AI has even begun to inspire novel cosmological insights. I will discuss the current status, promises, and challenges and outline a path towards achieving the goals of reconstructing the detailed initial conditions of the universe at its cosmic beginning and understanding the formation of cosmic structures much more completely than ever before.

        Speaker: Ben Wandelt
      • 11:45
        Simulation-based machine learning for gravitational-wave analysis 45m

        Gravitational waves (GWs) provide a unique window to the universe, enabling us to study mergers of black holes and/or neutron stars. In my talk, I will highlight how machine learning can address critical limitations in GW data analysis. I will present key innovations in this field, driven by unusually high requirements for accuracy, reliability and interpretability. Finally, I will discuss how insights gained from GW science transfer to other application domains of machine learning.

        Speaker: Maximilian Dax
    • 12:30 14:00
      Break: Lunch
    • 14:00 15:30
      Talks
      Convener: Sergei Gleyzer (University of Alabama (US))
      • 14:00
        Interpretable Machine Learning for Particle Physics 45m

        The term "interpretability" encompasses various strategies to scrutinize the decisions made by machine learning algorithms. In this talk, I argue that interpretability, at least in the context of particle physics, should be considered as part of the broader goal of assessing systematic uncertainties. I provide examples from my own research on jet physics at the Large Hadron Collider, where some of the goals of interpretability can be achieved through modified machine learning architectures and training paradigms. I also comment on the growing interest in foundation models, which are forcing us to rethink how we specify machine learning tasks and whether there can be statistical gains from incorporating auxiliary information.

        Speaker: Jesse Thaler (MIT/IAIFI)
      • 14:45
        Understanding and mitigating failures in anomaly detection: a probabilistic perspective 45m

        In this talk, we present an overview of anomaly detection from a probabilistic machine learning perspective, with a focus on work emerging from the machine learning literature. First, we discuss empirical failures of deep generative models for anomaly detection and why they occur, as well as their implications for deep generative modeling and anomaly detection. Then, we discuss the endeavor of robust anomaly detection and what is required to achieve it. We conclude with recent work that applies these insights to jet anomaly detection in high-energy physics.

        Speaker: Lily Zhang
    • 15:30 16:00
      Break: Coffee
    • 16:00 18:00
      Talks
      Convener: Robert Cousins Jr
      • 16:00
        Statistical tests for anomaly detection at the LHC 45m

        Signal-agnostic data exploration could unveil very subtle statistical deviations of collider data from the expected Standard Model of particle physics. However, the extreme size, rate and complexity of the datasets generated at the Large Hadron Collider (LHC) pose unique challenges for data analysis. Making assumptions about what is relevant becomes unavoidable to scale the information down to a human readable level.

        Machine learning can be exploited at multiple stages in the experimental pipeline to reduce these assumptions, trading assumed knowledge for a richer basis of acquired experimental evidence. In this talk I will discuss the main challenges behind the design of efficient, robust and at the same time interpretable statistical tests for anomaly detection at the LHC, such as model selection, uncertainty quantification and scalability. I will present recent solutions and future prospects based on existing literature, and examples from my own research on the Neyman-Pearson strategy for signal-agnostic likelihood-ratio-test.

        Speaker: Gaia Grosso
      • 16:45
        Detecting New Physics as data anomalies at the LHC: Transitioning from small-scale toy datasets to millions of complex proton collisions 25m

        Anomaly detection has emerged as a promising technique for identifying subtle New Physics signals amidst a dominant Standard Model background. Due to the novelty of these techniques, they are often proposed and demonstrated on toy datasets that mimic real LHC data before being deployed in actual experiments. In this talk, we will discuss the challenges encountered during the transition from research and development to practical implementation in experiments, highlighting lessons learned and future prospects. Various techniques will be compared, including outlier detection versus over-density detection, along with different levels of supervision, ranging from weakly supervised to semi-supervised and unsupervised methods. We will address the statistical challenges involved in applying these methods for data analysis and ultra-fast event filtering in the CMS Experiment at CERN. The discussion will focus on the robustness of these methods and the challenges associated with validating them.

        Speaker: Thea Aarrestad (ETH Zurich (CH))
      • 17:10
        Anomaly aware machine learning for dark matter direct detection at the DARWIN experiment 25m

        This talk presents a novel approach to dark matter direct detection using anomaly-aware machine learning techniques in the DARWIN next-generation dark matter direct detection experiment. I will introduce a semi-unsupervised deep learning pipeline that falls under the umbrella of generalized Simulation-Based Inference (SBI), an approach that allows one to effectively learn likelihoods straight from simulated data, without the need for complex functional dependence on systematics or nuisance parameters. I also present an inference procedure to detect non-background physics utilizing an anomaly function derived from the loss functions of the semi-unsupervised architecture. The pipeline’s performance is evaluated using pseudo-data sets in a sensitivity forecasting task, and the results suggest that it offers improved sensitivity over traditional methods.

        Speaker: Andre Joshua Scaffidi
      • 17:35
        Feldman-Cousins’ ML Cousin 25m

        The statistical treatment of sterile neutrino searches suffers from the fact that Wilks’ theorem, a beneficial simplifying assumption, does not hold across all regions of parameter space. The alternative, the Feldman-Cousins algorithm, suffers from expensive computational run times that prohibit its application into many-experiment global fits. This contribution introduces a deep learning-based method (which does not assume Wilks’ theorem) that can fit electron (anti)neutrino disappearance experiments in a tractable amount of time. Though this procedure’s utility for sterile neutrino searches are presented here, it will be useful for a variety of particle physics analyses.

        Speaker: Joshua Villarreal
    • 18:00 19:15
      Social: Poster Session & Welcome Drinks
      • 18:00
        Uncertainty-aware machine learning for the LHC 1m

        Estimating uncertainties is a fundamental aspect in every physics problem, no measurements or calculations comes without uncertainties. Hence it is crucial to consider the effect of training a neural network to problems in physics. I will present our work on amplitude regression, using loop amplitudes from LHC processes, as an example to examine the impact of different uncertainties on the outcome of the network. We test the behavior of different neural networks with uncertainty estimation, including Bayesian neural networks and repulsive ensembles.

        Speaker: Nina Elmer
      • 18:01
        Generative models: their evaluation and their limitations 1m

        I will present and discuss several proposed metrics, based on integral probability measures, for the evaluation of generative models (and, more generally, for the comparison of different generators). Some of the metrics are particularly efficient to be computed in parallel, and show good performances. I will first compare the metrics on toy multivariate/multimodal distributions, and then focus on HEP examples from the JetNet jet dataset. I will discuss the power of the tests and their implementation in TensorFlow2, making the opportunity to discuss, more generally, the limitations in the usage of generative models in HEP.

        Speaker: Samuele Grossi (Università degli studi di Genova & INFN sezione di Genova)
      • 18:02
        Limits to classification performance by relating Kullback-Leibler divergence to Cohen’s Kappa 1m

        The performance of machine learning classification algorithms are evaluated by estimating metrics, often from the confusion matrix, using training data and cross-validation. However, these do not prove that the best possible performance has been achieved. Fundamental limits to error rates can be estimated using information distance measures. To this end, the confusion matrix has been formulated to comply with the Chernoff-Stein Lemma. This links the error rates to the Kullback-Leibler divergences between the probability density functions describing the two classes. This leads to a key result that relates Cohen’s Kappa to the Resistor Average Distance which is the parallel resistor combination of the two Kullback-Leibler divergences. The Resistor Average Distance has units of bits and is estimated from the same training data used by the classification algorithm, using kNN estimates of the Kullback-Leibler divergences. The classification algorithm gives the confusion matrix and Kappa. Theory and methods are discussed in detail and then applied to Monte Carlo data and real datasets. Four very different real datasets - Breast Cancer, Coronary Heart Disease, Bankruptcy, and Particle Identification - are analysed, with both continuous and discrete values, and their classification performance compared to the expected theoretical limit. In all cases this analysis shows that the algorithms could not have performed any better due to the underlying probability density functions for the two classes. Important lessons are learnt on how to predict the performance of algorithms for imbalanced data. Preprint available at arXiv:2403.01571.

        Speaker: Stephen Watts
      • 18:03
        Graph neural networks on the test bench in HEP applications 1m

        Data analyses in the high-energy particle physics (HEP) community more and more often exploit advanced multivariate methods to separate signal from background processes. In this talk, a maximally unbiased, in-depth comparison of the graph neural network (GNN) architecture, which is of increasing popularity in the HEP community, with the already well-established technology of fully connected feed-forward deep neural networks (DNNs) is presented. When it comes to choosing a suitable machine-learning model, it is not a priori clear, what model this should be to benefit from inherent properties of the task. Also, the design of a fair and unbiased benchmark is non-trivial. This GNN vs. DNN comparison is insightful in terms of the details it reveals as to which aspects of GNNs are superior to DNNs - and which are not. The study is performed on a typical data set of a complex challenge recently faced at the Large Hadron Collider: the classification of events with top quark-antiquark pairs with additional heavy flavour jets originating from gluon splittings, Z or Higgs bosons.

        The study is documented in the paper “A Case Study of Sending Graph Neural Networks Back to the Test Bench for Applications in High‐Energy Particle Physics” published in Computing and Software for Big Science, https://doi.org/10.1007/s41781-024-00122-3.

        Speaker: Emanuel Lorenz Pfeffer (KIT - Karlsruhe Institute of Technology (DE))
      • 18:04
        Interpolated Likelihoods for Fast Reinterpretations 1m

        We present a method to accelerate Effective Field Theory reinterpretations using interpolated likelihoods. By employing Radial Basis Functions for interpolation and Gaussian Processes to strategically select interpolation points, we show that we can reduce the computational burden while maintaining accuracy. We apply this in the context of the Combined Higgs Boson measurement at CMS, a complex statistical model with many thousands of parameters requiring large computing power to evaluate.

        Speaker: Tom Runting (Imperial College (GB))
      • 18:05
        Efficient machine learning for statistical hypothesis testing 1m

        Traditional statistical methods are often not adequate to perform inclusive and signal-agnostic searches at modern collider experiments delivering large amounts of multivariate data. Machine learning provides a set of tools to enhance analyses in large scale regimes, but the adoption of these methodologies comes with new challenges, such as the lack of efficiency and robustness, and potential hidden biases. In this talk, I will discuss these aspects in the context of a recent proposal for a likelihood-ratio-based goodness-of-fit test powered by large-scale implementations of kernel methods, nonparametric learning models that can approximate any continuous function given enough data.

        Speaker: Dr Marco Letizia
      • 18:06
        Integrating Explainable AI in Data Analyses of ATLAS Experiment at CERN 1m

        “The Multi-disciplinary Use Cases for Convergent Approaches to AI Explainability (MUCCA) project is pioneering efforts to enhance the transparency and interpretability of AI algorithms in complex scientific fields. This study focuses on the application of Explainable AI (XAI) in high-energy physics (HEP), utilising a range of machine learning (ML) methodologies, from classical boosted decision trees to Graph Neural Networks (GNNs), to explore new physics models.

        Our work leverages case studies from the ATLAS experiment at the Large Hadron Collider (LHC) to demonstrate the potential of ML in improving sensitivity to new physics. Notably, GNNs outperformed traditional Convolutional Neural Networks (CNNs), with the DarkJetGraphs code achieving a 2-5% improvement in detection accuracy and twice the background rejection efficiency. These findings affirm the value of GNNs in extending the search space for new physics models that conventional methods may not adequately capture. Balancing the use of cutting-edge ML with transparency and interpretability through XAI techniques is critical to ensuring both scientific rigor and robust result interpretation.

        The presented research highlights this balance, with further developments in XAI techniques, including Kappa pruning and differentiable programming, proved to further enhance performance. A full publication of these methods and results is anticipated in Autumn 2024.

        Speaker: Joseph Carmignani (University of Liverpool (GB))
      • 18:07
        Proximal Nested Sampling with Data-Driven AI Priors 1m

        Bayesian model selection provides a powerful framework for objectively comparing models directly from observed data, without reference to ground truth data. However, Bayesian model selection requires the computation of the marginal likelihood (model evidence), which is computationally challenging, prohibiting its use in many high-dimensional Bayesian inverse problems. With Bayesian imaging applications in mind, we introduce the proximal nested sampling methodology to objectively compare alternative Bayesian imaging models for applications that use images to inform decisions under uncertainty. The methodology is based on nested sampling, a Monte Carlo approach specialised for model comparison, and exploits proximal Markov chain Monte Carlo techniques to scale efficiently to large problems and to tackle models that are log-concave and not necessarily smooth (e.g., involving l1 or total-variation priors). Taking one step further, we show how proximal nested sampling can be extended using Tweedie’s formula to support data-driven priors, such as deep neural networks learned from training data. We demonstrate our method by carrying out Bayesian model comparison between data-driven and hand-crafted priors in imaging applications like radio-interferometric image reconstruction. Based on arXiv:2106.03646 and arXiv:2307.00056

        Speaker: Henry Aldridge (UCL)
      • 18:08
        Generative models of astrophysical fields with scattering transforms on the sphere 1m

        Scattering transforms are a new type of summary statistics recently developed for the study of highly non-Gaussian processes, which have been shown to be very promising for astrophysical studies. In particular, they allow one to build generative models of complex non-linear fields from a limited amount of data, and have also been used as the basis of new statistical component separation algorithms. In the context of upcoming cosmological surveys, such as LiteBIRD for the cosmic microwave background polarization or Rubin-LSST and Euclid for study of the large scale structures of the Universe, the extension of these tools to spherical data is necessary. We develop scattering transforms on the sphere and focus on the construction of maximum-entropy generative models of several astrophysical fields. We construct, from a single target field, generative models of homogeneous astrophysical and cosmological fields, whose samples are quantitatively compared to the target fields using common statistics (power spectrum, pixel probability density function and Minkowski functionals). Our sampled fields agree well with the target fields, both statistically and visually. These generative models therefore open up a wide range of new applications for future astrophysical and cosmological studies; particularly those for which very little simulated data is available. We make our code available to the community so that this work can be easily reproduced and developed further.

        Speaker: Matt Price
      • 18:09
        Advanced techniques for Simulation Based Inference in collider physics 1m

        We present an application of Simulation-Based Inference (SBI) in collider physics, aiming to constrain anomalous interactions beyond the Standard Model (SM). This is achieved by leveraging Neural Networks to learn otherwise intractable likelihood ratios. We explore methods to incorporate the underlying physics structure into the likelihood estimation process. Specifically, we compare two approaches: morphing-aware likelihood estimation and derivative learning. Furthermore, we illustrate how uncertainty-aware networks can be employed to compare the performance of these methods. Additionally, we demonstrate two new techniques for enhancing the accuracy and reliability of the network training. First, we introduce of a new way to treat the outliers in the target reconstruction-level distributions by repeated smearing and modifying their parton-level weights accordingly (dubbed fractional smearing). Second, we utilise Lorentz-equivariant network architectures to exploit the symmetry structure inherent in the underlying particle physics amplitudes.

        Speaker: Giovanni De Crescenzo (University of Heidelberg)
      • 18:10
        SBI for wide field weak lensing 1m

        The standard approach to inference from cosmic large-scale structure data employs summary statistics that are compared to analytic models in a Gaussian likelihood with pre-computed covariance. To overcome many of the idealising assumptions that go into this type of explicit likelihood inference, and to take advantage of the high-fidelity wide field data that Euclid and LSST will provide, we can employ simulation-based inference (SBI). In previous work we have demonstrated the power of SBI in the context of performing a full re-analysis of the KiDS-1000 survey (MvWK & KL 2024) whilst modelling anisotropic observational systematics with forward simulations that significantly impact the outcome of the inference and must be included in the analysis of Euclid and LSST data. We further report on the effects of different levels of Gaussianity imposed on the inference. Our current work explores the use of wavelet summary statistics to construct higher order summary statistics directly from the field without any machine learning. This method thus does not suffer from many of the problems that are associated with using neural methods whilst being fast and capable of extracting large amounts of information. We present here an outlook of how we can tackle the task of SBI applied to upcoming wide field weak lensing surveys.

        Speaker: Kiyam Lin
      • 18:11
        Exhaustive Symbolic Regression: Learning Astrophysics directly from Data 1m

        A key challenge in the field of AI is to make machine-assisted discovery interpretable, enabling it not only to uncover correlations but also to improve our physical understanding of the world. A nascent branch of machine learning – Symbolic Regression (SR) – aims to discover the optimal functional representations of datasets, producing perfectly interpretable outputs (equations) by construction.

        SR is traditionally done using a “genetic algorithm”which stochastically selects trial functions by analogy with natural selection; I will describe the more ambitious approach of exhaustively searching and evaluating function space.

        Coupled to an information-theoretic model selection principle based on minimum description length, our algorithm “Exhaustive Symbolic Regression” (ESR) is guaranteed to find the simple functions that optimally balance accuracy with simplicity on a dataset. This gives it broad application across science. I will detail the method, its relation to Bayesian statistics and an optional language model-based prior on functions designed to enhance their physicality. Then I will use ESR to quantify the extent to which state-of-the-art astrophysical theories – FLRW cosmology, General Relativity and Inflation – are implied by the current data.

        Speaker: Harry Desmond (University of Portsmouth)
      • 18:12
        Usage of weakly correlated observables for nuisance parameter fits 1m

        Precision measurements at the Large Hadron Collider (LHC), such as the measurement of the top quark mass, are essential for advancing our understanding of fundamental particle physics. Profile likelihood fits have become the standard method to extract physical quantities and parameters from the measurements. These fits incorporate nuisance parameters to include systematic uncertainties. The results depend critically on the selection of observables. Including multiple observables from the measurements is beneficial for precision, as it helps to restrict the nuisance parameters, leading to more reliable fits. Usually, the used observables are assumed to be independent; however, including more observables can introduce correlations that complicate the analysis, as these correlations violate the assumption of independence. At the current precision of the top quark mass measurement, introducing more observables with minor correlations might already lead to a significant distortion of the results of the profile likelihood fit. This project aims to investigate the threshold of correlation at which the accuracy of likelihood fits begins to degrade. We utilize the realistically correlated reconstructed top and W mass, applying an uncorrelated single event likelihood fit for the analysis. To assess the accuracy of our fits, we calculate the pull for the nuisance parameters alongside the top quark mass distribution. Subsequently, we will explore machine learning techniques, such as normalizing flows, that take correlations into account instead of avoiding them.

        Speaker: Lars Stietz (Hamburg University of Technology (DE))
      • 18:13
        Accounting for Selection Effects in Supernova Cosmology with Simulation-Based Inference and Hierarchical Bayesian Modelling 1m

        Type Ia supernovae (SNe Ia) are thermonuclear exploding stars that can be used to put constraints on the nature of our universe. One challenge with population analyses of SNe Ia is Malmquist bias, where we preferentially observe the brighter SNe due to limitations of our telescopes. If untreated, this bias can propagate through to our posteriors on cosmological parameters. In this work, we develop a novel technique of using a normalising flow to learn the non-analytical likelihood of observing a SN Ia for a given survey from simulations, that is independent of any cosmological model. The learnt likelihood is then used in a hierarchical Bayesian model with Hamiltonian Monte Carlo sampling to put constraints on different sets of cosmological parameters conditioned on the observed data. The technique is verified on toy model simulations finding excellent agreement with analytically-derived posteriors to within 1σ. We conclude by discussing plans to show the generalisation of our flexible method to real survey selection effects where analytical solutions are intractable.

        Speaker: Benjamin Boyd (University of Cambridge)
      • 18:14
        COmoving Computer Acceleration (COCA): Correcting Emulation Errors for Trustworthy N-Body Simulations 1m

        Neural networks are increasingly used to emulate complex simulations due to their speed and efficiency. Unfortunately, many ML algorithms, including (deep) neural networks, lack interpretability. If machines predict something humans do not understand, how can we check (and trust) the results? Even if we could identify potential mistakes, current methods lack effective mechanisms to correct them, limiting the reliability of these emulators. To address these issues, we introduce COmoving Computer Acceleration (COCA), a novel hybrid framework that integrates machine learning with traditional N-body simulators. COCA solves the correct physical equations of motion within an emulated frame of reference, inherently correcting any prediction errors. Our framework significantly reduces emulation errors in particle trajectories but also requires far fewer force evaluations compared to conventional simulations. This method effectively addresses the critical challenges of interpretability and accuracy in cosmological applications of machine learning, ensuring both speed and trustworthiness in complex simulations.

        Speaker: Deaglan Bartlett (Institut d'Astrophysique de Paris)
      • 18:15
        Application of Machine Learning Based Top Quark and Jet Tagging to Hadronic Four-Top Final States Induced by SM as well as BSM Processes 1m

        The aim of this work is to solve the problem of hadronic jet substructure recognition using classical subjettiness variables available in the parameterized detector simulation package, Delphes. Jets produced in simulated proton-proton collisions are identified as either originating from the decay of a top quark or a W boson and are used to reconstruct the mass of a hypothetical scalar resonance decaying into a pair of top quarks in events where a total of four top quarks are produced. We compare a simple cut-based tagging method for the stacked histograms of a mixture of the Standard Model and new physics processes with a multi-layer perceptron classifier and a gradient boosting classifier. Due to the sufficient amount of data, we applied various undersampling techniques to the training sets. Our findings demonstrate that gradient boosting methods provide better results than the other tested approaches.

        Speaker: Monika Machalová
      • 18:16
        Accelerating High-Dimensional Cosmological Inference with COSMOPOWER 1m

        A new generation of astronomical surveys, such as the recently launched European Space Agency’s Euclid mission, will soon deliver exquisite datasets with unparalleled amounts of cosmological information, poised to change our understanding of the Universe. However, analysing these datasets presents unprecedented statistical challenges. Multiple systematic effects need to be carefully accounted for, which would otherwise lead to incorrect physical interpretations. Efficiently navigating the large, complex parameter spaces required for robust systematics modelling is critical to ensuring that the anticipated increase in precision translates into equally enhanced accuracy in the final cosmological measurements.

        My software COSMOPOWER tackles this challenge by training neural networks to emulate the computationally intensive calculation of cosmological power spectra. The emulation produces orders-of-magnitude acceleration in the inference pipeline. COSMOPOWER also enables the creation of fully-differentiable pipelines to leverage gradient-based sampling methods, which scale efficiently to high-dimensional parameter spaces. The acceleration provided by COSMOPOWER has led major international collaborations (e.g. Euclid, KiDS, ACT, Simons Observatory), working on both large-scale structure and Cosmic Microwave Background data, to adopt this software in their cosmological inference pipelines.

        In this poster, I will showcase how recent advancements in COSMOPOWER pave the way for a new paradigm in cosmological inference, allowing for comprehensive Bayesian analyses —including both parameter estimation and model selection —to be conducted in a fraction of the time required by traditional methods, and extracting more information from the data than the traditional approach based on two-point statistics. The framework not only accelerates the inference process but also improves our ability to accurately model the underlying physics, especially for beyond-standard models, making it a critical tool for the future of cosmological research.

        Speaker: Alessio Spurio Mancini (Royal Holloway, University of London)
      • 18:17
        Learning Optimal and Interpretable Summary Statistics of Galaxy Catalogs with SBI 1m

        How much cosmological information can we reliably extract from existing and upcoming large-scale structure observations? Many summary statistics fall short in describing the non-Gaussian nature of the late-time Universe and modelling uncertainties from baryonic physics. Using simulation based inference (SBI) with automatic data-compression from graph neural networks, we learn optimal summary statistics for galaxy catalogs in the context of cosmological parameter estimation. By construction these summaries do not require the ability to write down an explicit likelihood. We demonstrate that they can be used for efficient parameter inference, outperforming existing (ML) methods for the same parameter estimation. These summary statistics offer a new avenue for analyzing different simulation models for baryonic physics with respect to their relevance for the resulting cosmological features. The learned summary statistics are low-dimensional, feature the underlying simulation parameters, and are similar across different network architectures. To link our models, we identify the relevant scales associated to our summary statistics (e.g. in the range of modes between k = 5 − 30h/M pc) and we are able to match the summary statistics to underlying simulation parameters across various simulation models. Furthermore, we compare different baryonic feedback models in latent space and find differences in the flexibility of their parametrizations.

        Speaker: Kai Lehman (LMU Munich)
      • 18:18
        How to Unfold Top Decays 1m

        Many physics analyses at the LHC rely on algorithms to remove detector effect, commonly known as unfolding. Whereas classical methods only work with binned, one-dimensional data, Machine Learning promises to overcome both problems. Using a generative unfolding pipeline, we show how it can be build into an existing LHC analysis, designed to measure the top mass. We discuss the model-dependence of our algorithm, i.e. the bias of our measurement towards the top mass used in simulation and propose a method to reliably achieve unbiased results.

        Speakers: Sofia Palacios Schweitzer (Heidelberg), Tilman Plehn (Heidelberg University)
      • 18:19
        Noise injection node regularization for robust learning 1m

        We introduce Noise Injection Node Regularization (NINR), a method that injects structured noise into Deep Neural Networks (DNNs) during the training stage, resulting in an emergent regularizing effect. We present both theoretical and empirical evidence demonstrating substantial improvements in robustness against various test data perturbations for feed-forward DNNs trained under NINR. The novelty of our approach lies in the interplay between adaptive noise injection and initialization conditions, such that noise becomes the dominant driver of dynamics at the start of training. Since this method simply requires the addition of external nodes without altering the existing network structure or optimization algorithms, it can be easily incorporated into many standard problem specifications. We observe improved stability against a range of data perturbations, including domain shifts, with the most dramatic improvement occurring for unstructured noise, where our technique outperforms existing methods such as Dropout or L2 regularization in some cases. Additionally, we show that desirable generalization properties on clean data are generally maintained. This method is well-suited for many physical scenarios where robust predictions are critical to neural network performance. Currently, we are employing this method to improve networks’ ability to discriminate between prompt and non-prompt photons in highly noisy processes in the most recent simulations of the ATLAS detector.

        Speaker: Noam Levi (Tel Aviv University)
      • 18:20
        Modeling Smooth Backgrounds at Collider Experiments With Log Gaussian Cox Processes 1m

        Background modeling is one of the critical elements of searches for new physics at experiments at the Large Hadron Collider. In many searches, backgrounds are modeled using analytic functional forms. Finding an acceptable function can be complicated, inefficient and time-consuming. This poster presents a novel approach to estimating the underlying PDF of a 1D dataset of samples using Log Gaussian Cox Processes (LGCP). Using LGCP allows inferring a posterior distribution for the PDF, and unlike a Gaussian Process fit, it does not require binning the dataset. Markov Chain Monte Carlo (MCMC) is used to optimize the LCGP fit. The final result is a fit model which is highly flexible, does not rely on an analytic fit assumption, and carries uncertainty bands.

        Speaker: Yuval Yitzhak Frid (Tel Aviv University (IL))
      • 18:21
        Precision Machine Learning for the Matrix Element Method 1m

        The matrix element method is the LHC inference method of choice for limited statistics, as it allows for optimal use of available information. We present a dedicated machine learning framework, based on efficient phase-space integration, a learned acceptance and transfer function. It is based on a choice of INN and diffusion networks, and a transformer to solve jet combinatorics. We showcase this setup for the CP-phase of the top Yukawa coupling in associated Higgs and single-top production.

        Speakers: Nathan Huetsch (Heidelberg), Tilman Plehn (Heidelberg University)
      • 18:22
        The Landscape of Unfolding with Machine Learning 1m

        Recent innovations from machine learning allow for data unfolding, without binning and including correlations across many dimensions. We describe a set of known, upgraded, and new methods for ML-based unfolding. The performance of these approaches are evaluated on the same two datasets. We find that all techniques are capable of accurately reproducing the particle-level spectra across complex observables. Given that these approaches are conceptually diverse, they offer an exciting toolkit for a new class of measurements that can probe the Standard Model with an unprecedented level of detail and may enable sensitivity to new phenomena

        Speakers: Tilman Plehn (Heidelberg University), Xavier Marino (Heidelberg)
    • 09:00 10:35
      Talks
      Convener: Daniel Mortlock
      • 09:00
        Theoretical and Societal Topics in AI and Deep Learning for Physicists 45m

        I plan to touch on several theoretical topics (overparameterization, neural balance, attention and transformers) and their applications in physics and end on a proposal to solve some of the societal issues raised by AI inspired by physics.

        Speaker: Pierre Baldi
      • 09:45
        AI Benchmarks and Science 25m
        Speaker: Jeyan Thiyagalingam (Rutherford Appleton Laboratory, Science and Technology Facilities Council)
      • 10:10
        Improved Weak Lensing Photometric Redshift Calibration via StratLearn and Hierarchical Modeling 25m

        Discrepancies between cosmological parameter estimates from cosmic shear surveys and from recent Planck cosmic microwave background measurements challenge the ability of the highly successful ΛCDM model to describe the nature of the Universe. To rule out systematic biases in cosmic shear survey analyses, accurate redshift calibration within tomographic bins is key. In this work, we improve photo-z calibration via Bayesian hierarchical modeling of full galaxy photo-z conditional densities, by employing StratLearn, a recently developed statistical methodology, which accounts for systematic differences in the distribution of the spectroscopic training/source set and the photometric target set. Using realistic simulations that were designed to resemble the KiDS+VIKING-450 dataset, we show that StratLearn-estimated conditional densities improve the galaxy tomographic bin assignment, and that our StratLearn-Bayesian framework leads to nearly unbiased estimates of the target population means. This leads to a factor of ∼ 2 improvement upon often used and state-of-the-art photo-z calibration methods. Our approach delivers a maximum bias per tomographic bin of ∆⟨z⟩ = 0.0095 ± 0.0089, with an average absolute bias of 0.0052 ± 0.0067 across the five tomographic bins.

        Speaker: Maximilian Autenrieth (Imperial College London)
    • 10:35 11:00
      Break: Coffee
    • 11:00 12:30
      Talks
      Convener: Olaf Behnke (Deutsches Elektronen-Synchrotron (DE))
      • 11:00
        Simulation-based Inference (SBI) 45m

        A powerful class of statistical inference methods are starting to be used in across fields that leverage the power of machine learning (ML) to perform inference directly from high-dimensional data. They can be used, for instance, to estimate fundamental physics parameters from data collected in high energy physics experiments, or cosmological / astrophysics observations and work with both Bayesian and frequentist settings. In these application cases, the exact likelihood itself is typically intractable, but it can be sampled from using sophisticated forward simulation software. ML models can then learn to estimate statistical quantities such as posteriors, likelihoods or likelihood ratios from simulated samples and also exploit additional knowledge available at intermediate steps of the simulation, which are inaccessible in real data. This talk will focus on important statistical questions with regard to the application of SBI, with a bias towards frequentist statistical framework typically used in high-energy physics.

        Speaker: Aishik Ghosh (University of California Irvine (US))
      • 11:45
        Extending Unfolding Methods With Machine Learning 45m

        Correcting experimental measurements for detector effects, or unfolding, is a standard technique used at the LHC to report multi-differential cross section measurements. These techniques rely on binned data and are limited to low dimensional observables. In this talk, I will cover recent ideas to extend standard methods of unfolding using machine learning, enabling the measurements of unbinned and high-dimensional differential cross sections. These include methods using classifiers as approximators to the likelihood ratio, generative models, and high dimensional interpolation techniques.

        Speaker: Vinicius Mikuni (LBL)
    • 12:30 14:00
      Break: Lunch/Poster Session
    • 14:00 16:00
      Talks
      Convener: Ryan James Nichol (UCL)
      • 14:00
        Generative Unfolding 25m

        Looking for a way modern machine learning transforms LHC physics, unfolding has for a long time been one of our goal, and only modern networks allow us to do this meaningfully. It does not only make analyses with a wide range of theory hypotheses more efficient, it also allows the LHC collaborations to publish their data. I will show how generative networks can be used for this purpose, directly learning conditional probabilities in complete analogy to a forward simulations. I will then illustrate the power of this method using a challenging CMS case study.

        Speaker: Tilman Plehn (Heidelberg University)
      • 14:25
        lsbi: linear simulation based inference 25m

        Simulation-based inference is undergoing a renaissance in statistics and machine learning. With several packages implementing the state-of-the-art in expressive AI [mackelab/sbi] [undark-lab/swyft], it is now being effectively applied to a wide range of problems in the physical sciences, biology, and beyond.

        Given the rapid pace of AI/ML, there is little expectation that the implementations of the future will resemble these current first generation neural network-based approaches. This talk will present a new framework for simulation-based inference, linear simulation-based inference (lsbi), which abstracts the core principles of SBI from the specific details of machine learning, implementing a plug-and-play framework of linear and mixture models. lsbi has several use-cases:

        1. It is pedagogically helpful to separate out the general principles of SBI from the specific details of neural networks (particularly for ML skeptics).
        2. It is practically useful for producing expressive examples with known ground truths.
        3. It is pragmatically useful, since in many cases, lsbi is competitive with neural approaches in terms of accuracy, whilst being faster and more interpretable.

        An evolving code-driven PyPI/conda research package is available at: https://github.com/handley-lab/lsbi

        Speaker: Dr William Handley
      • 14:50
        Robust signal detection with classifiers decorrelated via optimal transport 25m

        New physics searches are usually done by training a supervised classifier to separate a signal model from the known Standard Model physics (also called the background model). However, even when the signal model is correct, systematic errors in the background model can influence supervised classifiers and might adversely affect the signal detection procedure. To tackle this problem, one approach is to use the (possibly misspecified) classifier only to perform a signal-enrichment step and then to carry out a model-agnostic search in the signal-rich region using only the real experimental data. For this procedure to work, we need a classifier constrained to be decorrelated with one or more protected variables used for the signal detection step. We do this by considering an optimal transport map of the classifier output that makes it independent of the protected variable(s) for the background. We then fit a semi-parametric mixture model to the distribution of the protected variable after making cuts on the transformed classifier to detect the presence of a signal. We compare and contrast this decorrelation method with previous approaches, show that the decorrelation procedure is robust to background misspecification, and analyse the power of the signal detection test.

        Speaker: Dr Purvasha Chakravarti (UCL)
      • 15:15
        The statistics of deep generative models for the LHC 45m

        In recent years, deep generative models (DGMs) have become essential for various steps in the LHC simulation and analysis chain. While there are many types of DGMs, no Swiss-army-knife architecture exists that can effectively handle speed, precision, and control simultaneously. In this talk, I will explore different DGMs, outline their strengths and weaknesses, and illustrate typical applications in high-energy physics. Moreover, I will introduce several methods to quantify and enhance model quality, including Bayesian neural networks for uncertainty estimation, the classifier test to identify failure modes, and reweighting using the DCTR approach. Lastly, to spark further discussion, I will discuss the *amplification methods for addressing the question, "How many more events can a DGM generate?"

        Speaker: Ramon Winterhalder (UCLouvain)
    • 16:00 16:30
      Break: Coffee
    • 16:30 18:05
      Talks
      Convener: Glen Cowan (Royal Holloway, University of London)
      • 16:30
        Fairness Methods in Particle Physics Event Classification 25m

        In social sciences, fairness in Machine Learning (ML) comprises the attempt to correct or eliminate algorithmic bias of gender, ethnicity, or sexual orientation from ML models. Many high-energy physics (HEP) analyses that search for a resonant decay of a particle employ mass-decorrelated event classifiers, as the particle mass is often used to perform the final signal extraction fit. These classifiers are designed to maintain fairness with respect to the mass, which is accomplished primarily by retaining mass-correlated information during training.

        Our studies present a first proof-of-concept for systematically applying, testing and comparing fairness methods for ML-based event classifiers in HEP analyses. We explore techniques that mitigate mass correlation during and after training. Through simulations and a case studies, we demonstrate the effectiveness of these methods in maintaining fairness while preserving the classifier performance.

        Speaker: Oliver Rieger (Nikhef National institute for subatomic physics (NL))
      • 16:55
        Systematics: Misery or Muse? 45m

        Systematic uncertainties usually have a negative connotation since they reduce the sensitivity of an experiment. However, the practical and conceptual challenges posed by various types of systematic uncertainty also have a long track record of motivating new ideas. I will outline some examples for my own career where systematics were my muse for innovation.

        Speaker: Kyle Stuart Cranmer (University of Wisconsin Madison (US))
      • 17:40
        Development of systematic-aware neural network trainings for binned-likelihood-analyses at the LHC 25m

        We demonstrate a neural network training, capable of accounting for the effects of systematic variations of the utilized data model in the training process and describe its extension towards neural network multiclass classification. We show the importance of adjusting backpropagation to be able to handle derivatives of histogram bins during training and add an interpretation of the optimization process itself, highlighting the differences between the systematic aware and conventional training strategies. Trainings for binary and multiclass classification with seven output classes are performed, based on a comprehensive data model with 86 nontrivial shape-altering systematic variations, as used for a previous measurement. The neural network output functions are used to infer the signal strengths for inclusive Higgs boson production, as well as for Higgs boson production via gluon-fusion (rggH) and vector boson fusion (rqqH). With respect to a conventional training, based on cross-entropy, we observe improvements of 12 and 16 %, for the sensitivity in rggH and rqqH, respectively.

        Speaker: Artur Monsch (KIT - Karlsruhe Institute of Technology (DE))
    • 19:00 21:00
      Social: Workshop dinner at Ognisko
    • 09:00 10:15
      Talks
      Convener: David A. van Dyk (Imperial College London)
      • 09:00
        Model misspecification meets ML: a HEP perspective 45m

        The field of high energy physics (HEP) benefits immensely from sophisticated simulators and data-driven techniques to perform measurements of nature at increasingly higher precision. Using the example of HEP, I will describe how and where uncertainties are incorporated into data analysis to address model misspecification concerns. My focus will be how machine learning (ML), in the variety of ways it is employed in practice, affects considerations around mis-modeling.

        Speaker: Alexander Held (University of Wisconsin Madison (US))
      • 09:45
        Learned harmonic mean estimation of the Bayesian evidence with normalizing flows 30m

        Computing the Bayesian evidence is an important task in Bayesian model selection, providing a principled quantitative way to compare models. In this work, we introduce normalizing flows to improve the learned harmonic mean estimator of the Bayesian evidence. This recently presented estimator leverages machine learning to address the exploding variance problem associated with the original harmonic mean. The improved method provides an accurate, robust and scalable estimator of the Bayesian evidence. Moreover, it is agnostic to the sampling strategy, meaning it can be combined with various efficient MCMC sampling techniques or variational inference approaches. We present numerical experiments demonstrating the effectiveness of the use of normalizing flows for the learned harmonic mean. We also apply the method to practical cosmological examples, including a 37-dimensional cosmic shear analysis using CosmoPower-JAX, a JAX-based implementation of the CosmoPower framework that accelerates cosmological inference by building differentiable neural emulators of cosmological power spectra, observing significant speed-up compared to the conventional method. We also successfully perform a 3x2pt analysis in a 157-dimensional setting, where using conventional methods is not feasible. This shows that the scalability CosmoPower-JAX and the learned harmonic mean estimator offer could allow for the comparison between models of unprecedented complexity, thus unlocking the full potential of Bayesian analysis even in high-dimensional settings.

        Speaker: Alicja Polanska (University College London)
    • 10:15 10:45
      Break: Coffee
    • 10:45 12:35
      Talks
      Convener: Jason McEwen
      • 10:45
        Interpretability in Semi-Supervised Classifier Tests for Model-Independent Searches of New Physics 30m

        Many model-independent search methods can be understood as performing a high-dimensional two-sample test. The test is typically performed by training a neural network over the high-dimensional feature space. If the test indicates a significant deviation from the background, it would be desirable to be able to characterize the "signal" the network may have found. In this talk, I will describe our work on interpreting semi-supervised classifier tests using active subspaces to understand the properties of the detected signal. Additionally, I will show how to extract the signal strength parameter from the trained classifier.

        Speaker: Mikael Kuusela (Carnegie Mellon University (US))
      • 11:15
        pop-cosmos: investigating the explainability of a high-dimensional, data-driven generative model in cosmology 30m

        I will present a perspective that explainability — model interrogation and validation rooted in domain knowledge — is a more important desideratum in fundamental science than interpretability in its strict meaning. In order to illustrate this point, I will draw on our recent work on pop-cosmos: a forward modelling framework for photometric galaxy survey data, where galaxies are modelled as draws from a population prior distribution over redshift, mass, dust properties, metallicity, and star formation history. After showing how the model is composed in terms of a diffusion model population prior and calibrated using simulation-based optimal-transport optimisation, I will discuss how to view this type of approach from a hierarchical Bayesian perspective. I will showcase the parallels between validating this type of model and standard practices in validating any complex physics-based parametric model in the field.

        Speaker: Hiranya Peiris
      • 11:45
        Identifying Tau Neutrinos in IceCube 25m

        Astrophysical tau neutrinos were predicted for a long time, but only recently has IceCube been able to identify those at the 5 sigma significance level. The key to this discovery was using machine learning methods to analyse the data. In this talk, I will first give a brief overview of the analysis and results before we dive deeper into the neural nets. We will try to understand how they work and discuss what surprising things we learned.

        Speaker: Philipp Eller (Wisconsin)
      • 12:10
        Conditional generation 25m

        “If you can simulate it, you can learn it.” The concept of conditional generation is powerful and versatile. The heavy lifting is distributed over a generator of a latent distribution of interest and an embedding network to encode the information contained in the data. Concrete applications to the reconstruction of neutrino kinematics in LHC collisions and associated interpretability questions will be presented.

        Speaker: Tobias Golling (Universite de Geneve (CH))
    • 12:35 14:00
      Break: Lunch
    • 14:00 15:30
      Talks
      Convener: Kaisey Mandel (University of Cambridge)
    • 15:30 16:00
      Break: Coffee
    • 16:00 17:45
      Talks
      Convener: Lydia Brenner (Nikhef National institute for subatomic physics (NL))