-
Gregor Kasieczka (Hamburg University (DE))08/09/2025, 09:15
-
Mikael Kuusela (Carnegie Mellon University (US))08/09/2025, 09:30Plenary
-
Genevieve Belanger, Genevieve Belanger08/09/2025, 10:00Plenary
-
LANCE Dixon08/09/2025, 10:30Plenary
-
Dr Wenshuai Wang (Institute of High Energy Physics)08/09/2025, 11:00
The High Energy cosmic-Radiation Detection (HERD) facility is a space astronomy and particle astrophysics experiment planned to be installed on the China Space Station. HERD is a China-led mission with Italy leading key European contributions. Its primary scientific goals include detecting dark matter in cosmic space, precisely measuring the energy spectrum and composition of cosmic rays, and...
Go to contribution page -
Akshay Malige (Brookhaven National Laboratory (US))08/09/2025, 11:00
Modern beam telescopes play a crucial role in high-energy physics experiments to precisely track particle interactions. Accurate alignment of detector elements in real-time is essential to maintain the integrity of reconstructed particle trajectories, especially in high-rate environments like the ATLAS experiment at the Large Hadron Collider (LHC). Any misalignment in the detector geometry can...
Go to contribution page -
Ianna Osborne (Princeton University)08/09/2025, 11:00
Awkward Array provides efficient handling of large, irregular data structures in Python, playing a key role in high-energy physics analysis. This work presents ongoing efforts to optimize Awkward Arrays for GPUs using CUDA, aiming to achieve performance parity with or surpass CPU kernel implementations. Key improvements focus on optimized memory management, leveraging CUDA-specific features,...
Go to contribution page -
Kylian Schmidt (KIT - Karlsruhe Institute of Technology (DE))08/09/2025, 11:00
The design of modern high-energy physics detectors is a highly intricate process, aiming to maximize their physics potential while balancing various manufacturing constraints. As detectors become larger and more sophisticated, it becomes increasingly difficult to maintain a comprehensive understanding of the entire system. To address this challenge, we aim to translate the design process into...
Go to contribution page -
Benjamin Galewsky (Univ. Illinois at Urbana Champaign (US))08/09/2025, 11:00
Data analysts working with large datasets require absolute certainty that each file is processed exactly once. ServiceX addresses this challenge by using well established transaction processing architectures. This system implements a fully transactional workflow powered by PostgreSQL and RabbitMQ, ensuring data integrity throughout the processing pipeline. This presentation details both the...
Go to contribution page -
CMS Collaboration, Thomas Reis (Science and Technology Facilities Council STFC (GB))08/09/2025, 11:00
Many physics analyses using the CMS detector at the LHC require accurate, high resolution electron and photon energy measurements. The CMS electromagnetic calorimeter (ECAL) is a fundamental component of these analyses. The excellent resolution of ECAL was of central importance to the discovery of the Higgs boson in 2012, and is being used for increasingly precise measurements of Higgs boson...
Go to contribution page -
Giacomo De Pietro (Karlsruhe Institute of Technology)08/09/2025, 11:00
Workflow Management Systems (WMSs) are essential tools for structuring an arbitrary sequence of tasks in a clear, maintainable, and repeatable way. The popular Python-based WMS luigi helps building complex workflows. It handles task dependency resolution and input/output tracking as well providing a simple workflow visualisation and a convenient command-line integration.
Go to contribution page
The extension... -
Valerio Pastore (INAF)08/09/2025, 11:00
At INAF (Istituto nazionale di astrofisica), in the contest of AGILE mission (Astro-Rivelatore Gamma a Immagini Leggero), we developed PacketLib, an open-source C++ software library designed for building applications that handle satellite telemetry source packets, provided they comply with the CCSDS Telemetry and Telecommand Standards.
Go to contribution page
As part of the ASTRI (Astrofisica con Specchi a... -
Alan Malta Rodrigues (University of Notre Dame (US)), CMS Collaboration08/09/2025, 11:00
The CMS Experiment had to manage and process data volumes approaching the exascale during the LHC Run 3. This required a seamless synergy between the workload and data management systems, namely WMCore and Rucio. Following up to the integration of Rucio into the CMS infrastructure, the workload management system has undergone substantial adaptations to harness new data management capabilities...
Go to contribution page -
Yao Zhang08/09/2025, 11:00
Charged particle tracking for drift chamber is a task in high-energy physics. In this work, we propose using reinforcement learning (RL) to the reconstruction of particle trajectories in drift chambers. By framing the tracking problem as a decision-making process, RL enables the development of more efficient and adaptive tracking algorithms. This approach offering improved performance and...
Go to contribution page -
Daniel Thomas Murnane (Niels Bohr Institute, University of Copenhagen)08/09/2025, 11:00
Particle physics is a field hungry for high quality simulation, to match the precision with which data is gathered at collider experiments such as the Large Hadron Collider (LHC). The computational demands of full detector simulation often lead to the use of faster but less realistic parameterizations, potentially compromising the sensitivity, generalizability, and robustness of downstream...
Go to contribution page -
Sergey Volkov08/09/2025, 11:00Track 3: Computations in Theoretical Physics: Techniques and MethodsPoster
QED in the classical Coulomb field of a nucleus serves as a good approximation for obtaining high-precision results in atomic physics. The external field is a subject to radiative corrections. These corrections can be explained perturbatively in terms of the QED Feynman diagrams containing scalar-like propagators for the external field together with the usual QED propagators.
A calculation...
Go to contribution page -
Dr Guang Zhao (Institute of High Energy Physics (CAS))08/09/2025, 11:00
Particle identification (PID) plays a crucial role in particle physics experiments. A groundbreaking advancement in PID involves cluster counting (dN/dx), which measures primary ionizations along a particle’s trajectory within a pixelated time projection chamber (TPC), as opposed to conventional dE/dx measurements. A pixelated TPC with a pixel size of 0.5x0.5 mm2 has been proposed as the...
Go to contribution page -
Minghua Liao (Sun Yat-Sen University (CN))08/09/2025, 11:00
Multiple visualization methods have been implemented in the Jiangmen Underground Neutrino Observatory (JUNO) experiment and its satellite experiment JUNO-TAO. These methods include event display software developed based on ROOT and Unity. The former is developed based on the JUNO offline software system and ROOT EVE, which provides an intuitive way for users to observe the detector geometry,...
Go to contribution page -
Mr Yujie Zeng (Sun Yat-Sen University (CN))08/09/2025, 11:00
Detector and event visualization software is essential for modern high-energy physics (HEP) experiment. It plays important role in the whole life circle of any HEP experiment, from detector design, simulation, reconstruction, detector construction and installation, to data quality monitoring, physics data analysis, education and outreach. In this talk, we will discuss two frameworks and their...
Go to contribution page -
Jamie Gooding (Technische Universitaet Dortmund (DE))08/09/2025, 11:00
Simulation-based inference (SBI) is a set of statistical inference approaches in which Machine Learning (ML) algorithms are trained to approximate likelihood ratios. It has been shown to provide an alternative to the likelihood fits commonly performed in HEP analyses. SBI is particularly attractive in analyses performed over many dimensions, in which binning data would be computationally...
Go to contribution page -
Liyan Qian (Chinese Academy of Sciences (CN))08/09/2025, 11:00
We present an end-to-end track reconstruction algorithm based on Graph Neural Networks (GNNs) for the main drift chamber of the BESIII experiment at the BEPCII collider. The algorithm directly processes detector hits as input to simultaneously predict the number of track candidates and their kinematic properties in each event. By incorporating physical constraints into the model, the...
Go to contribution page -
CMS Collaboration, Felix Philipp Zinn (Rheinisch Westfaelische Tech. Hoch. (DE))08/09/2025, 11:00
Statistical analyses in high energy physics often rely on likelihood functions of binned data. These likelihood functions can then be used for the calculation of test statistics in order to assess the statistical significance of a measurement.
evermore is a python package for building and evaluating these likelihood functions using JAX – a powerful python library for high performance...
Go to contribution page -
Doğa Elitez (CERN)08/09/2025, 11:00
A new all-silicon tracking detector (ITk) for ATLAS is under construction to meet the demands of the HL-LHC, and updated track reconstruction performance predictions have recently been published by the ATLAS Collaboration. The new detector, designed to operate at an average of up to 200 simultaneous proton–proton interactions, will provide a wider pseudorapidity coverage, an increased...
Go to contribution page -
Lars Stietz (Hamburg University of Technology (DE))08/09/2025, 11:00
The availability of precise and accurate simulation is a limiting factor for interpreting and forecasting data in many fields of science and engineering. Often, one or more distinct simulation software applications are developed, each with a relative advantage in accuracy or speed. The quality of insights extracted from the data stands to increase if the accuracy of faster, more economical...
Go to contribution page -
Mr Sami Ullah Khan (University of Padua & INFN Turin)08/09/2025, 11:00
The development of radiation-hard CMOS Monolithic Active Pixel Sensors (MAPS) is a key advancement for next-generation high-energy physics experiments. These sensors offer improved spatial resolution and integration capabilities but require efficient digital readout and data acquisition (DAQ) systems to operate in high-radiation environments. My research focuses on the FPGA-based digital...
Go to contribution page -
CMS Collaboration, Davide Valsecchi (ETH Zurich (CH))08/09/2025, 11:00
Precise simulation-to-data corrections, encapsulated in scale factors, are crucial for achieving high precision in physics measurements at the CMS experiment. Traditional methods often rely on binned approaches, which limit the exploitation of available information and require a time-consuming fitting process repeated for each bin. This work presents a novel approach utilizing modern...
Go to contribution page -
Sascha Diefenbacher (Lawrence Berkeley National Lab. (US))08/09/2025, 11:00
As generative models start taking an increasingly prominent role in both particle physics and everyday life, quantifying the statistical power and expressiveness of such generative models becomes a more and more pressing question.
In past work, we have seen that a generative mode can, in fact, be used to generate samples beyond the initial training data. However, the exact quantification...
Go to contribution page -
Jay Chan (Lawrence Berkeley National Lab. (US)), Xiangyang Ju (Lawrence Berkeley National Lab. (US))08/09/2025, 11:00
The simulation of particle interactions with detectors plays a critical role in understanding the detector performances and optimizing physics analysis. Without the guidance of the first-principle theory, the current state-of-the-art simulation tool, \textsc{Geant4}, exploits phenomenology-inspired parametric models, which must be combined and carefully tuned to experimental observations. The...
Go to contribution page -
Xiaoshuai Qin (Shandong University (CN))08/09/2025, 11:00
Track reconstruction is one of the most important and challenging tasks in the offline data processing of collider experiments. The Super Tau-Charm Facility (STCF) is a next-generation electron-positron collider running in the tau-charm energy region proposed in China, where conventional track reconstruction methods face enormous challenges from the higher background environment introduced by...
Go to contribution page -
Abdulrahman Al Marzouqi (University of Bahrain (BH)), CMS Collaboration08/09/2025, 11:00
The CMS experiment requires massive computational resources to efficiently process the large amounts of detector data. The computational demands are expected to grow as the LHC enters the high-luminosity era. Therefore, GPUs will play a crucial role in accelerating these workloads. The approach currently used in the CMS software (CMSSW) submits individual tasks to the GPU execution queues,...
Go to contribution page -
CMS Collaboration, Maria Michailidi (National and Kapodistrian University of Athens (GR))08/09/2025, 11:00
The CMS experiment requires massive computational resources to efficiently process the large amounts of detector data. The computational demands are expected to grow as the LHC enters the high-luminosity era. Therefore, GPUs will play a crucial role in accelerating these workloads. The approach currently used in the CMS software (CMSSW) relies on explicit memory management techniques, where...
Go to contribution page -
Patrick Asenov (Universita & INFN Pisa (IT))08/09/2025, 11:00
Precision measurements of particle properties, such as the leading hadronic contribution to the muon magnetic moment anomaly, offer critical tests of the Standard Model and probes for new physics. The MUonE experiment aims to achieve this through precise reconstruction of muon-electron elastic scattering events using silicon strip tracking stations and low-Z targets, while accounting for...
Go to contribution page -
hyuan hyuan (中国科学院高能物理研究所)08/09/2025, 11:00
Machine learning (ML), a cornerstone of data science and statistical analysis, autonomously constructs hierarchical mathematical models—such as deep neural networks—to extract complex patterns and relationships from data without explicit programming. This capability enables accurate predictions and the extraction of critical insights, making ML a transformative tool across scientific...
Go to contribution page -
Xiangyang Ju (Lawrence Berkeley National Lab. (US))08/09/2025, 11:00
Machine Learning (ML) plays an important role in physics analysis in High Energy Physics. To achieve better physics performance, physicists are training larger and larger models with larger dataset. Therefore, many workflow developments focus on distributed training of large ML models, inventing techniques like model pipeline parallelism. However, not all physics analyses need to train large...
Go to contribution page -
Merna Abumusabh (IPHC - Strasbourg)08/09/2025, 11:00
The calibration of Belle II data involves two key processes: prompt
Go to contribution page
calibration and reprocessing. Prompt calibration represents the initial
step in continuously deriving calibration constants in a timely manner for
the data collected over the previous couple of weeks. Currently, this process
is managed by b2cal, a Python-based plugin built on Apache
Airflow to handle calibration jobs.... -
Nikolay Gagunashvili08/09/2025, 11:00
Unfolding can be considered a procedure for estimating an unknown probability density function. Both external and internal quality assessment methods can be used for this purpose.
Go to contribution page
In some cases, external criteria exist that allow for the gauging of the quality of deconvolution. A typical example is the deconvolution of a blurred image, where the sharpness of the unblurred image can be used... -
Mr Andrea Paccagnella08/09/2025, 11:00
The LHCf experiment aims to study forward neutral particle production at the LHC, providing crucial data for improving hadronic interaction models used in cosmic ray physics. A key challenge in this context is the
Go to contribution page
measurement of $(K^0)$ production, indirectly reconstructed from the four photons originated by its decay. The main challenge in this measurement is the reconstruction of events... -
Muhammad Numan Anwar (Universita e INFN, Bari (IT))08/09/2025, 11:00
Cluster counting is a highly promising particle identification technique for drift chambers in particle physics experiments. In this paper, we trained neural network models, including a Long Short-Term Memory (LSTM) model for the peak-finding algorithm and a Convolutional Neural Network (CNN) model for the clusterization algorithm, using various hyperparameters such as loss functions,...
Go to contribution page -
Marie Hein (RWTH Aachen University)08/09/2025, 11:00
Weakly supervised anomaly detection has been shown to find new physics with a high significance at low injected signal cross sections. If the right features and a robust classifier architecture are chosen, these methods are sensitive to a very broad class of signal models. However, choosing the right features and classification architecture in a model-agnostic way is a difficult task as the...
Go to contribution page -
Andrzej Novak (Massachusetts Inst. of Technology (US)), Tristan Fillinger (KEK / IPNS)08/09/2025, 11:00
Visualizing pre-binned histograms is a HEP domain specific concern which is not adequately supported within the greater pythonic ecosystem. In recent years, [mplhep][1] has emerged as a leading package providing this basic functionality in a user-friendly interface. It also supplies styling templates for the four big LHC experiments - ATLAS, CMS, LHCb, and ALICE. At the same time, the...
Go to contribution page -
Jan Gavranovic (Jozef Stefan Institute (SI)), Jernej Debevc (Jozef Stefan Institute (SI)), Lara Calic (Lund University (SE))08/09/2025, 11:00
In a physics data analysis, "fake" or non-prompt backgrounds refer to events that would not typically satisfy the selection criteria for a given signal region, but are nonetheless accepted due to misreconstructed particles. This can occur, for example, when particles from secondary decays are incorrectly identified as originating from the hard scatter interaction point (resulting in non-prompt...
Go to contribution page -
Michael Boehler (University of Freiburg (DE))08/09/2025, 11:00
At many Worldwide LHC Computing Grid (WLCG) sites, HPC resources are already integrated, or will be integrated in the near future, into the experiment specific workflows. The integration can be done either in an opportunistic way to use otherwise unused resources for a limited period of time, or in a permanent way. The WLCG ATLAS Tier-2 cluster in Freiburg has been extended in both ways:...
Go to contribution page -
CMS Collaboration, Simone Rossi Tisbeni (Universita Di Bologna (IT))08/09/2025, 11:00
The High-Level Trigger (HLT) of the Compact Muon Solenoid (CMS) processes event data in real time, applying selection criteria to reduce the data rate from hundreds of kHz to around 5 kHz for raw data offline storage. Efficient lossless compression algorithms, such as LZMA and ZSTD, are essential in minimizing these storage requirements while maintaining easy access for subsequent analysis....
Go to contribution page -
Marc Neu08/09/2025, 11:00
Machine learning algorithms are being used more frequently in the {first-level triggers in collider experiments}, with Graph Neural Networks (GNNs) pushing the hardware requirements of FPGA-based triggers beyond the current state of the art. {As a first online event processing stage, first-level trigger systems process $O({10}~{\text{M}})$ events per second with a hard real-time latency...
Go to contribution page -
Maciej Mikolaj Glowacki (CERN)08/09/2025, 11:00
The CMS experiment at the LHC has entered a new phase in real-time data analysis with the deployment of two complementary unsupervised anomaly detection algorithms during Run 3 data-taking. Both algorithms aim to enhance the discovery potential for new physics by enabling model-independent event selection directly at the hardware trigger level, operating at the 40 MHz LHC collision rate within...
Go to contribution page -
Massimiliano Galli (Princeton University (US))08/09/2025, 11:00
The Analysis Grand Challenge (AGC) showcases an example of HEP analysis. Its reference implementation uses modern Python packages to realize the main steps, from data access to statistical model building and fitting. The packages used for data handling and processing (coffea, uproot, awkward-array) have recently undergone a series of performance optimizations.
Go to contribution page
While not being part of the HEP... -
Manfred Peter Fackeldey (Princeton University (US))08/09/2025, 11:00
In recent years, Awkward Array, Uproot, and related packages have become the go-to solutions for performing High-Energy Physics (HEP) analyses. Their development is driven by user experience and feedback, with the community actively shaping their evolution. User requests for new features and functionality play a pivotal role in guiding these projects.
For example, the Awkward development...
Go to contribution page -
Corentin Allaire (IJCLab, Université Paris-Saclay, CNRS/IN2P3)08/09/2025, 11:00
The simulation of calorimeter showers is computationally expensive, leading to the development of generative models as an alternative. Many of these models face challenges in balancing generation quality and speed. A key issue damaging the simulation quality is the inaccurate modeling of distribution tails. Normalizing flow (NF) models offer a trade-off between accuracy and speed, making them...
Go to contribution page -
Tomas Lindén (Helsinki Institute of Physics (FI))08/09/2025, 11:00
The HEP communities have developed an increasing interest in High Performance Computing (HPC) centres, as these hold the potential of providing significant computing resources to the current and future experiments. At the Large Hadron Collider (LHC), the ATLAS and CMS experiments are challenged with a scale up of several factors in computing for the High Luminosity LHC (HL-HLC) Run 4,...
Go to contribution page -
CMS Collaboration, Tomas Lindén (Helsinki Institute of Physics (FI))08/09/2025, 11:00
The HEP communities have developed an increasing interest in High Performance Computing (HPC) centres, as these hold the potential of providing significant computing resources to the current and future experiments. At the Large Hadron Collider (LHC), the ATLAS and CMS experiments are challenged with a scale up of several factors in computing for the High Luminosity LHC (HL-HLC) Run 4,...
Go to contribution page -
CMS Collaboration, David Walter (Massachusetts Inst. of Technology (US))08/09/2025, 11:00
The High Luminosity Large Hadron Collider (HL-LHC) and future big science experiments will generate unprecedented volumes of data, necessitating new approaches to physics analysis infrastructure. We present the SubMIT Physics Analysis Facility, an implementation of the emerging Analysis Facilities (AF) concept at MIT. Our solution combines high-throughput computing capabilities with modern...
Go to contribution page -
Mr Jiaheng Zou (IHEP, Beijing), Wenxing Fang08/09/2025, 11:00
The JUNO offline software (JUNOSW) is built upon the SNiPER framework. Its multithreaded extension, MT-SNiPER, enables inter-event parallel processing and has successfully facilitated JUNOSW's parallelization. Over the past year, two rounds of JUNO Data Challenge (DC) have been conducted to validate the complete data processing chain. During these DC tasks, the performance of MT-SNiPER was...
Go to contribution page -
Vito Conforti08/09/2025, 11:00
The ASTRI (Astrofisica con Specchi a Tecnologia Replicante Italiana) Project was born as a collaborative international effort led by the Italian National Institute for Astrophysics (INAF) to design and realize an end-to-end prototype of the Small-Sized Telescope (SST) of the Cherenkov Telescope Array (CTA) in a dual-mirror configuration (2M). The prototype, named ASTRI-Horn, has been...
Go to contribution page -
Dr Humberto Reyes-González (RWTH Aachen)08/09/2025, 11:00
Jet tagging, i. e. determining the origin of high-energy hadronic jets, is a key challenge in particle physics. Jets are ubiquitous observables in collider experiments, made of complex collections of particles, that need to be classified. Over the past decade, machine learning-based classifiers have greatly enhanced our jet tagging capabilities, with increasingly sophisticated models driving...
Go to contribution page -
Wenxing Fang08/09/2025, 11:00
The Jiangmen Underground Neutrino Observatory (JUNO) aims to determine the neutrino mass ordering (NMO) with a 3-sigma confidence level within six years. The experiment is currently in the commissioning phase, focusing on filling the liquid scintillator and evaluating detector performance. During physics data taking, the expected data rate after the global trigger is approximately 40 GB/s,...
Go to contribution page -
Marek Mateusz Mytkowski (Warsaw University of Technology (PL))08/09/2025, 11:00
Identifying products of ultrarelativistic collisions delivered by the LHC and RHIC colliders is one of the crucial objectives of experiments such as ALICE and STAR, which are specifically designed for this task. They allow for a precise Particle Identification (PID) over a broad momentum range.
Traditionally, PID methods rely on hand-crafted selections, which compare the recorded signal of...
Go to contribution page -
Marcin Wolter (Polish Academy of Sciences (PL))08/09/2025, 11:00
Patrick Asenov (Universita & INFN Pisa (IT)), Anna Driutti (Universita & INFN Pisa (IT)), Mateusz Jacek Goncerz (Polish Academy of Sciences (PL)), Emma Hess (Universita & INFN Pisa (IT)), Marcin Kucharczyk (Polish Academy of Sciences (PL)), Damian Mizera (Cracow University of Technology (PL)), Marcin Wolter (Polish Academy of Sciences (PL)), Milosz Zdybal (Polish Academy of Sciences...
Go to contribution page -
Benjamin Huth (CERN)08/09/2025, 11:00
Graph Neural Networks (GNNs) have been in the focus of machine-learning-based track reconstruction for high-energy physics experiments during the last years. Within ATLAS, the GNN4ITk group has investigated this type of algorithm for track reconstruction at the High-Luminosity LHC (HL-LHC) using the future full-silicon Inner Tracker (ITk).
The Event Filter (EF) is part of the ATLAS Trigger...
Go to contribution page -
Ayse Asu Guvenli (Hamburg University (DE)), CMS Collaboration08/09/2025, 11:00
Searches for long-lived particles (LLPs) have attracted much interest lately due to their high discovery potential in the LHC Run-3. Signatures featuring LLPs with long lifetimes and decaying inside the muon detectors of the CMS experiment at CERN are of particular interest. In this talk, we will describe a novel Level-1 trigger algorithm that significantly improves CMS's signal efficiency for...
Go to contribution page -
Silia Taider08/09/2025, 11:00
The ROOT software package is a widely used data analysis framework in the High Energy Physics (HEP) community. As many other Python packages, ROOT features a powerful performance-oriented core which can be accessed in a Python application thanks to dynamic bindings. These facilitate the usage and integration of ROOT with the broader Python ecosystem.
Go to contribution page
Despite these capabilities, there is... -
Paul Gilles (RWTH Aachen University)08/09/2025, 11:00
With the consequences of global warming becoming abundantly clear, physics research needs to do its part in becoming more sustainable, including its computing aspects. Many measures in this field expend great effort to keep the impact on users minimal. However, even greater savings can be gained when compromising on these expectations.
Go to contribution page
In any such approach affecting the user experience, the... -
Jonas Eppelt (Karlsruher Insititute of Technology (KIT))08/09/2025, 11:00
The Belle~II electromagnetic calorimeter (ECL) is not only used for measuring electromagnetic particles but also for identifying and determining the position of hadrons, particularly neutral hadrons.
Go to contribution page
Recent data-taking periods have presented two challenges for the current clustering method:
Firstly, the record-breaking luminosities achieved by the SuperKEKB accelerator have increased... -
Cristina Botta (CERN)08/09/2025, 11:30Plenary
-
Sascha Caron (Nikhef National institute for subatomic physics (NL))08/09/2025, 12:00Plenary
-
Shih-Chieh Hsu (University of Washington Seattle (US))08/09/2025, 12:30Plenary
-
Luca Beccatini08/09/2025, 14:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Accurate and efficient predictions of scattering amplitudes are essential for precision studies in high-energy physics, particularly for multi-jet processes at collider experiments. In this work, we introduce a novel neural network architecture designed to predict amplitudes for multi-jet events. The model leverages the Catani–Seymour factorization scheme and uses MadGraph to compute...
Go to contribution page -
Jay Chan (Lawrence Berkeley National Lab. (US))08/09/2025, 14:30
Track reconstruction is a cornerstone of modern collider experiments, and the HL-LHC ITk upgrade for ATLAS poses new challenges with its increased silicon hit clusters and strict throughput requirements. Deep learning approaches compare favorably with traditional combinatorial ones — as shown by the GNN4ITk project, a geometric learning tracking pipeline that achieves competitive physics...
Go to contribution page -
Maciej Mikolaj Glowacki (CERN)08/09/2025, 14:30Track 1: Computing Technology for Physics ResearchOral
We present an MLOps approach for managing the end-to-end lifecycle of machine learning algorithms deployed on FPGAs in the CMS Level-1 Trigger (L1T). The primary goal of the pipeline is to respond to evolving detector conditions by automatically acquiring up-to-date training data, retraining and re-optimising the model, validating performance, synthesising firmware, and deploying validated...
Go to contribution page -
Yang Zhang (Institute of High Energy Physics, Chinese Academy of Science)08/09/2025, 14:50
Precision measurements of Higgs, W, and Z bosons at future lepton colliders demand jet energy reconstruction with unprecedented accuracy. The particle flow approach has proven to be an effective method for achieving the required jet energy resolution. We present CyberPFA, a particle flow algorithm specifically optimized for the particle-flow-oriented crystal bar electromagnetic calorimeter...
Go to contribution page -
Javier Mariño Villadamigo (Institut für Theoretische Physik - University of Heidelberg)08/09/2025, 14:50Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Direct simulation of multi-parton QCD processes at full-color accuracy is computationally expensive, making it often impractical for large-scale LHC studies. A two-step approach has recently been proposed to address this: events are first generated using a fast leading-color approximation and reweighted to full-color accuracy. We build upon this strategy by introducing a machine-learning...
Go to contribution page -
Greta Sophie Heine08/09/2025, 14:50Track 1: Computing Technology for Physics ResearchOral
The high-luminosity environment at Belle II leads to growing beam-induced background, posing major challenges for the Belle II Level-1 (L1) trigger system. To maintain trigger rates within hardware constraints, effective background suppression is essential. Hit filtering algorithms based on Graph Neural Networks (GNNs), including the Interaction Network (IN), have demonstrated successful...
Go to contribution page -
Cilicia Uzziel Perez (La Salle, Ramon Llull University (ES))08/09/2025, 15:10
We present lightweight, attention-enhanced Graph Neural Networks (GNNs) tailored for real-time particle reconstruction and identification in LHCb’s next-generation calorimeter. Our architecture builds on node-centric GarNet layers, which eliminate costly edge message passing and are optimized for FPGA deployment, achieving sub-microsecond inference latency. By integrating attention mechanisms...
Go to contribution page -
Lauri Antti Olavi Laatu (Imperial College (GB))08/09/2025, 15:10Track 1: Computing Technology for Physics ResearchOral
Transformers are the state-of-the-art model architectures and widely used in application areas of machine learning. However the performance of such architectures is less well explored in the ultra-low latency domains where deployment on FPGAs or ASICs is required. Such domains include the trigger and data acquisition systems of the LHC experiments.
We present a transformer-based algorithm...
Go to contribution page -
Remco De Boer08/09/2025, 15:10Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
One of the central tools in hadron spectroscopy is amplitude analysis (partial-wave analysis) to interpret the experimental data. Amplitude models are fitted to data with large statistics to extract information about resonances and branching fractions. In amplitude analysis, we require flexibility to implement models with different decay hypotheses, spin formalisms, and resonance...
Go to contribution page -
CMS Collaboration, Efe Yigitbasi (Rice University (US))08/09/2025, 15:30Track 1: Computing Technology for Physics ResearchOral
The CMS Experiment at the CERN Large Hadron Collider (LHC) relies on a Level-1 Trigger system (L1T) to process in real time all potential collisions, happeing at a rate of 40 MHz, and select the most promising ones for data acquisition and further processing. The CMS upgrades for the upcoming high-luminosity LHC run will vastly improve the quality of the L1T event reconstruction, providing...
Go to contribution page -
Katharina sophia Schaeuble, Ulrich Einhaus (KIT - Karlsruhe Institute of Technology (DE))08/09/2025, 15:30
We present a versatile GNN-based end-to-end reconstruction algorithm for highly granular calorimeters that can include track and timing information to aid the reconstruction of particles. The algorithm starts directly from calorimeter hits and possibly reconstructed tracks, and outputs a coordinate transformation in which all shower objects are well separated from each other and assigned...
Go to contribution page -
Mathijs Fraaije08/09/2025, 15:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
I will present work on Tropical sampling from Feynman measures:
We introduce an algorithm that samples a set of loop momenta distributed as a given Feynman integrand. The algorithm uses the tropical sampling method and can be applied to evaluate phase-space-type integrals efficiently. We provide an implementation,
Go to contribution pagemomtrop, and apply it to a series of relevant integrals from the... -
Enrico Lupi (CERN, INFN Padova (IT))08/09/2025, 15:50
With the upcoming High-Luminosity upgrades at the LHC, data generation rates are expected to increase significantly. This calls for highly efficient architectures for machine learning inference in experimental workflows like event reconstruction, simulation, and data analysis.
Go to contribution page
At the ML4EP team at CERN, we have developed SOFIE, a tool within the ROOT/TMVA package that translates externally... -
Andreas Stefl (CERN)08/09/2025, 16:40Track 1: Computing Technology for Physics ResearchOral
The ATLAS experiment has engaged in a modernization of the reconstruction software to cope with the challenging running conditions expected for HL-LHC operations. The use of the experiment-independent ACTS toolkit for track reconstruction is a major component of this effort, involving the complete redesign of several elements of the ATLAS reconstruction software. This contribution will...
Go to contribution page -
Aishik Ghosh (University of California Irvine (US))08/09/2025, 16:40
Particle physics experiments rely on the (generalised) likelihood ratio test (LRT) for searches and measurements. This is not guaranteed to be optimal for composite hypothesis tests, as the Neyman-Pearson lemma pertains only to simple hypothesis tests. An improvement in the core statistical testing methodology would have widespread ramifications across experiments. We discuss an alternate test...
Go to contribution page -
Joshua Davies (University of Liverpool)08/09/2025, 16:40Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
For several decades, the FORM computer algebra system has been a crucial software package for the large-scale symbolic manipulations required by computations in theoretical high-energy physics. In this talk I will present version 5, which includes an updated built-in diagram generator, greatly improved polynomial arithmetic performance through an interface to FLINT, and enhanced capabilities...
Go to contribution page -
Simone Balducci (Universita e INFN, Bologna (IT))08/09/2025, 17:00Track 1: Computing Technology for Physics ResearchOral
CLUEstering is a versatile clustering library based on CLUE, a density-based weighted clustering algorithm optimized for high-performance computing that supports clustering in an arbitraty. The library offers a user-friendly Python interface and a C++ backend to maximize performance. CLUE’s parallel design is tailored to exploit modern hardware accelerators, enabling it to process large-scale...
Go to contribution page -
Daniel Stremmer (KIT)08/09/2025, 17:00Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
We present the program package ${\tt ggxy}$, which in its first version can be used to calculate partonic and hadronic cross sections to Higgs boson pair production at NLO QCD. The 2-loop virtual amplitudes are implemented using analytical approximations in different kinematic regions, while all other parts of the calculation are exact. This implementation allows to freely modify the masses of...
Go to contribution page -
R D Schaffer (Université Paris-Saclay (FR))08/09/2025, 17:00
Neural Simulation-Based Inference (NSBI) is a powerful class of machine learning (ML)-based methods for statistical inference that naturally handle high dimensional parameter estimation without the need to bin data into low-dimensional summary histograms. Such methods are promising for a range of measurements at the Large Hadron Collider, where no single observable may be optimal to scan over...
Go to contribution page -
Marian I Ivanov (GSI - Helmholtzzentrum fur Schwerionenforschung GmbH (DE))08/09/2025, 17:20
We present a modular, data-driven framework for calibration and performance correction in the ALICE experiment. The method addresses time- and parameter-dependent effects in high-occupancy heavy-ion environments, where evolving detector conditions (e.g., occupancy and cluster overlaps, gain drift, space charge, dynamic distortions, and reconstruction or calibration deficiencies) require...
Go to contribution page -
Enrico Bothmann (CERN)08/09/2025, 17:20Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Significant computing resources are used for parton-level event generation for the Large Hadron Collider (LHC). The resource requirements of this part of the simulation toolchain are expected to grow further in the High-Luminosity (HL-LHC) era. At the same time, the rapid deployment of computing hardware different from the traditional CPU+RAM model in data centers around the world mandates a...
Go to contribution page -
Mateusz Jakub Fila (CERN)08/09/2025, 17:20Track 1: Computing Technology for Physics ResearchOral
The Julia programming language is considered a strong contender as a future language for high-energy physics (HEP) computing. However, transitioning to the Julia ecosystem will be a long process and interoperability between Julia and C++ is required. So far several successful attempts have been made to wrap HEP C++ packages for use in Julia. It is also important to explore the reverse...
Go to contribution page -
Hongyue Duyang (Shandong University)08/09/2025, 17:40
Jiangmen Underground Neutrino Observatory (JUNO) is a next generation 20-kton liquid scintillator detector under construction in southern China. It is designed to determine neutrino mass ordering via the measurement of reactor neutrino oscillation, and also to study other physics topics including atmospheric neutrinos, supernova neutrinos and more. The detector's large mass and high...
Go to contribution page -
Hideki Okawa (Chinese Academy of Sciences (CN))08/09/2025, 17:40Track 1: Computing Technology for Physics ResearchOral
Future colliders such as the High Luminosity Large Hadron Collider and Circular Electron Positron Collider will face enormous increase in dataset in the coming decades. Quantum and quantum-inspired algorithms may allow us to overcome some of such challenges. There is an important class of problems the so-called combinatorial optimization problems. They are non-deterministic polynomial time...
Go to contribution page -
Gregor Kasieczka (Hamburg University (DE))09/09/2025, 09:45
-
Tilman Plehn, Tilman Plehn09/09/2025, 10:00Plenary
-
Dr Michael Hudson Kirby (Brookhaven National Laboratory (US)), Michael Hudson Kirby (Brookhaven National Laboratory (US))09/09/2025, 10:30Plenary
-
Yuan-Tang Chou (University of Washington (US))09/09/2025, 11:30Plenary
-
Daniel Ratner (SLAC)09/09/2025, 12:00Plenary
-
Dimitrios Danopoulos (CERN)09/09/2025, 14:30Track 1: Computing Technology for Physics ResearchOral
The increasing reliance on deep learning for high-energy physics applications demands efficient FPGA-based implementations. However, deploying complex neural networks on FPGAs is often constrained by limited hardware resources and prolonged synthesis times. Conventional monolithic implementations suffer from scalability bottlenecks, necessitating the adoption of modular and resource-aware...
Go to contribution page -
Laurits Tani (National Institute of Chemical Physics and Biophysics (EE))09/09/2025, 14:30
The application of foundation models in high-energy physics has recently been proposed as a way to use large unlabeled datasets to efficiently train powerful task-specific models. The aim is to train a task-agnostic model on an existing large dataset such that the learned representation can later be utilized for subsequent downstream physics tasks.
Go to contribution page
The pretrained model can reduce the training... -
Bryan Bliewert (Deutsches Elektronen-Synchrotron (DE))09/09/2025, 14:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Physics programs at future colliders cover a wide range of diverse topics and set high demands for precise event reconstruction. Recent analyses have stressed the importance of accurate jet clustering in events with low boost and high jet multiplicity. This contribution present how machine learning can be applied to jet clustering while taking desired properties such as infrared and collinear...
Go to contribution page -
Roope Oskari Niemi09/09/2025, 14:50Track 1: Computing Technology for Physics ResearchOral
Machine learning model compression techniques—such as pruning and quantization—are becoming increasingly important to optimize model execution, especially for resource-constrained devices. However, these techniques are developed independently of each other, and while there exist libraries that aim to unify these methods under a single interface, none of them offer integration with hardware...
Go to contribution page -
Anna Hallin (University of Hamburg)09/09/2025, 14:50
OmniJet-alpha, the first cross-task foundation model for particle physics, was first presented at ACAT 2024. In its base configuration, OmniJet-alpha is capable of transfer learning between an unsupervised problem (jet generation) and a classic supervised task (jet tagging). Since its release, we have also shown that it can sucessfully transfer from CMS Open data to simulation, and even...
Go to contribution page -
Aishik Ghosh (University of California Irvine (US))09/09/2025, 14:50Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
The process of neutrino model building using flavor symmetries requires a physicist to select a group, determine field content, assign representations, construct the Lagrangian, calculate the mass matrices matrix, and perform statistical fits of the resulting free parameters. This process is constrained by the physicist's time and their intuition regarding mathematically complex groups,...
Go to contribution page -
Lea Reuter (Karlsruhe Institute of Technology)09/09/2025, 15:10
Large backgrounds and detector aging impact the track finding in the Belle II central drift chamber, reducing both purity and efficiency in events. This necessitates the development of new track algorithms to mitigate detector performance degradation. Building on our previous success with an end-to-end multi-track reconstruction algorithm for the Belle II experiment at the SuperKEKB collider...
Go to contribution page -
Joaquin Iturriza Ramirez (Centre National de la Recherche Scientifique (FR))09/09/2025, 15:10Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Fast and precise evaluations of scattering amplitudes even
Go to contribution page
in the case of precision calculations is essential for event generation
tools at the HL-LHC. We explore the scaling behavior of the achievable
precision of neural networks in this regression problem for multiple
architectures, including a Lorentz symmetry aware multilayer perceptron
and the L-GATr architecture. L-GATr is... -
Federico Lazzari (Universita di Pisa & INFN Pisa (IT))09/09/2025, 15:10Track 1: Computing Technology for Physics ResearchOral
With plans for upgrading the detector in order to collect data at a luminosity up to 1.5×1034 cm-2s-2 being ironed out (Upgrade II - LHC Run5), the LHCb Collaboration has sought to implement new data taking solutions already starting from the upcoming LHC Run4 (2030-2033).
The first stage of the LHCb High Level Trigger (HLT1), currently implemented on GPUs and aiming at reducing the event...
Go to contribution page -
CMS Collaboration, Filippo Cattafesta (Scuola Normale Superiore & INFN Pisa (IT))09/09/2025, 15:30
Detailed event simulation at the LHC is taking a large fraction of computing budget. CMS developed an end-to-end ML based simulation that can speed up the time for production of analysis samples of several orders of magnitude with a limited loss of accuracy. As the CMS experiment is adopting a common analysis level format, the NANOAOD, for a larger number of analyses, such an event...
Go to contribution page -
Liv Helen Vage (Princeton University (US))09/09/2025, 15:30Track 1: Computing Technology for Physics ResearchOral
Fast machine learning (ML) inference is of great interest in the HEP community, especially in low-latency environments like triggering. Faster inference often unlocks the use of more complex ML models that improve physics performance, while also enhancing productivity and sustainability. Logic gate networks (LGNs) currently achieve some of the fastest inference times for standard image...
Go to contribution page -
Sophia Vent09/09/2025, 15:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Modern ML-based taggers have become the gold standard at the LHC, outperforming classical algorithms. Beyond pure efficiency, we also seek controllable and interpretable algorithms. We explore how we can move beyond black-box performance and toward physically meaningful understanding of modern taggers. Using explainable AI methods, we can connect tagger outputs with well-known physics...
Go to contribution page -
CMS Collaboration, Dr Florian Bury (University of Bristol)09/09/2025, 15:50
The Matrix Element Method (MEM) offers optimal statistical power for hypothesis testing in particle physics, but its application is hindered by the computationally intensive multi-dimensional integrals required to model detector effects. We present a novel approach that addresses this challenge by employing Transformers and generative machine learning (ML) models. Specifically, we utilize ML...
Go to contribution page -
Yuan-Tang Chou (University of Washington (US))09/09/2025, 15:50Track 1: Computing Technology for Physics ResearchOral
The rising computational demands of growing data rates and complex machine learning (ML) algorithms in large-scale scientific experiments have driven the adoption of the Services for Optimized Network Inference on Coprocessors (SONIC) framework. SONIC accelerates ML inference by offloading tasks to local or remote coprocessors, optimizing resource utilization. Its portability across diverse...
Go to contribution page -
David Lawrence09/09/2025, 16:40Track 1: Computing Technology for Physics ResearchOral
Stable operation of detectors, beams, and targets is crucial for reducing systematic errors and achieving high-precision measurements in accelerator-based experiments. Historically, this stability was achieved through extensive post-acquisition calibration and systematic data studies, as not all operational parameters could be precisely controlled in real time. However, recent advances in...
Go to contribution page -
Quentin Führing (Technische Universitaet Dortmund (DE), University of Cambridge (UK))09/09/2025, 16:40
Measurements of neutral, oscillating mesons are a gateway to quantum mechanics and give access to the fundamental interactions of elementary particles. For example, precise measurements of $CP$ violation in neutral $B$ mesons can be taken in order to test the Standard Model of particle physics. These measurements require knowledge of the $B$-meson flavour at the time of its production, which...
Go to contribution page -
Tianji Cai09/09/2025, 16:40Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
AI for fundamental physics is now a burgeoning field, with numerous efforts pushing the boundaries of experimental and theoretical physics, as well as machine learning research itself. In this talk, I will introduce a recent innovative application of Natural Language Processing to the state-of-the-art precision calculations in high energy particle physics. Specifically, we use Transformers to...
Go to contribution page -
Lorenz Vogel (ITP, Heidelberg University)09/09/2025, 17:00Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
ATLAS explores modern neural networks for a multi-dimensional calibration of its calorimeter signal defined by clusters of topologically connected cells (topo-clusters). The Bayesian neural network (BNN) approach yields a continuous and smooth calibration function, including uncertainties on the calibrated energy per topo-cluster. In this talk the performance of this BNN-derived calibration is...
Go to contribution page -
Sebastian Pitz (ITP, Heidelberg University)09/09/2025, 17:00
We construct Lorentz-equivariant transformer and graph networks using the concept of local canonicalization. While many Lorentz-equivariant architectures use specialized layers, this approach allows to take any existing non-equivariant architecture and make it Lorentz-equivariant using transformations with equivariantly predicted local frames. In addition, data augmentation emerges as a...
Go to contribution page -
Anwar Ibrahim, Prof. fedor. ratnikov09/09/2025, 17:00Track 1: Computing Technology for Physics ResearchOral
Abstract:
Go to contribution page
Optimizing control systems in particle accelerators presents significant challenges, often requiring extensive manual effort and expert knowledge. Traditional tuning methods are time-consuming and may struggle to navigate the complexity of modern beamline architectures. To address these challenges, we introduce a simulation-based framework that leverages Reinforcement Learning (RL)... -
Gage DeZoort (Princeton University (US))09/09/2025, 17:20
Modern machine learning (ML) algorithms are sensitive to the specification of non-trainable parameters called hyperparameters (e.g., learning rate or weight decay). Without guiding principles, hyperparameter optimization is the computationally expensive process of sweeping over various model sizes and, at each, re-training the model over a grid of hyperparameter settings. However, recent...
Go to contribution page -
Nina Elmer (Heidelberg University)09/09/2025, 17:20Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Neural networks for LHC physics must be accurate, reliable, and well-controlled. This requires them to provide both precise predictions and reliable quantification of uncertainties - including those arising from the network itself or the training data. Bayesian networks or (repulsive) ensembles provide frameworks that enable learning systematic and statistical uncertainties. We investigate...
Go to contribution page -
Gustavo Uribe (Universidad Antonio Narino (CO))09/09/2025, 17:20Track 1: Computing Technology for Physics ResearchOral
The ATLAS detector at CERN and its supporting infrastructure form a highly complex system. It covers numerous interdependent sub-systems and requires collaboration across a team of multi-disciplinary experts. The ATLAS Technical Coordination Expert System provides an interactive description of the technical infrastructure and enhances its understanding. It features tools to assess the impact...
Go to contribution page -
Jose Miguel Munoz Arias (MIT)09/09/2025, 17:40Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
One of the main goals of theoretical nuclear physics is to provide a first-principles description of the atomic nucleus, starting from interactions between nucleons (protons and neutrons). Although exciting progress has been made in recent years thanks to the development of many-body methods and nucleon-nucleon interactions derived from chiral effective field theory, performing accurate...
Go to contribution page -
Samuele Grossi (Università degli studi di Genova & INFN sezione di Genova)09/09/2025, 17:40
Deep generative models have become powerful tools for alleviating the computational burden of traditional Monte Carlo generators in producing high-dimensional synthetic data. However, validating these models remains challenging, especially in scientific domains requiring high precision, such as particle physics. Two-sample hypothesis testing offers a principled framework to address this task....
Go to contribution page -
Gordon Watts (University of Washington (US))09/09/2025, 17:40Track 1: Computing Technology for Physics ResearchOral
In high energy physics, most AI/ML efforts focus on improving the scientific process itself — modeling, classification, reconstruction, and simulation. In contrast, we explore how Large Language Models (LLMs) can accelerate access to the physics by assisting with the broader ecosystem of work that surrounds and enables scientific discovery. This includes understanding complex documentation,...
Go to contribution page -
Gregor Kasieczka (Hamburg University (DE))10/09/2025, 09:15
-
Simone Zoia (CERN)10/09/2025, 09:30Plenary
-
Dr Andrea Bocci (CERN)10/09/2025, 10:00Plenary
-
Sascha Diefenbacher (Lawrence Berkeley National Lab. (US))10/09/2025, 10:30Plenary
-
Peter McKeown (CERN)10/09/2025, 11:00
Simulation plays an essential role in modern high energy physics experiments. However, the simulation of particle showers in the calorimeter systems of detectors with traditional Monte Carlo procedures represents a major computational bottleneck, and this subdetector system has long been the focus of fast simulation efforts. More recently, approaches based on deep generative models have shown...
Go to contribution page -
Maksim Melnik Storetvedt (Western Norway University of Applied Sciences (NO))10/09/2025, 11:00
With the emergence of increasingly complex workflows and data rates, accelerators have gained importance within ALICE and the Worldwide LHC Computing Grid (WLCG). Consequently, support for GPUs was added to JAliEn, the ALICE Grid middleware, in a transparent manner to automatically use these resources when available -- without breaking existing mechanisms for payload isolation and...
Go to contribution page -
ASRITH KRISHNA RADHAKRISHNAN (Universita e INFN, Bologna (IT))10/09/2025, 11:00
Unsupervised anomaly detection has become a pivotal technique for model-independent searches for new physics at the LHC. In high-energy physics (HEP), anomaly detection is employed to identify rare, outlier events in collision data that deviate significantly from expected distributions. A promising approach is the application of generative machine learning models, which can efficiently detect...
Go to contribution page -
Nick Smith (Fermi National Accelerator Lab. (US))10/09/2025, 11:00
With the move to HTTP/WebDAV and JSON Web Tokens as a standard protocol for transfers within the WLCG distributed storage network, a large amount of off-the-shelf technologies become viable for meeting the requirements of a Storage Element (SE). In this work, we explore the capabilities and performance of the OpenResty framework, which extends the nginx server with the LuaJIT scripting...
Go to contribution page -
Raghuvar Vijayakumar (University of Freiburg (DE))10/09/2025, 11:00
In the field of High Throughput Computing (HTC), the management and processing of large volumes of accounting data across different environments and use cases is a significant challenge. AUDITOR addresses this issue by providing a flexible framework for building accounting pipelines that can be adapted to a wide range of needs.
Go to contribution page
At its core, AUDITOR serves as a centralised storage solution for... -
144. Building a High-Availability and User-Friendly Neutron Scattering Data Computing InfrastructureMr Binbin Lee (Institute of High Energy Physics)10/09/2025, 11:00
To address the urgent need for efficient data analysis platforms in the neutron scattering field, this report presents a cloud-based computing infrastructure solution based on the technical architecture of OpenStack and WebRTC. Based on this infrastructure, a deeply integrated system for data management and storage is constructed to provide researchers with a one-stop analysis platform that...
Go to contribution page -
Valerii Kholoimov (Instituto de Física Corpuscular (Univ. of Valencia))10/09/2025, 11:00
The new fully software-based trigger of the LHCb experiment operates at a 30 MHz data rate, opening a search window into previously unexplored regions of physics phase space. The BuSca (Buffer Scanner) project at LHCb acquires, reconstructs and analyzes data in real time, extending sensitivity to new lifetimes and mass ranges though the recently deployed Downstream tracking algorithm. BuSca...
Go to contribution page -
Henry Day-Hall (DESY)10/09/2025, 11:00
This contribution presents the final iteration of the CaloClouds series. Simulation of photon showers in the granularities expected in a future Higgs factory is computationally challenging. A viable simulation must capture the find details exposed by such a detector, while also being fast enough to keep pace with the expected rate of observations. The Caloclouds model utilises point cloud...
Go to contribution page -
Zhipeng Yao10/09/2025, 11:00
The Super Tau Charm Facility (STCF) is a next-generation electron-positron collider proposed in China, operating at a center-of-mass energy of 2–7 GeV with a peak luminosity of 0.5×10³⁵ cm⁻²s⁻¹. In STCF experiments, the identification of high-momentum charged hadrons is critical for physics studies, driving the implementation of a dedicated particle identification (PID) system that combines...
Go to contribution page -
Tao Lin (Chinese Academy of Sciences (CN))10/09/2025, 11:00
The Jiangmen Underground Neutrino Observatory (JUNO) is a multipurpose neutrino experiment with the primary goals of the determining the neutrino mass ordering and precisely measuring oscillation parameters. The JUNO detector construction was completed at the end of 2024. It generate about 3 petabytes of data annually, requiring extensive offline processing. This processing, which is called...
Go to contribution page -
Zahraa Zaher10/09/2025, 11:00
Recent years have seen growing interest in leveraging secondary cosmic ray muons for tomographic imaging of large and unknown volumes. A key area of application is cargo scanning for border security, where muon tomography is used to detect concealed hazardous or illicit materials in trucks and shipping containers. We present recent developments in TomOpt, a Python-based, end-to-end software...
Go to contribution page -
CMS Collaboration, Christine Zeh (Vienna University of Technology (AT))10/09/2025, 11:00
Efficient data processing using machine learning relies on heterogeneous computing approaches, but optimizing input and output data movements remains a challenge. In GPU-based workflows data already resides on GPU memory, but machine learning models requires the input and output data to be provided in specific tensor format, often requiring unnecessary copying outside of the GPU device and...
Go to contribution page -
Yuan-Tang Chou (University of Washington (US))10/09/2025, 11:00
Charge particle track reconstruction is the foundation of the collider experiments. Yet, it's also the most computationally expensive part of the particle reconstruction. The innovation in tracking reconstruction with graph neural networks (GNNs) has shown the promising capability to cope with the computing challenges posed by the High-Luminosity LHC (HL-LHC) with Machine learning. However,...
Go to contribution page -
Alina Lazar (Youngstown State University (US))10/09/2025, 11:00
The TrackML dataset, a benchmark for particle tracking algorithms in High-Energy Physics (HEP), presents challenges in data handling due to its large size and complex structure. In this study, we explore using a heterogeneous graph structure combined with the Hierarchical Data Format version 5 (HDF5) not only to efficiently store and retrieve TrackML data but also to speed up the training and...
Go to contribution page -
Vivekanand Gyanchand Sahu (University of California San Diego)10/09/2025, 11:00
We present a suite of optimizations to the Particle Transformer (ParT), a state-of-the-art model for jet tagging, targeting the stringent latency and memory constraints of real-time environments such as HL-LHC triggers. To address the quadratic scaling and compute bottlenecks of standard attention, we integrate FlashAttention for exact, fused-kernel attention with reduced memory I/O, and...
Go to contribution page -
Lukas Lang (RWTH Aachen University)10/09/2025, 11:00
Weakly supervised anomaly detection has been shown to be a sensitive and robust tool for Large Hadron Collider (LHC) analysis. The effectiveness of these methods relies heavily on the input features of the classifier, influencing both model coverage and the detection of low signal cross sections. In this talk, we demonstrate that improvements in both areas can be achieved by using energy flow...
Go to contribution page -
Raees Ahmad Khan (University of Pittsburgh (US)), Ms Tania Korchuganova10/09/2025, 11:00
The Production and Distributed Analysis (PanDA) workload management system was designed with flexibility to adapt to emerging computing technologies in processing, storage, networking, and distributed computing middleware for the global data distribution. PanDA can coordinate processing over heterogeneous computing resources, including dozens of geographically separated high-performance...
Go to contribution page -
Doug Benjamin (Brookhaven National Laboratory (US))10/09/2025, 11:00
This study evaluates the portability, performance, and adaptability of the Liquid Argon TPC (LAr TPC) detector simulations on different HPC platforms, specifically Polaris, Frontier, and Perlmutter. Lar TCP workflow is a computationally complex workflow which mimics neutrino interactions and the resultant detector responses in a modular liquid argon TPC, integrating various subsystems to...
Go to contribution page -
Leonardo Beltrame (Politecnico di Milano (IT))10/09/2025, 11:00
The Next Generation Trigger project aims to improve the computational efficiency of the CMS reconstruction software (CMSSW) to increase the data processing throughput at the High-Luminosity Large Hadron Collider. As part of this project, this work focuses on improving the common Structure of Arrays (SoA) used in CMSSW for running both on CPUs and GPUs. We introduce a new SoA feature that...
Go to contribution page -
Chance Alan Lavoie (Carnegie Mellon University), Chance Lavoie, Rocky Bala Garg (Stanford University (US))10/09/2025, 11:00
Particle tracking is among the most sophisticated and complex parts of the full event reconstruction chain. Various reconstruction algorithms work in sequence to build trajectories from detector hits. Each of these algorithms requires numerous configuration parameters that need fine-tuning to properly account for the detector/experimental setup, the available CPU budget, and the desired...
Go to contribution page -
Jonas Spinner10/09/2025, 11:00
Generative networks are an exciting tool for fast LHC event generation. Usually, they
Go to contribution page
are used to generate configurations with a fixed number of particles. Autoregressive
transformers allow us to generate events with variable numbers of particles, very much
in line with the physics of QCD jet radiation. We show how they can learn a factorized
likelihood for jet radiation and extrapolate in... -
David Rousseau (IJCLab-Orsay)10/09/2025, 11:00
Measurements and observations in Particle Physics fundamentally depend on one's ability to quantify their uncertainty and, thereby, their significance. Therefore, as Machine Learning methods become more prevalent in HEP, being able to determine the uncertainties of an ML method becomes more important. A wide range of possible approaches has been proposed, however, there has not been a...
Go to contribution page -
Fedor Ratnikov, Foma Shipilov10/09/2025, 11:00
In the end-cap region of the SPD detector complex, particle identification will be provided by a Focusing Aerogel RICH detector (FARICH). FARICH will primarily aid with pion / kaon separation in final open charmonia states (momenta below 5 GeV/c). A free-running (triggerless) data acquisition pipeline to be employed in the SPD results in a high data rate necessitating new approaches to event...
Go to contribution page -
Andrea Favorito10/09/2025, 11:00Track 3: Computations in Theoretical Physics: Techniques and MethodsPoster
I will present joint work on the behavior of Feynman integrals and perturbative expansions at large loop orders. Using the tropical sampling algorithm for evaluating Feynman integrals, along with a dedicated graph-sampling algorithm to generate representative sets of Feynman diagrams, we computed approximately $10^7$ integrals with up to 17 loops in four-dimensional $\phi^4$ theory. Through...
Go to contribution page -
Josef Ruzicka Gonzalez (Costa Rica Center for High Technology), Saverio Mariani (CERN), Sergio Arguedas Cuendis (Consejo Nacional de Rectores (CONARE) (CR))10/09/2025, 11:00
The performance of Particle Identification (PID) in the LHCb experiment is critical for numerous physics analyses. Classifiers, derived from detector likelihoods under various particle mass hypotheses, are trained to tag particles using calibration samples that involve information from the Ring Imaging Cherenkov (RICH) detectors, calorimeters, and muon identification chambers. However, these...
Go to contribution page -
Tadej Novak (Jozef Stefan Institute (SI))10/09/2025, 11:00
Simulating physics processes and detector responses is essential in high energy physics but accounts for significant computing costs. Generative machine learning has been demonstrated to be potentially powerful in accelerating simulations, outperforming traditional fast simulation methods. While efforts have focused primarily on calorimeters initial studies have also been performed on silicon...
Go to contribution page -
Miriam Lucio Martinez (Univ. of Valencia and CSIC (ES))10/09/2025, 11:00Track 3: Computations in Theoretical Physics: Techniques and MethodsPoster
Reconstructing the trajectories of charged particles as they traverse several detector layers is a key ingredient for event reconstruction at LHC and virtually any particle physics experiment. The limited bandwidth available, together with the high rate of tracks per second O(10^10) - where each track consists of a variable number of measurements - makes this problem exceptionally challenging...
Go to contribution page -
Doug Benjamin (Brookhaven National Laboratory (US))10/09/2025, 11:00
We present a novel integration of the PanDA workload management system (PanDA WMS) and Harvester with Globus Compute to enable secure, portable, and remote execution of ATLAS workflows on high-performance computing (HPC) systems. In our approach, Harvester, which runs on an external server, is used to orchestrate job submissions via Globus Compute’s multi-user endpoint (MEP). This MEP provides...
Go to contribution page -
Louis Moureaux (Hamburg University (DE))10/09/2025, 11:00
Despite compelling evidence for the incompleteness of the Standard Model and an extensive search programme, no hints of new physics have so far been observed at the LHC. Anomaly detection was proposed as way to enhance the sensitivity of generic searches not targetting any specific signal model. One of the leading methods in this field, CATHODE (Classifying Anomalies THrough Outer Density...
Go to contribution page -
Jiahui Zhuo (Univ. of Valencia and CSIC (ES))10/09/2025, 11:00
The new fully software-based trigger of the LHCb experiment operates at a 30 MHz data rate and imposes tight constraints on GPU execution time. Tracking reconstruction algorithms in this first-level trigger must efficiently select detector hits, group them, build tracklets, account for the LHCb magnetic field, extrapolate and fit trajectories, and select the best track candidates to filter...
Go to contribution page -
Mohammad Atif (Brookhaven National Laboratory), Doug Benjamin (Brookhaven National Laboratory (US))10/09/2025, 11:00
Recent advancements in large language models (LLMs) have paved the way for tools that can enhance the software development process for scientists. In this context, LLMs excel at two tasks -- code documentation in natural language and code generation in a given programming language. The commercially available tools are often restricted by the available context window size, encounter usage...
Go to contribution page -
Beijiang Liu, Changzheng YUAN, Ke Li (Chinese Academy of Sciences (CN)), Zhengde Zhang (中国科学院高能物理研究所)10/09/2025, 11:00
The data processing and analyzing is one of the main challenges at HEP experiments. To accelerate the physics analysis and drive new physics discovery, the rapidly developing Large Language Model (LLM) is the most promising approach, it have demonstrated astonishing capabilities in recognition and generation of text while most parts of physics analysis can be benefitted. In this talk we will...
Go to contribution page -
R D Schaffer (Université Paris-Saclay (FR))10/09/2025, 11:00
In view of reducing the disk size of the future analysis formats for High Luminosity LHC in the ATLAS experiment, we have explored the use of lossy compression in the newly developed analysis format known as PHYSLITE. Improvements in disk size are being obtained in migrating from the 'traditional' ROOT TTree format to the newly developed RNTuple format. Lossy compression can bring improvements...
Go to contribution page -
Dennis Daniel Nick Noll (Lawrence Berkeley National Lab (US))10/09/2025, 11:00
This contribution discusses an anomaly detection search for narrow-width resonances beyond the Standard Model that decay into a pair of jets. Using 139 fb−1 of proton-proton collision data at sqrt(s) = 13 TeV, recorded from 2015 to 2018 with the ATLAS detector at the Large Hadron Collider, we aim to identify new physics without relying on a specific signal model. The analysis employs two...
Go to contribution page -
Artur Cordeiro Oudot Choi (University of Washington (US))10/09/2025, 11:00
The ServiceX project aims to provide a data extraction and delivery service for HEP analysis data, accessing files from distributed stores and applying user-configured transformations on them. ServiceX aims to support many existing analysis workflows and tools in as transparent a manner as possible, while enabling new technologies. We will discuss the most recent backends added to ServiceX,...
Go to contribution page -
Adam Takacs (Heidelberg University), Alexandre Falcão (University of Bergen)10/09/2025, 11:00
To compare collider experiments, measured data must be corrected for detector distortions through a process known as unfolding. As measurements become more sophisticated, the need for higher-dimensional unfolding increases, but traditional techniques have limitations. To address this, machine learning-based unfolding methods were recently introduced. In this work, we introduce OmniFold-HI, an...
Go to contribution page -
lei wang (Institute of High Energy Physics)10/09/2025, 11:00
When we try to move the software named ‘Hepsptycho’, which is a ptychography reconstruction program originally based on multiple Nvidia GPU and MPI techs, to run on the Hygon DCU architectures, we found that the reconstructed object and probe encountered an error while the results running on Nvidia GPUs are correct. We profiled the ePIE algorithm using NVIDIA Nsight Systems and Hygon's...
Go to contribution page -
Benjamin Fischer (RWTH Aachen University (DE))10/09/2025, 11:00
Workflow tools provide the means to codify complex multi-step processes, thus enabling reproducibility, preservation, and reinterpretation efforts. Their powerful bookkeeping also directly supports the research process, especially where intermediate results are produced, inspected, and iterated upon frequently.
Go to contribution page
In Luigi, such a complex workflow graph is composed of individual tasks that... -
Chiara Signorile10/09/2025, 11:00Track 3: Computations in Theoretical Physics: Techniques and MethodsPoster
In this presentation, I will discuss recent advancements in NNLO+PS predictions for top-quark pair production and decay within the MiNNLO framework. MiNNLO provides a robust method for incorporating next-to-next-to-leading order (NNLO) QCD corrections directly into fully differential predictions, offering unprecedented accuracy. This approach enables a consistent treatment of both production...
Go to contribution page -
Thorsten Lars Henrik Buss (Universität Hamburg)10/09/2025, 11:00
Simulating showers of particles in highly-granular detectors is a key frontier in the application of machine learning to particle physics. Achieving high accuracy and speed with generative machine learning models can enable them to augment traditional simulations and alleviate a major computing constraint.
Recent developments have shown how diffusion based generative shower simulation...
Go to contribution page -
Karthik Suresh (College of William and Mary)10/09/2025, 11:00
Artificial Intelligence is set to play a transformative role in designing large and complex detectors, such as the ePIC detector at the upcoming Electron-Ion Collider (EIC). The ePIC setup features a central detector and additional systems positioned in the far forward and far backward regions. Designing this system involves balancing many factors—performance, physics goals, and cost—while...
Go to contribution page -
Saverio Monaco10/09/2025, 11:00
Quantum generative modeling provides an alternative framework for simulating complex processes in high-energy physics. Calorimeter shower simulations, in particular, involve high-dimensional, stochastic data and are essential for particle identification and energy reconstruction at experiments such as those at the LHC. As these simulations increase in complexity—especially in large-scale...
Go to contribution page -
Siyang Wu (Shandong University)10/09/2025, 11:00
The $Z_c(3900)$ was first discovered by the Beijing Spectrometer (BESIII) detector in 2013. As one of the most attractive discoveries of the BESIII experiment, $Z_c(3900)$ itself has inspired extensive theoretical and experimental research on its properties. In recent years, the rapid growth of massive experimental data at high energy physics (HEP) experiments have driven a lot of novel...
Go to contribution page -
Alessandro Brusamolino (KIT - Karlsruhe Institute of Technology (DE)), Katharina Schäuble10/09/2025, 11:00
One of the main points of object reconstruction is the definition of the targets of our reconstruction. In this talk we present recent developments on the topic, focusing on how we can embed detector constraints, mainly calorimeter granularity, in our truth information and how this can impact the performance of the reconstruction, in particular for Machine Learning based approaches. We will...
Go to contribution page -
Xiaofei Yan (Institute of High Energy Physics(IHEP))10/09/2025, 11:00
To standardize the evaluation of computational capabilities across various hardware architectures in data centers, we developed a CPU performance benchmarking tool within the HEP-Score framework. The tool uses the JUNO offline software as a realistic workload and produces standardized outputs aligned with HEP-Score requirements. Our tests demonstrate strong linear performance characteristics...
Go to contribution page -
Shuang Wang (IHEP)10/09/2025, 11:00
Astronomical satellites serve as critical infrastructure in the field of astrophysics, and data processing is one of the most essential processes for conducting scientific research on cosmic evolution, celestial activities, and dark matter. Recent advancements in satellite sensor resolution and sensitivity have led to petabyte (PB)-scale data volumes, characterized by unprecedented scale and...
Go to contribution page -
Umit Sozbilir (Universita e INFN, Bari (IT))10/09/2025, 11:00
Hyperparameter optimization plays a crucial role in achieving high performance and robustness for machine learning models, such those used in complex classification tasks in High Energy Physics (HEP).
Go to contribution page
In this study, we investigate and experience the usage of $\texttt{Optuna}$, a rather new, modern and scalable optimization tool in the framework of a realistic signal-versus-background... -
Raees Ahmad Khan (University of Pittsburgh (US))10/09/2025, 11:00
In large-scale distributed computing systems, workload dispatching and the associated data management are critical factors that determine key metrics such as resource utilization of distributed computing and resilience of scientific workflows. As the Large Hadron Collider (LHC) advances into its high luminosity era, the ATLAS distributed computing infrastructure must improve these metrics to...
Go to contribution page -
Volodymyr Svintozelskyi (Univ. of Valencia and CSIC (ES))10/09/2025, 11:00
The LHCb collaboration is currently using a pioneer system of data filtering in the trigger system, based on real-time particle reconstruction using Graphics Processing Units (GPUs). This corresponds to processing 5 TB/s of data and has required a huge amount of hardware and software developments. Among them, the corresponding power consumption and sustainability is an imperative matter in...
Go to contribution page -
FU Shiyuan fusy10/09/2025, 11:00
The High Energy Photon Source produces vast amounts of diverse, multi-modal data annually, with IO bottlenecks increasingly limiting scientific computational efficiency. To overcome this challenge, our approach introduces a threefold solution. First, we develop daisy-io, which has a unified IO interface designed for cross-disciplinary applications, which integrates accelerated data retrieval...
Go to contribution page -
Andrey Shevelev, Fedor Ratnikov10/09/2025, 11:00
The increasing reliance on machine learning (ML) and particularly deep learning (DL) in scientific and industrial applications requires models that are not only accurate, but also reliable under varying conditions. This is especially important for automated machine learning and fault-tolerant systems where there is limited or no human control. In this paper, we present a novel,...
Go to contribution page -
Jeremy Couthures (Laboratoire d'Annecy de physique des particules, CNRS / Univ. Savoie Mont Blanc (FR))10/09/2025, 11:00
As the High-Luminosity LHC (HL-LHC) era approaches, significant improvements in
Go to contribution page
reconstruction software are required to keep pace with the increased data rates and
detector complexity. A persistent challenge for high-throughput event reconstruction is
the estimation of track parameters, which is traditionally performed using iterative
Kalman Filter-based algorithms. While GPU-based track... -
Gordon Watts (University of Washington (US))10/09/2025, 11:00
ServiceX, a data extraction and delivery service for HEP experiments, is being used in ATLAS to prepare training data for a long-lived particle search. The training data contains low-level features not available in the ATLAS experiment’s PHYSLITE format - making ServiceX’s ability to read complex event data (e.g. ATLAS’s xAOD format) ideally suited to solving this problem. This poster will...
Go to contribution page -
Luigi Favaro (Universite Catholique de Louvain (UCL) (BE))10/09/2025, 11:00
The speed and fidelity of detector simulations in particle physics pose compelling questions on future LHC analysis and colliders. The sparse high-dimensional data combined with the required precision provide a challenging task for modern generative networks. We present a general framework to train generative networks on any detector geometry with minimal user input. Vision transformers allow...
Go to contribution page -
Martin Foll (University of Oslo (NO))10/09/2025, 11:00
The ROOT software framework is widely used in HEP for storage, processing, analysis and visualization of large datasets. With the large increase in usage of ML for experiment workflows, especially lately in the last steps of the analysis pipeline, the matter of exposing ROOT data ergonomically to ML models becomes ever more pressing. In this contribution we discuss the experimental component...
Go to contribution page -
Theo Heimel (UCLouvain)10/09/2025, 11:10Track 3: Computations in Theoretical Physics: Techniques and MethodsPoster
Machine learning methods enable unbinned and full-dimensional unfolding. However, existing approaches, both classifier-based and generative, suffer from prior dependence. We propose a new method for ML-based unfolding that is completely prior independent and infers the unfolded distribution in a fully frequentist manner. Using several benchmark datasets, we demonstrate that the method can...
Go to contribution page -
Adriano Di Florio (CC-IN2P3), CMS Collaboration10/09/2025, 11:30Track 1: Computing Technology for Physics ResearchOral
For over two decades, computing resources in the Worldwide LHC Computing Grid (WLCG) have been based exclusively on the x86 architecture. However, in the near future, heterogeneous non-x86 architectures are expected to make up a significant portion of the resources available to LHC experiments driven also by their adoption in current and upcoming world-class HPC facilities. In response to this...
Go to contribution page -
Ayodele Ore10/09/2025, 11:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Significant efforts are currently underway to improve the description of hadronization using Machine Learning. While modern generative architectures can undoubtedly emulate observations, it remains a key challenge to integrate these networks within principled fragmentation models in a consistent manner. This talk presents developments in the HOMER method for extracting Lund fragmentation...
Go to contribution page -
Gagik Gavalian (Jefferson National Lab)10/09/2025, 11:30
Charged track reconstruction is a critical task in nuclear physics experiments, enabling the identification and analysis of particles produced in high-energy collisions. Machine learning (ML) has emerged as a powerful tool for this purpose, addressing the challenges posed by complex detector geometries, high event multiplicities, and noisy data. Traditional methods rely on pattern recognition...
Go to contribution page -
Jamal Slim (DESY)10/09/2025, 11:50Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
We present a quantum generative model that extends Quantum Born Machines (QBMs) by incorporating a parametric Polynomial Chaos Expansion (PCE) to encode classical data distributions. Unlike standard QBMs relying on fixed heuristic data-loading strategies, our approach employs a trainable Hermite polynomial basis to amplitude-encode classical data into quantum states. These states are...
Go to contribution page -
Francesco Fenu (Agenzia Spaziale Italiana)10/09/2025, 11:50
The Compton Spectrometer and Imager (COSI) is a NASA Small Explorer (SMEX) satellite mission planned to fly in 2027. It has the participation of institutions in the US, Europe and Asia and aims at the construction of a gamma-ray telescope for observations in the 0.2-5 MeV energy range. COSI consists of an array of germanium strip detectors cooled to cryogenic temperatures with millimeter...
Go to contribution page -
Alan Malta Rodrigues (University of Notre Dame (US)), CMS Collaboration10/09/2025, 11:50Track 1: Computing Technology for Physics ResearchOral
Flexible workload specification and management are critical to the success of the CMS experiment, which utilizes approximately half a million cores across a global grid computing infrastructure for data reprocessing and Monte Carlo production. TaskChain and StepChain specifications, responsible for over 95% of central production activities, employ distinct workflow paradigms: TaskChain...
Go to contribution page -
Timo Janssen (University of Göttingen)10/09/2025, 12:10Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
We apply for the first time the Flow Matching method to the problem of phase-space sampling for event generation in high-energy collider physics. By training the model to remap the random numbers used to generate the momenta and helicities of the collision matrix elements as implemented in the portable partonic event generator Pepper, we find substantial efficiency improvements in the studied...
Go to contribution page -
Dr Marcel Völschow (Hamburg University of Applied Sciences)10/09/2025, 12:10
Beyond the planet Neptune, only the largest solar system objects can be observed directly. However, there are tens of thousands of smaller objects whose frequency and distribution could provide valuable insights into the formation of our solar system - if we could see them.
Project SOWA (Solar-system Occultation Watch and Analysis) aims to systematically search for such invisible objects...
Go to contribution page -
Artur Cordeiro Oudot Choi (University of Washington (US))10/09/2025, 12:10Track 1: Computing Technology for Physics ResearchOral
As the HL-LHC prepares to deliver large volumes of data, the need for an efficient data delivery and transformation service becomes crucial. To address this challenge, a cross-experiment toolset—ServiceX—was developed to link the centrally produced datasets to flexible, user-level analysis workflows. Modern analysis tools such as Coffea benefit from ServiceX as the first step in event...
Go to contribution page -
Julian Simon Schliwinski (Humboldt University of Berlin (DE))10/09/2025, 12:30
The next generation of ground-based gamma-ray astronomy instruments will involve arrays of dozens of telescopes, leading to an increase in operational and analytical complexity. This scale-up poses challenges for both system operations and offline data processing, especially when conventional approaches struggle to scale effectively. To address these challenges, we are developing AI agents...
Go to contribution page -
Yu Hu10/09/2025, 12:30Track 1: Computing Technology for Physics ResearchOral
The High Energy Photon Source (HEPS), a new fourth-generation high-energy synchrotron radiation facility, is set to become fully operational by the end of 2025. With its significantly enhanced brightness and detector performance, HEPS will generate over 300 PB of experimental data annually across 14 beamlines in phase I, quickly reaching the EB scale. HEPS supports a wide range of experimental...
Go to contribution page -
Theo Heimel (UCLouvain)10/09/2025, 12:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
MadEvent7 is a new modular phase-space generation library written in C++ and CUDA, running on both GPUs and CPUs. It features a variety of different phase-space mappings, including the classic MadGraph multi-channel phase space and an optimized implementation of normalizing flows for neural importance sampling, as well as their corresponding inverse mappings. The full functionality is...
Go to contribution page -
Lia Lavezzi (INFN Torino), Lia Lavezzi (INFN Torino (IT))10/09/2025, 12:50Track 1: Computing Technology for Physics ResearchOral
In the last decade, the concept of Open Science has gained importance: there is a real effort to make tools and data shareable among different communities, with the goal of making data and software FAIR (Findable, Accessible, Interoperable and Reusable). This goal is shared by several scientific communities, including the Einstein Telescope (ET). ET is the third generation ground-based...
Go to contribution page -
Konrad Helms10/09/2025, 12:50Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Modern approaches to phase-space integration combine well-established Monte Carlo methods with machine learning techniques for importance sampling. Recent progress in generative models in the form of continuous normalizing flows, trained using conditional flow matching, offers the potential to improve the phase-space sampling efficiency significantly.
Go to contribution page
We present a multi-jet inclusive... -
Stephen Jiggins (Deutsches Elektronen-Synchrotron (DE))10/09/2025, 12:50
In many domains of science the likelihood function is a fundamental ingredient used to statistically infer model parameters from data, due to the likelihood ratio (LR) as an optimal test statistic. Neural based LR estimation using probabilistic classification has therefore had a significant impact in these domains, providing a scalable method for determining an intractable LR from simulated...
Go to contribution page -
Giuseppe Andronico (Universita e INFN, Catania (IT))10/09/2025, 13:10Track 1: Computing Technology for Physics ResearchOral
The Jiangmen Underground Neutrino Observatory (JUNO) is an underground 20 kton liquid scintillator detector being constructed in southern China. The JUNO physics program aims to explore neutrino properties, particularly through electron anti-neutrinos emitted from two nuclear power complexes at a baseline of approximately 53 km. Targeting an unprecedented relative energy resolution of 3% at 1...
Go to contribution page -
Nikita Schmal10/09/2025, 13:10Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
One primary goal of the LHC is the search for physics beyond the Standard Model, leading to the development of many different methods to look for new physics effects. In this context, we employ Machine Learning methods, in particular we explore the applications of Simulation-Based Inference (SBI), to learn otherwise intractable likelihoods and fully exploit the information available, compared...
Go to contribution page -
Gregor Kasieczka (Hamburg University (DE))11/09/2025, 09:15
-
Paul Stevenson (University of Surrey)11/09/2025, 09:30Plenary
-
Sioni Paris Summers (CERN)11/09/2025, 10:00Plenary
-
Eugenio Valdano (INSERM)11/09/2025, 10:30Plenary
-
David Shih11/09/2025, 11:30Plenary
-
Bo-Cheng Lai11/09/2025, 12:00Plenary
-
Jay Ajitbhai Sandesara (University of Wisconsin Madison (US))11/09/2025, 12:30
Neural Simulation-Based Inference (NSBI) is an emerging class of statistical methods that harness the power of modern deep learning to perform inference directly from high-dimensional data. These techniques have already demonstrated significant sensitivity gains in precision measurements across several domains, outperforming traditional approaches that rely on low-dimensional summaries. This...
Go to contribution page -
Pia Bredt, Juergen Reuter (DESY Hamburg, Germany)11/09/2025, 14:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
We discuss recent developments in performance improvements for Monte Carlo integration and event sampling. (1) Massive parallelization of matrix element evaluations based on a new back end for the matrix element generator O'Mega targeting GPUs. This has already been integrated in a development version of the Monte Carlo event generator Whizard for realistic testing and profiling. (2) A...
Go to contribution page -
Laurent Dufour (CERN)11/09/2025, 14:30Track 1: Computing Technology for Physics ResearchOral
The LHCb experiment, one of the four major experiments at the Large Hadron Collider (LHC), excels in high-precision measurements of particles that are produced relatively frequently (strange, charmed and bottom hadrons). Key to LHCb's potential is its sophisticated trigger system that enables complete event reconstruction, selection, alignment and calibration in real-time. Through the Turbo...
Go to contribution page -
Simon Hiesl11/09/2025, 14:30
In anticipation of higher luminosities at the Belle II experiment, high levels of beam background
Go to contribution page
from outside of the interaction region are expected. To prevent track trigger rates
from surpassing the limitations of the data acquisition system, an upgrade of the first-level
neural track trigger becomes indispensable. This upgrade contains a novel track finding
algorithm based on... -
李亚康 liyk11/09/2025, 14:50Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
To characterize the structures and properties of samples in the analysis of experimental data of Small-Angle Neutron Scattering (SANS), a physical model must be selected corresponding to each sample for iterative fitting. However, the conventional method of model selection is primarily based on manual experience, which has a high threshold for analysis and low accuracy. Furthermore, the...
Go to contribution page -
Gagik Gavalian (Jefferson National Lab)11/09/2025, 14:50Track 1: Computing Technology for Physics ResearchOral
In this article, we present the High-Performance Output (HiPO) data format developed at Jefferson Laboratory for
Go to contribution page
storing and analyzing data from Nuclear Physics experiments. The format was designed to efficiently store large
amounts of experimental data, utilizing modern fast compression algorithms. The purpose of this development was
to provide organized data in the output, facilitating... -
Lennart Uecker (Heidelberg University (DE))11/09/2025, 14:50
The LHCb experiment at the Large Hadron Collider (LHC) operates a fully software-based trigger system that processes proton-proton collisions at a rate of 30 MHz, reconstructing both charged and neutral particles in real time. The first stage of this trigger system, running on approximately 500 GPU cards, performs a track pattern recognition to reconstruct particle trajectories with low...
Go to contribution page -
Iason Krommydas (Rice University (US))11/09/2025, 15:10Track 1: Computing Technology for Physics ResearchOral
High-energy physics (HEP) analyses routinely handle massive datasets, often exceeding the available resources. Efficiently interacting with these datasets requires dedicated techniques for data loading and management. Awkward Array is a Python library widely used in high-energy physics (HEP) to efficiently handle complex, irregularly structured...
Go to contribution page -
Giulio Cordova (Universita & INFN Pisa (IT))11/09/2025, 15:10
The upgraded LHCb experiment is pioneering the landscape of real-time data-processing techniques using an heterogeneous computing infrastructure, composed of both GPUs and FPGAs, aimed at boosting the performance of the HLT1 reconstruction. Amongst the novelties in the reconstruction infrastructure made for the Run 3, the introduction of a real-time VELO hit-finding FPGA-based architecture...
Go to contribution page -
Daniel Schiller (Institute for Theoretical Physics Heidelberg)11/09/2025, 15:10Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Foundation models are a very successful approach to linguistic tasks. Naturally, there is the desire to develop foundation models for physics data. Currently, existing networks are much smaller than publicly available Large Language Models (LLMs), the latter having typically billions of parameters. By applying pretrained LLMs in an unconventional way, we introduce large networks for...
Go to contribution page -
Jolly Chen (CERN & University of Twente (NL)), Leonardo Beltrame (Politecnico di Milano (IT)), Dr Oliver Gregor Rietmann (CERN)11/09/2025, 15:30Track 1: Computing Technology for Physics ResearchOral
Based on previous experience with parallel event data processing, a Structure-of-Arrays (SoA) layout for objects frequently performs better than Array-of-Structures (AoS), especially on GPUs. However, AoS is widespread in existing code, and in C++, changing the data layout from AoS to SoA requires changing the data structure declarations and the access syntax. This work is repetitive, time...
Go to contribution page -
Dr Ramon Winterhalder (Università degli Studi di Milano)11/09/2025, 15:30Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
The increasing complexity of modern neural network architectures demands fast and memory-efficient implementations to mitigate computational bottlenecks. In this work, we evaluate the recently proposed BitNet architecture in HEP applications, assessing its performance in classification, regression, and generative modeling tasks. Specifically, we investigate its suitability for quark-gluon...
Go to contribution page -
CMS Collaboration, Mario Masciovecchio (Univ. of California San Diego (US))11/09/2025, 15:30
Charged particle track reconstruction is one the heaviest computational tasks in the event reconstruction chain at Large Hadron Collider (LHC) experiments. Furthermore, projections for the High Luminosity LHC (HL-LHC) show that the required computing resources for single-threaded CPU algorithms will exceed those that are expected to be available. It follows that experiments at the HL-LHC will...
Go to contribution page -
Qi Bin Lei (Stanford University (US)), Rocky Bala Garg (Stanford University (US))11/09/2025, 15:50
The exponential time scaling of traditional primary vertex reconstruction algorithms raises significant performance concerns for future high-pileup environments, particularly with the upcoming High Luminosity upgrade to the Large Hadron Collider. In this talk, we introduce PV-Finder, a deep learning-based approach that leverages reconstructed track parameters to directly predict primary vertex...
Go to contribution page -
Jens Braun (Karlsruhe Institute of Technology (KIT))11/09/2025, 15:50Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
We present FeynGraph, a modern high-performance Feynman diagram generator designed to integrate seamlessly with modern computational workflows to calculate scattering amplitudes. FeynGraph is designed as a high-performance Rust library with easy-to-use Python bindings, allowing it to be readily used in other tools. With additional features like arbitrary custom diagram selection filters and...
Go to contribution page -
Giacomo Parolini (CERN)11/09/2025, 15:50Track 1: Computing Technology for Physics ResearchOral
RNTuple is ROOT's next-generation columnar format, replacing TTree as the primary storage solution for LHC event data.
Go to contribution page
After 6 years of R&D, the RNTuple format reached the 1.0 milestone in 2024, promising full backward compatibility with any data written from that moment on.
Designing a new on-disk format does not only allow for significant improvements on file sizes and read/write speed,... -
Florian Ernst (Heidelberg University (DE))11/09/2025, 16:40Track 1: Computing Technology for Physics ResearchOral
For high-energy physics experiments, the generation of Monte Carlo events, and particularly the simulation of the detector response, is a very computationally intensive process. In many cases, the primary bottleneck in detector simulation is the detailed simulation of the electromagnetic and hadronic showers in the calorimeter system.
ATLAS is currently using its state-of-the-art fast...
Go to contribution page -
Antoine Petitjean (ITP, Universität Heidelberg)11/09/2025, 16:40
Unfolding detector-level data into meaningful particle-level distributions remains a key challenge in collider physics, especially as the dimensionality of the relevant observables increases. Traditional unfolding techniques often struggle with such high-dimensional problems, motivating the development of machine learning-based approaches.We introduce a new method for generative unfolding that...
Go to contribution page -
Dr Anthony Hartin (LMU)11/09/2025, 16:40Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Non perturbative QED is used in calculations of Schwinger pair creation, in precision QED tests with ultra-intense lasers and to predict beam backgrounds at the interaction point of colliders. In order to predict these phenomena, custom built monte carlo event generators based on a suitable non perturbative theory have to be developed. One such suitable theory uses the Furry Interaction...
Go to contribution page -
Lorenzo Valente (University of Hamburg (DE))11/09/2025, 17:00Track 1: Computing Technology for Physics ResearchOral
The accurate simulation of particle showers in collider detectors remains a critical bottleneck for high-energy physics research. Current approaches face fundamental limitations in scalability when modeling the complete shower development process.
Deep generative models offer a promising alternative, potentially reducing simulation costs by orders of magnitude. This capability becomes...
Go to contribution page -
Sofia Palacios Schweitzer (ITP, University Heidelberg)11/09/2025, 17:00
Two shortcomings of classical unfolding algorithms, namely that they are defined on binned, one-dimensional observables, can be overcome when using generative machine learning. Many studies on generative unfolding reduce the problem to correcting for detector smearing, however a full unfolding pipeline must also account for background, acceptance and efficiency effects. To fully integrate...
Go to contribution page -
York Schröder .11/09/2025, 17:00Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
The determination of the hot QCD pressure has a long history, and has -- due to its phenomenological relevance in cosmology, astrophysics and heavy-ion collisions -- spawned a number of important theoretical advances in perturbative thermal field theory applicable to equilibrium thermodynamics.
We present major progress towards the determination of the last missing piece for the pressure of...
Go to contribution page -
Dr Mirko Bunse (Lamarr Institute for Machine Learning and Artificial Intelligence, Dortmund, Germany)11/09/2025, 17:20
Measured distributions are usually distorted by a finite resolution of the detector. Within physics research, the necessary correction of these distortions is know as Unfolding. Machine learning research uses a different term for this very task: Quantification Learning. For the past two decades, this difference in terminology - together with several differences in notation - have prevented...
Go to contribution page -
Hayden Richard Hollenbeck (University of Virginia (US))11/09/2025, 17:20Track 1: Computing Technology for Physics ResearchOral
Next-generation High Energy Physics (HEP) experiments face unprecedented computational demands. The High-Luminosity Large Hadron Collider anticipates data processing needs that will exceed available resources, while intensity frontier experiments such as DUNE and LZ are dominated by the simulation of high-multiplicity optical photon events. Heterogeneous architectures, particularly GPU...
Go to contribution page -
Sanjaye Ramgoolam (Queen Mary University of London)11/09/2025, 17:20Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Random matrix theory has a long history of applications in the study of eigenvalue distributions arising in diverse real-world ensembles of matrix data. Matrix models also play a central role in theoretical particle physics, providing tractable mathematical models of gauge-string duality, and allowing the computation of correlators of invariant observables in physically interesting sectors of...
Go to contribution page -
Severin Diederichs (CERN)11/09/2025, 17:40Track 1: Computing Technology for Physics ResearchOral
The simulation throughput of LHC experiments is increasingly limited by detector complexity in the high-luminosity phase. As high-performance computing shifts toward heterogeneous architectures such as GPUs, accelerating Geant4 particle transport simulations by offloading parts of the workload to GPUs can improve performance. The AdePT plugin currently offloads electromagnetic showers in...
Go to contribution page -
Prof. Marco Meyer-Conde (Tokyo City University (JP), University Of Illinois (US))11/09/2025, 17:40Track 3: Computations in Theoretical Physics: Techniques and MethodsOral
Gravitational Wave (GW) Physics has entered a new era of Multi-Messenger Astronomy (MMA), characterized by increasing detections from GW observatories such as the LIGO, Virgo, and KAGRA collaborations. This presentation will introduce the KAGRA experiment, outlining the current workflow from data collection to physics interpretation, and demonstrate the transformative role of machine learning...
Go to contribution page -
David Walter (Massachusetts Inst. of Technology (US))11/09/2025, 17:40
The High-Luminosity LHC era will deliver unprecedented data volumes, enabling measurements on fine-grained multidimensional histograms containing millions of bins with thousands of events each. Achieving ultimate precision requires modeling thousands of systematic uncertainty sources, creating computational challenges for likelihood minimization and parameter extraction. Fast minimization is...
Go to contribution page -
Maciej Besta (ETHZ)12/09/2025, 09:10Plenary
-
Peter McKeown (CERN)12/09/2025, 10:00
The need for fast calorimeter shower simulation tools has spurred the development of numerous surrogate approaches based on deep generative models. While these models offer significant reductions in compute times with respect to traditional Monte Carlo methods, their development consumes significant amounts of time, manpower and computing resources.
In order to reduce the time to design a...
Go to contribution page -
12/09/2025, 10:30Plenary
-
Nick Smith (Fermi National Accelerator Lab. (US))12/09/2025, 11:30
-
Thea Aarrestad12/09/2025, 11:50
-
Tianji Cai12/09/2025, 12:10
-
David Britton (University of Glasgow (GB))12/09/2025, 12:30
-
Jens Behrmann (Apple)Plenary
-
Alexandru Calotoiu (ETHZ)Plenary
Choose timezone
Your profile timezone: