Help us make Indico better by taking this survey! Aidez-nous à améliorer Indico en répondant à ce sondage !

SMARTHEP Edge Machine Learning School

Europe/Zurich
222/R-001 (CERN)

222/R-001

CERN

200
Show room on map
Anna Sfyrla (Universite de Geneve (CH)), Maurizio Pierini (CERN), Sioni Paris Summers (CERN), Thea Aarrestad (ETH Zurich (CH))
Description

Welcome to the NextGen and SMARTHEP School on Edge Machine Learning! 

smarthep edge ml school

This  training program is designed to provide participants with knowledge and hands-on experience in the emerging field of Edge Machine Learning, tailored for applications at the LHC.

The program includes introductory lectures, tutorials, and seminars covering a variety of related topics, with ample time for discussion. Tutorials include:

  • Model compression techniques
  • Efficient GPU inference
  • Efficient FPGA inference
  • Neuromorphic computing and spiking neural networks
  • Heterogeneous accelerated computing

 

We will also feature talks on real-time inference applications for autonomous vehicles, satellites, gravitational wave physics, high energy particle physics and more.

Additionally, the program features contributed flash talks and poster sessions, allowing attendees to share their work and insights. And a PhD student poster prize awaits!

Participants
  • Ahmed Elghareeb
  • Alejandro Pérez Aguilera
  • Alina Lazar
  • Andreas Salzburger
  • Aryaman Pattnayak
  • Aryan Saini
  • Benjamin Ramhorst
  • Carlos Eduardo Cocha Toapaxi
  • Caue Evangelista de Sousa
  • Chang Sun
  • Christine Zeh
  • Christof Sauer
  • Christopher Edward Brown
  • Cristina Botta
  • Daniel Estrada
  • Daniela Katherinne Paredes Hernandez
  • David Reikher
  • Dolores Garcia
  • Dylan Sheldon Rankin
  • Efe Yigitbasi
  • Eleni Xochelli
  • Elvira Rossi
  • Eric Anton Moreno
  • Fanqiang Meng
  • Fotis Giasemis
  • Francesco Vaselli
  • Geetika Gupta
  • Georgios Krintiras
  • Gianluca Cerminara
  • Gregor Krzmanc
  • Guillermo Hijano
  • Hampus Linander
  • Haoyang Li
  • Iaroslava Bezshyiko
  • Ioannis Xiotidis
  • Irene Andreou
  • Jaroslaw Szumega
  • Joaquin Hoya
  • Joerg Stelzer
  • Joshua Falco Beirer
  • Kaare Endrup Iversen
  • Kunihiro Nagano
  • Kyungmin Park
  • Laura Boggia
  • Leon Bozianu
  • Leonid Burmistrov
  • Licheng ZHANG
  • Mansoora Shamim
  • Manuel Gonzalez Berges
  • Marco Fariselli
  • Martin Roosen
  • Martino Borsato
  • Max Cohen
  • Maximilian Amerl
  • Michael Kagan
  • Micol Olocco
  • Nicolò Ghielmetti
  • Noah Clarke Hall
  • Noemi D'Abbondanza
  • Panagiotis Bellos
  • Parth Joshi
  • Patin Inkaew
  • Patricia Rebello Teles
  • Paul Thompson
  • Piero Viscone
  • Pierre Pelissou
  • Qibin Liu
  • Rimsky Alejandro Rojas Caballero
  • Robin Syring
  • Roope Oskari Niemi
  • Santosh Parajuli
  • Sebastian Schmitt
  • Sebastien Rettie
  • Sergio Perez
  • Simone Capelli
  • Simone Machetti
  • Stanislaw Wozniak
  • Stefan Katsarov
  • Stefano Veneziano
  • Sten Åstrand
  • Steven Schramm
  • Stylianos Tzelepis
  • Tanguy Dietrich
  • Thomas Ortner
  • Tianjia Du
  • Tjark Miener
  • Tobias Becker
  • Towsifa Akhter
  • Una Helena Alberti
  • Valentina Camagni
  • Vasilis Belis
  • Viktor Shcherbakov
  • Vilius Čepaitis
  • Yifan Yang
  • Yuan-Tang Chou
  • Zhibin Yang
  • +85
Videoconference
SMARTHEP Edge Machine Learning School
Zoom Meeting ID
64035676853
Host
Anna Sfyrla
Alternative host
Maurizio Pierini
Useful links
Join via phone
Zoom URL
    • 09:00 10:45
      Lectures: ML 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Anna Sfyrla (Universite de Geneve (CH))
      • 09:00
        Introduction to ML 30m
        Speaker: Michael Kagan (SLAC National Accelerator Laboratory (US))
      • 09:30
        Discussion 15m
      • 09:45
        ML on the Edge at the LHC experiments 45m
        Speaker: Dylan Sheldon Rankin (University of Pennsylvania (US))
      • 10:30
        Discussion 15m
    • 10:45 11:15
      Coffee break 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 11:15 12:00
      Lectures: GPUs 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      • 11:15
        Introduction to GPUs 30m
        Speaker: NVIDIA team
      • 11:45
        Discussion 15m
    • 12:00 13:30
      Lunch break 1h 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 13:30 14:30
      Lectures: GPUs 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Thea Aarrestad (ETH Zurich (CH))
      • 13:30
        ML inference on GPUs 45m
        Speaker: NVIDIA team
      • 14:15
        Discussion 15m
    • 14:30 16:00
      Tutorial: GPUs 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Thea Aarrestad (ETH Zurich (CH))
      • 14:30
        Tutorial on network optimization for GPUs 1h 30m
        Speaker: NVIDIA team
    • 16:00 16:30
      Coffee break 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map

      Coffee offered at the mezzanine outside the main auditorium

    • 16:30 18:00
      Flash talks / poster session: Flash talks and poster session in Auditorium Mezzanine 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map

      All contributors will present a poster, which they will need to print by themselves. Poster holders will be available on Monday at the mezzanine. Please also upload the pdf of the poster in the agenda.

      Some contributors have also asked for a flash talk. The duration of the flash talks is 3 mins. Please upload the flash talk contribution by Monday morning 11am. Thanks!

      Convener: Thea Aarrestad (ETH Zurich (CH))
      • 16:30
        FlashSim: Towards a Digital Twin of the CMS Experiment using Normalizing Flows and Flow Matching 3m

        The simulation of high-energy physics collision events is a key element for data analysis at present and future particle accelerators. The comparison of simulation predictions to data allows us to look for rare deviations that can be due to new phenomena not previously observed. The CMS Collaboration is investigating how novel machine learning algorithms, specifically Normalizing Flows and Flow Matching, can be used to perform accurate simulations with several orders of magnitude of speed-up compared to traditional approaches, contributing to the development of a "Digital Twin" of the CMS Experiment, a simulation framework named FlashSim. The classical simulation chain computes energy deposits, electronics response, and reconstruction from a physics process. We propose an end-to-end approach, directly simulating the final high-level format from physical inputs, skipping intermediate steps. The speed and accuracy of the proposed approach make it a compelling tool for the present and future needs of CMS Collaboration.

        Speaker: Francesco Vaselli (Scuola Normale Superiore & INFN Pisa (IT))
      • 16:33
        Improving the Inference of Graph Neural Networks for Track Reconstruction 3m

        Optimizing the inference of Graph Neural Networks (GNNs) for track finding is crucial for enhancing the computing performance of particle collision event reconstruction. Track finding involves identifying and reconstructing the paths of particles from complex, noisy detector data. By leveraging GNNs, we can model the relationships between detector hits as a graph, where nodes represent hits and edges represent potential connections between them. To speed up the inference of these GNN models, it is important to reduce computational overhead, improve model architecture, and exploit hardware accelerators such as GPUs. Techniques like quantization and pruning can be employed to minimize model size and inference time without sacrificing accuracy.

        Speakers: Alina Lazar (Youngstown State University (US)), Henry Paschke, James jsgaboriaultwhit@student.ysu.edu, Jay Chan (Lawrence Berkeley National Lab. (US)), Minh Tuan Dang (CERN), Paolo Calafiura (Lawrence Berkeley National Lab. (US)), Xiangyang Ju (Lawrence Berkeley National Lab. (US))
      • 16:36
        An Open-Source RISC-V-based GPGPU Accelerator for Machine Learning-based Edge Computing Applications 3m

        In recent years, the demand for real-time machine learning (ML)-based computing solutions has driven the rapid growth of edge computing. The adopted hardware must strike a delicate balance by providing sufficient computational power to meet stringent real-time constraints while minimizing energy consumption. General-purpose graphics processing units (GPGPU) are a commonly adopted solution to maximize the data parallelism of ML algorithms thanks to more specialized hardware.

        This work presents an open-source RISC-V-based GPGPU accelerator designed to support research in the ML-based edge computing domain. The accelerator features a low-power GPGPU streaming multiprocessor (SM) and offers two selectable memory hierarchies: cache-based and scratchpad-based. Its high configurability, regarding the number of threads, warps, and memory sizes, enables matching specific application requirements. The accelerator has been integrated into our in-house-designed eXtendible Heterogeneous Energy Efficient Platform (X-HEEP) microcontroller to improve its data processing capabilities and provide a real-world integration example.

        Speaker: Simone Machetti
      • 16:40
        Accelerating Machine Learning algorithms in FPGAs for the trigger system of a SiPM-based upgraded camera of the CTA Large-Sized Telescopes 3m

        Current Imaging Atmospheric Cherenkov Telescopes use combined analog and digital electronics for their trigger systems, implementing simple but fast algorithms. Such trigger techniques are used due to high data rates and strict timing requirements. In recent years, in the context of a possible upgraded camera for the Large-Sized Telescopes (LSTs) of the Cherenkov Telescope Array (CTA) based on Silicon PhotoMultipliers, a new fully digital trigger system incorporating Machine Learning (ML) algorithms is being developed. The main concept is to implement those algorithms in FPGAs to increase the sensitivity and efficiency of the real-time decision making while being able to fulfill timing constraints. The project is full of challenges, such as complex printed circuit board design, complex FPGA logic design, and translating high level ML models to FPGA synthesizable code. We are currently developing a test bench as a proof of concept and to evaluate the FPGA performance of the algorithms.

        Speakers: Alejandro Pérez Aguilera (IPARCOS-UCM), Prof. Juan Abel Barrio (IPARCOS-UCM), Dr Luis Ángel Tejedor (IPARCOS-UCM)
      • 16:46
        Nanosecond ML for calorimeter segmentation 3m

        Effective pile-up suppression, particle ID and clustering are essential for maximising the physics performance of the Phase-II Global trigger in ATLAS. To address this, we train both convolutional and DeepSets neural networks to exploit cluster topologies to accurately predict calorimeter cell labels, and benchmark performance against existing approaches. We optimise the networks for firmware deployment and obtain resource and timing estimates.

        Speakers: Alex Martynwood (UCL), Naoki Kimura (UCL), Nikos Konstantinidis (UCL), Noah Clarke Hall (University College London)
      • 16:49
        NextGen Trigger WP 2.2: Enhancing the L0 Muon Trigger: Project goals and needs 3m

        The NGT WP 2.2 aims to improve the robustness of the L0 muon trigger system and include additional trigger strategies for non-pointing signatures from decay of long-lived exotic particles implementing novel trigger strategies in firmware. Here we present the goals of the project, as well as the needs and requirements from available tools such as HLS4ML.

        Speakers: Maria Carnesale (Sapienza Universita e INFN, Roma I (IT)), Oliver Kortner (Max Planck Society (DE)), Rimsky Alejandro Rojas Caballero (University of Massachusetts (US)), Verena Ingrid Martinez Outschoorn (University of Massachusetts (US))
      • 16:52
        Nanosecond AI for anomaly detection with decision trees on FPGA 3m

        We present an interpretable implementation of the autoencoding algorithm, used as an anomaly detector, built with a forest of deep decision trees on FPGA, field programmable gate arrays. Scenarios at the Large Hadron Collider are considered for which the autoencoder is trained using the Standard Model. The design is then deployed for anomaly detection of unknown processes. The inference is made with a latency value of 30 ns at percent-level resource usage using the Xilinx Virtex UltraScale+ VU9P FPGA. The work is documented at https://arxiv.org/abs/2304.03836

        Speaker: Joerg Stelzer (University of Pittsburgh (US))
      • 16:55
        Neural Architectures and Data Processing Pipelines for Irradiation Experiments: from the Automatic Assessment of Proposals to the Monitoring of the Beam Quality 3m

        Irradiation facilities, infrastructures for assessing devices and material radiation-hardness, face a variety of challenges, from the management of the experiment-selection process to the monitoring of the beam quality they need. While addressing vastly different issues, the answers may be found in carefully engineered Machine Learning and Artificial Intelligence (AI) solutions.
        The applications of AI models in High-Energy-Physics (HEP) data analysis are well-established, in particular with neural networks and deep-learning algorithms. We suggest that recent advances in Natural-Language-Processing techniques such as transformer architectures may be used for the experiments’ proposals assessment and the development of new attention-based monitoring and anomaly detection tools used during their execution.
        We provide supporting evidence for our approach by describing 1) how we help assess HEP-related scientific proposals within the RADNEXT EU-project and 2) how we monitor and evaluate the transverse beam profile quality in real-time at the CERN IRRAD facility in the EURO-LABS EU-project.

        Speaker: Jaroslaw Szumega (CERN EP-DT-DD, Mines ParisTech (FR))
      • 17:00
        Real-time search for Dark Photons at the Upgraded LHCb experiment (Poster Upload) 3m

        This work presents a new search for soft dark photons from charm decays, made possible by the novel real-time analysis (RTA) capabilities of the upgraded LHCb detector. The challenge consists in finding a peak on top of an irreducible non-resonant background of several kHz. In LHC Run 3, LHCb can read out the entire detector in real time (at 30 MHz) and filter interesting events through a two-stage software trigger using farms of GPUs (first stage) and CPUs (second stage). ML-based classification algorithms are employed at both stages to select charm decays, identify the extremely soft electrons that dark photons decay into, and reduce the overwhelming combinatorial background. The data throughput is further reduced by writing to disk only the interesting part of each event.

        Speakers: Carlos Eduardo Cocha Toapaxi (Ruprecht Karls Universitaet Heidelberg), Martino Borsato (Universita & INFN, Milano-Bicocca (IT))
      • 17:05
        Flavour Tagging in Run 3 at the LHCb experiment (Poster Upload) 3m

        One of the main goals of the LHCb experiment is to study charge-parity violation by looking at the decays of the large variety of beauty mesons created in pp collisions at LHC. Such studies are particularly challenging in the presence of $B$$-\overline{B}$ oscillations as the $B$ meson flavour at production time might be different from the flavour at its decay time.

        Flavour Tagging algorithms exploit the correlations between the $B$ meson production flavour and features of the global event to tag the candidate as $B$ or $\overline{B}$. Together with the tagging decision, the probability of a wrong tagging decision must be provided which is estimated through the application of Machine Learning algorithms.

        The purpose of this contribution is to introduce the strategy and developments of the Flavour Tagging algorithms for Run 3.

        Speaker: Micol Olocco (Technische Universitaet Dortmund (DE))
      • 17:10
        Track Reconstruction with Graph Neural Networks on Heterogeneous Architectures (Poster Upload) 1m

        The next decade will see an order of magnitude increase in data collected by high-energy physics experiments, driven by the High-Luminosity LHC (HL-LHC). The reconstruction of charged particle trajectories (tracks) has always been a critical part of offline data processing pipelines. The complexity of HL-LHC data will however increasingly mandate track finding in all stages of an experiment's real-time processing. This paper presents a GNN-based track-finding pipeline tailored for the Run 3 LHCb experiment's vertex detector and benchmarks its physics performance and computational cost against existing classical algorithms on GPU architectures. A novelty of our work compared to existing GNN tracking pipelines is batched execution, in which the GPU evaluates the pipeline on hundreds of events in parallel. We evaluate the impact of neural-network quantisation on physics and computational performance, and comment on the outlook for GNN tracking algorithms for other parts of the LHCb track-finding pipeline.

        Speaker: Fotis Giasemis (Centre National de la Recherche Scientifique (FR))
      • 17:12
        Ultrafast Jet Classification at the HL-LHC (Poster Upload) 1m

        Three machine learning models are used to perform jet origin classification.
        These models are optimized for deployment on a field-programmable gate array device.
        In this context, we demonstrate how latency and resource consumption scale with the input size and choice of algorithm.
        Moreover, the models proposed here are designed to work on the type of data and under the foreseen conditions at the CERN LHC during its high-luminosity phase.
        Through quantization-aware training and efficient synthetization for a specific field programmable gate array, we show that $\mathcal{O}(100)$ ns inference of complex architectures such as Deep Sets and Interaction Networks is feasible at a relatively low computational resource cost.

        Speaker: Denis-Patrick Odagiu (ETH Zurich (CH))
      • 17:13
        Optimizing Multi-Model Compression for Resource-Constrained Devices (Poster Upload) 1m

        We address the challenge of compressing a sequence of models for deployment on computing- and memory-constrained devices. This task differs from single model compression, as the decision to apply compression schemes either independently or jointly across all sub-networks introduces a new degree of freedom. We evaluate the performance of pruning and quantization techniques for model compression in the context of a prototypical image restoration and object detection multi-model system. We propose an adaptation of Quantization Aware Training (QAT) and pruning techniques, where the multi-model system is fine-tuned as a single unit with an adapted loss function, rather than applying these techniques to each model individually.

        Speaker: João Luís Prado
    • 09:00 11:00
      Lectures: FPGAs 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Sioni Paris Summers (CERN)
      • 09:00
        Introduction to FPGAs and FPGA inference 45m
        Speaker: Mario Ruiz Noguera (AMD)
      • 09:45
        AMD Brevitas QAT and the FINN compiler 1h 15m
        Speaker: Mario Ruiz Noguera (AMD)
    • 11:00 11:30
      Coffee break 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 11:30 12:30
      Lectures: Heterogeneous Accelerated Compute Cluster - HACC 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Dylan Sheldon Rankin (University of Pennsylvania (US))
    • 12:30 13:45
      Lunch break 1h 15m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 13:45 16:00
      Tutorial: FPGAs 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Anna Sfyrla (Universite de Geneve (CH))
      • 13:45
        QONNX 30m
        Speakers: AMD/XILINX team, Marius Köppel (ETH Zurich (CH))
      • 14:15
        Discussion 15m
      • 14:30
        Tutorial: Working with QONNX 45m
        Speaker: Marius Köppel (ETH Zurich (CH))
      • 15:15
        Tutorial: Training a fully heterogeneously quantized NN with HGQ 45m
        Speaker: Chang Sun (California Institute of Technology (US))
    • 16:00 16:30
      Coffee break 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 16:30 18:00
      Tutorial: fwXmachina 500/1-001 - Main Auditorium

      500/1-001 - Main Auditorium

      CERN

      400
      Show room on map
      Convener: Tae Min Hong (University of Pittsburgh (US))
      • 16:30
        Introduction to fwX 30m
        Speaker: Tae Min Hong (University of Pittsburgh (US))
      • 17:00
        Tutorial 1h
        Speakers: Ben Carlson, Elangovan Yuvaraj, Isabelle Eleonore Taylor (University of Pittsburgh (US)), Joerg Stelzer (University of Pittsburgh (US)), Pavel Serhiayenka, Rajat Gupta (University of Pittsburgh (US)), Santiago Cane (University of Pittsburgh (US)), Tae Min Hong (University of Pittsburgh (US))
        • Using the egamma example we train a BDT and generate the fwX BDT description 15m

          Taking an example of electrons and photons we train an BDT and write out a fwX represenation of the BDT (demo from the Auditorium)

          Speaker: Rajat Gupta (University of Pittsburgh (US))
        • VHDL Synthesis 20m

          Taking the output fwX BDT description we generate the firmware image with Vivado (demo from remote)

          Speaker: Isabelle Eleonore Taylor (University of Pittsburgh (US))
        • FPGA application: running the testbench and checking the output signals 25m

          In a first step (3a), we run the behavioral testbench (simulation) and make the block design for the FPGA testbench (burning FPGA). In a second step (3b) we look at the waveforms (output signals) of the FPGA testbench (only video)

          Note: these two video's are in webm-format, which only Chrome supports for certain.

          Speaker: Pavel Serhiayenka (University of Pittsburgh)
    • 08:45 09:30
      Lectures: Generative AI in Research and Science 222/R-001

      222/R-001

      CERN

      200
      Show room on map
      Convener: Andreas Salzburger (CERN)
      • 08:45
        Generative AI in Research and Science 30m

        The intersection of generative AI and science is revolutionizing research and discovery across various scientific disciplines. Generative AI, capable of creating new content based on learned patterns, accelerates hypothesis generation, drug discovery, material design, and personalized medicine by simulating experiments and predicting outcomes. It enhances climate science, environmental modeling, and scientific communication through advanced data analysis and visualization. However, this integration also presents challenges, including data quality, interpretability, and ethical concerns, necessitating careful management to ensure responsible and effective use of AI in scientific research.

        We are exploring use cases and relevant examples of how GenAI has been adopted by the research community and how NVIDIA has invested in tools, libraries and models to enable open research and collaboration.

        Speaker: Geetika Gupta (NVIDIA, Director Product Management)
      • 09:15
        Discussion 15m
    • 09:30 11:15
      Tutorial: FPGAs 222/R-001

      222/R-001

      CERN

      200
      Show room on map
      Convener: Sioni Paris Summers (CERN)
    • 11:15 11:45
      Coffee break 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 11:45 12:45
      Lectures: Edge ML out of HEP 222/R-001

      222/R-001

      CERN

      200
      Show room on map
      Convener: Sioni Paris Summers (CERN)
      • 11:45
        Efficient architectures with geometric deep learning 45m
        Speaker: Hampus Otto Linander (Verses AI)
      • 12:30
        Discussion 15m
    • 12:45 14:00
      Lunch break 1h 15m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 14:00 15:00
      Lectures: Neuromorphic computing 222/R-001

      222/R-001

      CERN

      200
      Show room on map
      Convener: Dr Jean-Roch Vlimant (California Institute of Technology (US))
      • 14:00
        Neuromorphic computing 45m
        Speakers: Stanislaw Wozniak (IBM Research), Thomas Ortner (IBM Research)
      • 14:45
        Discussion 15m
    • 15:00 15:30
      Coffee break 30m 222/R-001

      222/R-001

      CERN

      200
      Show room on map
    • 15:30 17:00
      Tutorial: Neuromorphic computing 222/R-001

      222/R-001

      CERN

      200
      Show room on map
      Convener: Dr Jean-Roch Vlimant (California Institute of Technology (US))
      • 15:30
        Tutorial on neuromorphic computing 1h 30m
        Speaker: IBM Research
    • 17:00 18:00
      Lectures: Satellite 222/R-001

      222/R-001

      CERN

      200
      Show room on map
      Convener: Dr Jean-Roch Vlimant (California Institute of Technology (US))
      • 17:00
        Real time applications in earth-monitoring satellites 30m
        Speakers: Agenium Space, François Devieilleville (Agenium)
      • 17:30
        Discussion 15m