Help us make Indico better by taking this survey! Aidez-nous à améliorer Indico en répondant à ce sondage !

3rd ATLAS Machine Learning Workshop

Europe/Zurich
503/1-001 - Council Chamber (CERN)

503/1-001 - Council Chamber

CERN

162
Show room on map
Dan Guest (University of California Irvine (US)), David Rousseau (LAL-Orsay, FR)
Description

This is the third workshop on the use of machine learning (ML) across the ATLAS Collaboration. This follows the the first and second in Atlas Machine Learning workshops, and will be similar in focus.

  • The agenda is final
  • On Wednesday morning, we have a special public session with non ATLAS speakers, or ATLAS speakers talking about public material
  • The other sessions are restricted so that ATLAS work in progress can be shown
  • We are in Council Chamber, except Tuesday afternoon (starting at 3PM)  in IT Auditorium, some distance from building 40 or main building.
  • For news concerning this workshop and other ATLAS machine learning events, please make sure you are registered to atlas-comp-ml-announce@cern.ch 
  • All sessions are on Vidyo (not necessarily on the same Vidyo room, check the session-wise link )
  • Participation is free
  • Please please register (left hand column) and specify whether you'll be physically present at CERN during most of the workshop.

We'll have a workshop dinner Tuesday night at Café de l'aviation 19h30 (easily reachable with Tram, stop Blandonnet). For 50 CHF, we'll have salade mêlée, the famous steak Caravelle with unlimited french fries, one glass of wine, mineral water and coffee (special dishes, in particular vegetarian will be possible, as well as additional wine). People will pay there. As the room is somewhat limited, please register if you are reasonably sure you are coming, by clicking on the "Registration" menu on the left hand side, then "Dinner registration", registration will close at 7PM Monday, or when all available seats are taken (or if someone cancel). Please cancel, if you've finally decided to not come.

Contact: ATLAS ML conveners, Dan Guest and David Rousseau.

Registration
Registration
Participants
  • Aaron Paul O'Neill
  • Ada Farilla
  • Adam Abed Abud
  • Adam James Ruby
  • Adrian Buzatu
  • Aishik Ghosh
  • Aishik Ghosh
  • Alan Litke
  • Alex Emerman
  • Alex Zeng Wang
  • Alex Zeng Wang
  • Alexander Gavrilyuk
  • Alexander Leopold
  • Alexander Linus Sopio
  • Alexander Linus Sopio
  • Alexander Sherman
  • Amartya Rej
  • Amir Farbin
  • Amir Farbin
  • Amit Harihar Kumar
  • Andrea Coccaro
  • Andreas Hönle
  • Andreas Salzburger
  • Andrew Bell
  • Andrew Stuart Bell
  • Anil Radhakrishnan
  • Anna Zaborowska
  • Antonio Giannini
  • Atanu Pathak
  • Auriane Canesse
  • Ava Chloe Audrey Lee
  • Ava Chloe Audrey Lee
  • Benjamin Henckel
  • Benjamin Henckel
  • Benjamin Henry Hooberman
  • Beojan Stanislaus
  • Berare Gokturk
  • Bibhuti Parida
  • Binbin Dong
  • Binish Batool
  • Blake Christopher Forland
  • Blake Christopher Forland
  • Brett Mayes
  • Carlos Moreno Martinez
  • Cecile Germain
  • Cecile Germain
  • Cecilia Tosciri
  • Cecilia Tosciri
  • Charlie Donaldson
  • Chen Zhou
  • Chloe Gray
  • Christer Steinfinsbo
  • Christian Schmitt
  • Christopher Deutsch
  • Cristiano Alpigiani
  • Cristiano Sebastiani
  • Dalila Salamani
  • Dan Guest
  • Dan Guest
  • Daniel Joseph Antrim
  • Daniel Joseph Antrim
  • Daniel Lellouch
  • Daniela Katherinne Paredes Hernandez
  • David Rousseau
  • Davide Pietro Mungo
  • Deepak Kar
  • Dinos Bachas
  • Diptaparna Biswas
  • Dmitry Tsybyshev
  • Dominique Godin
  • Douglas Raymond Davis
  • Dylan Frizzell
  • Ece Akilli
  • Eilam Gross
  • Eilam Gross
  • Elena Freundlich
  • Elena Villhauer
  • Elham E Khoda
  • Elham E Khoda
  • Elisabeth Schopf
  • Elisabeth Schopf
  • Elvira Rossi
  • Emma Torro Pastor
  • Eric Drechsler
  • Fabian Becherer
  • Federico Meloni
  • Ferdinand Schenck
  • Ferdinand Schenck
  • Florencia Luciana Castillo
  • Florian Fischer
  • Francesco Maria Follega
  • Frederik Gerhard Faye
  • Frederik Gerhard Faye
  • Gabriel Palacino
  • Gabriel Palacino
  • Giacomo Artoni
  • Giles Chatham Strong
  • Gilles Louppe
  • Gilles Louppe
  • Giuliano Gustavino
  • Gregory Barbour
  • Hadar Cohen
  • Haifeng Li
  • Hasib Ahmed
  • Helena Santos
  • Heng Li
  • Hok-Chuen (Tom) Cheng
  • Hongtao Yang
  • Ioannis Karkanias
  • Iskander Ibragimov
  • Ivan Kadochnikov
  • Jacob Julian Kempster
  • Jana Schaarschmidt
  • Jay Chan
  • Jay Chan
  • Jeanette Miriam Lorenz
  • Jeanette Miriam Lorenz
  • Jennet Elizabeth Dickinson
  • Jennifer Thompson
  • Jeremy Robert Love
  • Joao Victor Da Fonseca Pinto
  • Joe Davies
  • Johannes Erdmann
  • John Meshreki
  • John Meshreki
  • Johnny Raine
  • Jonathan Shlomi
  • Jonathan Shlomi
  • Josefina Alconada
  • Joseph Earl Lambert
  • Jovan Mitrevski
  • Judita Mamuzic
  • Judita Mamuzic
  • Juerg Beringer
  • Julian Fischer
  • Julien Donini
  • Jun Gao
  • Kalliopi Iordanidou
  • Karolos Potamianos
  • Karolos Potamianos
  • keith baker
  • Knut Zoch
  • Kunlin Han
  • Kyle Beyer
  • Le Li
  • Le Li
  • Leonid Serkin
  • Ligang Xia
  • Lily Asquith
  • Louis-Guillaume Gagnon
  • Louis-Guillaume Gagnon
  • Lucas Santiago Borgna
  • Lucas Santiago Borgna
  • Lukas Fabian Ehrke
  • Lukas Fabian Ehrke
  • Manfredi Ronzani
  • Maowu Nie
  • Marcus Matthias Morgenstern
  • Maria Giovanna Foti
  • Maria Moreno Llacer
  • Mariel Pettee
  • Mariel Pettee
  • Marilena Bandieramonte
  • Marko Stamenkovic
  • Mason Proffitt
  • Mats Joakim Robert Olsson
  • Matt Zhang
  • Matteo Barbetti
  • Matthew Basso
  • Matthew Epland
  • Matthew Feickert
  • Mauro Verzetti
  • Mauro Verzetti
  • Meng-Ju Tsai
  • Michael Glenn Eggleston
  • Michael James Fenton
  • Michael Kagan
  • Michela Biglietti
  • Michele Faucci Giannelli
  • Migle Stankaityte
  • Miha Zgubic
  • Mohamed Belfkir
  • Monika Mittal
  • Muhammad Alhroob
  • Muhammad Waqas
  • Nicola Orlando
  • Nicole Michelle Hartman
  • Nicole Michelle Hartman
  • Nils Erik Krumnack
  • Noel Aaron Nottbeck
  • Nurcan Ozturk
  • Olaf Nackenhorst
  • Olaf Nackenhorst
  • Oscar Estrada Pastor
  • Ourania Sidiropoulou
  • Pablo Martinez Agullo
  • Paul Nilsson
  • Peng Huo
  • Petr Manek
  • Pierfrancesco Butti
  • Pilar Casado
  • Quentin Buat
  • Quentin Buat
  • Rachel Jordan Hyneman
  • Rachel Jordan Hyneman
  • Rafael Teixeira De Lima
  • Ricardo Woelker
  • Robert Langenberg
  • Roger Naranjo
  • Rui Zhang
  • Sabrina Amrouche
  • Samuel David Jones
  • Sanmay Ganguly
  • Sanmay Ganguly
  • Santiago Rafael Paredes Saenz
  • Sascha Caron
  • Savannah Jennifer Thais
  • Sean Joseph Gasiorowski
  • Sean Joseph Gasiorowski
  • Sebastiano Spinali
  • Serena Palazzo
  • Sergei Chekanov
  • Sergio González Fernández
  • shaojun sun
  • Simone Francescato
  • Sitong An
  • Stefan Dahl Hasselgren
  • Stefan Dahl Hasselgren
  • Stefan Gadatsch
  • Stefan Gadatsch
  • Stefano Giagu
  • Stefano Giagu
  • Steven Schramm
  • Suyog Shrestha
  • Swapneel Sundeep Mehta
  • Takuya Nobe
  • Therese Berge Sjursen
  • Thomas Paul Charman
  • Thomas Paul Charman
  • Tiago Vale
  • Tobias Golling
  • Tom Stevenson
  • Torre Wenaus
  • Trygve Buanes
  • Tyler James Burch
  • Vallary Shashikant Bhopatkar
  • Victor Estrade
  • Vince Croft
  • Vincent Pascuzzi
  • Vincent Wai Sum Wong
  • Werner Spolidoro Freund
  • Werner Spolidoro Freund
  • Yann Coadou
  • Yann Coadou
  • Yavar Taheri Yeganeh
  • Yi-Lun Chung
  • Zekeriya Uysal
  • Zhi Zheng
  • Zhi Zheng
  • Ziyu Guo
  • Wednesday 17 October
    • 09:00 13:00
      External To Atlas: Public Session 503/1-001 - Council Chamber

      503/1-001 - Council Chamber

      CERN

      162
      Show room on map
      Convener: Michael Aaron Kagan (SLAC National Accelerator Laboratory (US))
      • 09:00
        Constraining Effective Field Theories with Machine Learning 20m

        We present powerful new analysis techniques to constrain effective field theories at the LHC. By leveraging the structure of particle physics processes, we extract extra information from Monte-Carlo simulations, which can be used to train neural network models that estimate the likelihood ratio. These methods scale well to processes with many observables and theory parameters, do not require any approximations of the parton shower or detector response, and can be evaluated in microseconds. We show that they allow us to put significantly stronger bounds on dimension-six operators than existing methods, demonstrating their potential to improve the precision of the LHC legacy constraints.

        Speakers: Johann Brehmer, Kyle Cranmer, Gilles Louppe, Juan Pavez
      • 09:25
        How to train taggers on data 20m

        The machine learning methods currently used in high energy particle physics often rely on Monte Carlo simulations of signal and background. A problem with this approach is that it is not always possible to distinguish whether the machine is learning physics or simply an artefact of the simulation. In this presentation I will explain how it is possible to perform a new physics search with a tagger that has been trained entirely on background data. I will show how jets of particles produced at the LHC, are prime targets for such an approach. To this end, we use an unsupervised learning method (Adversarial Auto Encoder), trained entirely on background jets, to detect any anomalous result as a new physics signal. I will show a range of applications for this approach, and describe how to practically include it in an experimental analysis by mass-decorrelating the network output in an adversarial framework and performing a bump hunt.

        Speaker: Jennifer Thompson (ITP Heidelberg)
      • 09:50
        Systematics aware learning: a case study in High Energy Physics 20m

        Experimental science often has to cope with systematic errors that coherently bias data. We analyze this issue on the analysis of data produced by experiments of the Large Hadron Collider at CERN as a case of supervised domain adaptation. The dataset used is a representative Higgs to tau tau analysis from ATLAS and released as part of the Kaggle Higgs ML challenge. Perturbations have been introduced into this dataset to mimick systematic errors. A classifier is trained to separate the Higgs signal from the background. The goal is to reduce the sensitivity of the classifier with respect to systematics uncertainty. The figure of merit is the total uncertainty, including statistical and systematics uncertainty.
        Systematics-aware learning should create an efficient representation that is insensitive to perturbations induced by the systematic effects. Different techniques have been experimented with and will be reported (i) Data Augmentation (training on a mix of data generated by varying the nuisance parameter), (ii) Adversarial Learning (using the Pivot technique, an adversarial network is trained simultaneously to the classifier to reduce the classifier sensitivity) (iii) Tangent Propagation (regularizing the partial derivative of the classifier score with respect to the nuisance parameter).

        Speaker: Victor Estrade (LRI)
      • 10:15
        HiggsML challenge with optimized DNN 20m

        A lot of work done in advancing the performance of deep-learning approaches often takes place in the realms of image recognition - many papers use famous benchmark datasets, such as Cifar or Imagenet, to quantify the advantages their idea offers. However it is not always obvious, when reading such papers, whether the concepts presented can also be applied to problems in other domains and still offer improvements.
        One such example of another domain is the task of event classification in high-energy particle-collisions, such as those which occur at the LHC. In this presentation, a classifier trained on publicly available physics data (from the HiggsML Kaggle challenge) is used to test the domain transferability of several recent Machine-Learning concepts.
        A system utilising relatively recent concepts, such as cyclical learning-rate schedules and data-augmentation, is found to slightly outperform the winning solution of the HiggsML challenge, whilst requiring less than 10% of the training time, no further feature engineering, and less specialised hardware

        Speaker: Giles Chatham Strong (LIP Laboratorio de Instrumentacao e Fisica Experimental de Part)
      • 10:40
        Coffee Break 25m
      • 11:10
        The Lund Jet Plane 20m

        Lund diagrams, a representation of the phase space within jets, have long been used in discussing parton showers and resummations. We point out here that they can also serve as a powerful tool for experimentally characterising the radiation pattern within jets. We briefly comment on some of their analytical properties and highlight their scope for constraining Monte Carlo simulations. We then examine the use of the Lund plane for boosted electroweak boson tagging. When used as an input to deep-learning methods it yields high performance. Furthermore, much of that performance can be reproduced by using the Lund plane as an input to simpler log-likelihood type discriminators. This suggests a potential for unique insight and experimental validation of the features being used by machine-learning approaches. In the context of our discussion, we also highlight the importance of accounting for detector effects when considering the performance of machine-learning approaches.

        Speaker: Frederic Alexandre Dreyer (Oxford)
      • 11:35
        The TrackML tracking challenge : status report at half time 20m

        LHC experiments need to reconstruct the trajectory of particles from the few precise measurements in the detector. One major process is to « connect the dots », that is associate together the points left by each particle. The complexity of the process is growing exponentially with the LHC luminosity, so that new algorithms are needed. The TrackML challenge is a two phases competition to tackle the issue: 100.000 points to be associated into 10.000 tracks in less than 100 seconds. The first phase (with no speed incentive) has run on Kaggle over the summer, while the second one (with a strong speed incentive) is just starting on Codalab https://sites.google.com/site/trackmlparticle/ . I will summarize the preliminary findings and perspective.

        Speaker: Andreas Salzburger (CERN)
      • 12:00
        HEP Machine Learning on HPCs 20m

        Overview of current efforts and plans in HEP to use HPCs for Machine Learning.

        Speaker: Amir Farbin (University of Texas at Arlington (US))
      • 12:25
        DeepFlavour and Tensorflow in CMSSW 20m

        Jet classification, especially focused towards heavy-flavour jets, is of paramount importance for CMS. Modern ML techniques have been applied to this classification task resulting in a new generation of DeepLearning-based taggers that sport an improved performance. DeepFlavour, DeepAK8 and DeepDoubleB/C are the latest incarnations of these new family of classifiers that have been developed by the experiment. After the development, a significant effort has been invested into the deployment of such new classifiers in the complex CMS reconstruction software, CMSSW. In this presentation we will review the improvements in the ML modelling as well as the technical difficulties encountered in such classifiers in the real world of computing operations.

        Speaker: Mauro Verzetti (CERN)