3rd ATLAS Machine Learning Workshop

Europe/Zurich
503/1-001 - Council Chamber (CERN)

503/1-001 - Council Chamber

CERN

162
Show room on map
Dan Guest (University of California Irvine (US)), David Rousseau (LAL-Orsay, FR)
Description

This is the third workshop on the use of machine learning (ML) across the ATLAS Collaboration. This follows the the first and second in Atlas Machine Learning workshops, and will be similar in focus.

  • The agenda is final
  • On Wednesday morning, we have a special public session with non ATLAS speakers, or ATLAS speakers talking about public material
  • The other sessions are restricted so that ATLAS work in progress can be shown
  • We are in Council Chamber, except Tuesday afternoon (starting at 3PM)  in IT Auditorium, some distance from building 40 or main building.
  • For news concerning this workshop and other ATLAS machine learning events, please make sure you are registered to atlas-comp-ml-announce@cern.ch 
  • All sessions are on Vidyo (not necessarily on the same Vidyo room, check the session-wise link )
  • Participation is free
  • Please please register (left hand column) and specify whether you'll be physically present at CERN during most of the workshop.

We'll have a workshop dinner Tuesday night at Café de l'aviation 19h30 (easily reachable with Tram, stop Blandonnet). For 50 CHF, we'll have salade mêlée, the famous steak Caravelle with unlimited french fries, one glass of wine, mineral water and coffee (special dishes, in particular vegetarian will be possible, as well as additional wine). People will pay there. As the room is somewhat limited, please register if you are reasonably sure you are coming, by clicking on the "Registration" menu on the left hand side, then "Dinner registration", registration will close at 7PM Monday, or when all available seats are taken (or if someone cancel). Please cancel, if you've finally decided to not come.

Contact: ATLAS ML conveners, Dan Guest and David Rousseau.

Registration
Registration
Participants
  • Wednesday 17 October
    • 09:00 13:00
      External To Atlas: Public Session 503/1-001 - Council Chamber

      503/1-001 - Council Chamber

      CERN

      162
      Show room on map
      Convener: Michael Aaron Kagan (SLAC National Accelerator Laboratory (US))
      • 09:00
        Constraining Effective Field Theories with Machine Learning 20m

        We present powerful new analysis techniques to constrain effective field theories at the LHC. By leveraging the structure of particle physics processes, we extract extra information from Monte-Carlo simulations, which can be used to train neural network models that estimate the likelihood ratio. These methods scale well to processes with many observables and theory parameters, do not require any approximations of the parton shower or detector response, and can be evaluated in microseconds. We show that they allow us to put significantly stronger bounds on dimension-six operators than existing methods, demonstrating their potential to improve the precision of the LHC legacy constraints.

        Speakers: Johann Brehmer, Kyle Cranmer, Gilles Louppe, Juan Pavez
      • 09:25
        How to train taggers on data 20m

        The machine learning methods currently used in high energy particle physics often rely on Monte Carlo simulations of signal and background. A problem with this approach is that it is not always possible to distinguish whether the machine is learning physics or simply an artefact of the simulation. In this presentation I will explain how it is possible to perform a new physics search with a tagger that has been trained entirely on background data. I will show how jets of particles produced at the LHC, are prime targets for such an approach. To this end, we use an unsupervised learning method (Adversarial Auto Encoder), trained entirely on background jets, to detect any anomalous result as a new physics signal. I will show a range of applications for this approach, and describe how to practically include it in an experimental analysis by mass-decorrelating the network output in an adversarial framework and performing a bump hunt.

        Speaker: Jennifer Thompson (ITP Heidelberg)
      • 09:50
        Systematics aware learning: a case study in High Energy Physics 20m

        Experimental science often has to cope with systematic errors that coherently bias data. We analyze this issue on the analysis of data produced by experiments of the Large Hadron Collider at CERN as a case of supervised domain adaptation. The dataset used is a representative Higgs to tau tau analysis from ATLAS and released as part of the Kaggle Higgs ML challenge. Perturbations have been introduced into this dataset to mimick systematic errors. A classifier is trained to separate the Higgs signal from the background. The goal is to reduce the sensitivity of the classifier with respect to systematics uncertainty. The figure of merit is the total uncertainty, including statistical and systematics uncertainty.
        Systematics-aware learning should create an efficient representation that is insensitive to perturbations induced by the systematic effects. Different techniques have been experimented with and will be reported (i) Data Augmentation (training on a mix of data generated by varying the nuisance parameter), (ii) Adversarial Learning (using the Pivot technique, an adversarial network is trained simultaneously to the classifier to reduce the classifier sensitivity) (iii) Tangent Propagation (regularizing the partial derivative of the classifier score with respect to the nuisance parameter).

        Speaker: Victor Estrade (LRI)
      • 10:15
        HiggsML challenge with optimized DNN 20m

        A lot of work done in advancing the performance of deep-learning approaches often takes place in the realms of image recognition - many papers use famous benchmark datasets, such as Cifar or Imagenet, to quantify the advantages their idea offers. However it is not always obvious, when reading such papers, whether the concepts presented can also be applied to problems in other domains and still offer improvements.
        One such example of another domain is the task of event classification in high-energy particle-collisions, such as those which occur at the LHC. In this presentation, a classifier trained on publicly available physics data (from the HiggsML Kaggle challenge) is used to test the domain transferability of several recent Machine-Learning concepts.
        A system utilising relatively recent concepts, such as cyclical learning-rate schedules and data-augmentation, is found to slightly outperform the winning solution of the HiggsML challenge, whilst requiring less than 10% of the training time, no further feature engineering, and less specialised hardware

        Speaker: Giles Chatham Strong (LIP Laboratorio de Instrumentacao e Fisica Experimental de Part)
      • 10:40
        Coffee Break 25m
      • 11:10
        The Lund Jet Plane 20m

        Lund diagrams, a representation of the phase space within jets, have long been used in discussing parton showers and resummations. We point out here that they can also serve as a powerful tool for experimentally characterising the radiation pattern within jets. We briefly comment on some of their analytical properties and highlight their scope for constraining Monte Carlo simulations. We then examine the use of the Lund plane for boosted electroweak boson tagging. When used as an input to deep-learning methods it yields high performance. Furthermore, much of that performance can be reproduced by using the Lund plane as an input to simpler log-likelihood type discriminators. This suggests a potential for unique insight and experimental validation of the features being used by machine-learning approaches. In the context of our discussion, we also highlight the importance of accounting for detector effects when considering the performance of machine-learning approaches.

        Speaker: Frederic Alexandre Dreyer (Oxford)
      • 11:35
        The TrackML tracking challenge : status report at half time 20m

        LHC experiments need to reconstruct the trajectory of particles from the few precise measurements in the detector. One major process is to « connect the dots », that is associate together the points left by each particle. The complexity of the process is growing exponentially with the LHC luminosity, so that new algorithms are needed. The TrackML challenge is a two phases competition to tackle the issue: 100.000 points to be associated into 10.000 tracks in less than 100 seconds. The first phase (with no speed incentive) has run on Kaggle over the summer, while the second one (with a strong speed incentive) is just starting on Codalab https://sites.google.com/site/trackmlparticle/ . I will summarize the preliminary findings and perspective.

        Speaker: Andreas Salzburger (CERN)
      • 12:00
        HEP Machine Learning on HPCs 20m

        Overview of current efforts and plans in HEP to use HPCs for Machine Learning.

        Speaker: Amir Farbin (University of Texas at Arlington (US))
      • 12:25
        DeepFlavour and Tensorflow in CMSSW 20m

        Jet classification, especially focused towards heavy-flavour jets, is of paramount importance for CMS. Modern ML techniques have been applied to this classification task resulting in a new generation of DeepLearning-based taggers that sport an improved performance. DeepFlavour, DeepAK8 and DeepDoubleB/C are the latest incarnations of these new family of classifiers that have been developed by the experiment. After the development, a significant effort has been invested into the deployment of such new classifiers in the complex CMS reconstruction software, CMSSW. In this presentation we will review the improvements in the ML modelling as well as the technical difficulties encountered in such classifiers in the real world of computing operations.

        Speaker: Mauro Verzetti (CERN)