In order to enable an iCal export link, your account needs to have an API key created. This key enables other applications to access data from within Indico even when you are neither using nor logged into the Indico system yourself with the link provided. Once created, you can manage your key at any time by going to 'My Profile' and looking under the tab entitled 'HTTP API'. Further information about HTTP API keys can be found in the Indico documentation.
Additionally to having an API key associated with your account, exporting private event information requires the usage of a persistent signature. This enables API URLs which do not expire after a few minutes so while the setting is active, anyone in possession of the link provided can access the information. Due to this, it is extremely important that you keep these links private and for your use only. If you think someone else may have acquired access to a link using this key in the future, you must immediately create a new key pair on the 'My Profile' page under the 'HTTP API' and update the iCalendar links afterwards.
Permanent link for public information only:
Permanent link for all public and protected information:
(Universite de Geneve (CH))
Amazon Web Services AI - Deep learning at scale30m
Deep learning is driving rapid progress in fields such as computer vision, natural language processing, and speech recognition. It has become one of the most disruptive technologies and nowadays many products feature artificial intelligence. However, creating production-ready deep learning models involves many challenges: starting from generating good high-quality training datasets to distributed model training, hyperparameter tuning, large scale deployment and many more.
This talk will give an overview of how to build, train, deploy deep learning models on AWS.
(Amazon Web Services AI)
This talk will cover the current state of machine learning (ML) in neutrino experiments. In experiments like the Deep Underground Neutrino Experiment (DUNE), NuMI Off-axis νe Appearance (NOvA), the Micro Booster Neutrino Experiment (MicroBooNE), and Argon Neutrino Teststand (ArgoNeuT), deep learning (DL) approaches based around convolutional neural networks have been developed to provide highly accurate and efficient selections of neutrino interactions. Moreover, there is also other exciting ongoing work in DUNE focused on applying ML to tasks such as developing Generative Adversarial Networks (GANs) for photon simulation; or using Graph Neural Networks for track vs. shower hit-classification. In addition to the above, DUNE is also collaborating with high-tech companies to run some of their ML models on Tensor Processing Unit (TPU), and Field-programmable gate array (FPGA), specific hardware provided by Google and Micron Technology, respectively. Finally, the Tokai to Kamioka (T2K) experiment is also using a novel DL approach for the 3D reconstruction of the SuperFGD detector events.
Scaling Deep Learning to Exascale – ACM Gordon Bell Prize 201830m
In this talk I will present technical details about the ACM Gordon Bell Prize winner project at Supercomputing 2018. In this work the joint team from NERSC and NVIDIA succeeded in scaling a Deep Learning training across 27.000+ GPUs in Summit (world’s largest HPC system) and obtained a high fraction of peak performance. This research showed that realistic scientific applications could leverage mixed precision with FP16 (without loss of accuracy). A Deep Learning approach to segmentation of Atmospheric Rivers (AR) and Tropical Cyclones (TC) was applied to achieve state-of-the-art pattern detection for characterizing extreme weather. The training was performed at an impressive performance of peak (sustained) FP16 performance of 1.13 EF/s (1.0 EF/s). To achieve this Exascale breakthrough several innovations were developed both in Software and Hardware. These new technologies in software (TensorFlow, NCCL, etc) and hardware (Volta, Tensor Cores, NVLINK, etc) will be explained during the presentation.
(Universite de Geneve (CH))
Galaxy Morphology classification using CNN20m
Galaxy morphology is one of the most important parameters to understand the assembly and evolution of galaxies in the universe. The most used classification nowadays was proposed by Hubble (1926). This classification is based on the present of disks-, arms-, bulges-like structures, and is carried out mostly by eye identification. Thanks to the upcoming large telescopes and surveys, the new catalogues will contain millions of galaxies, making impossible to carry out a visual classification. Then, convolutional neural networks (CNN) start to play an important role to classify them. In this presentation I sumarise the most recent results (including ours) using machine learning techniques to classify galaxies and future projects.
DrJose Antonio Vazquez-Mata
(Instituto de Astronomia, UNAM)
Perspectives of Deep learning techniques in Lattice 1+1d Scalar Field Theory20m
We explore the perspectives of machine learning techniques in the context of quantum field theories based on our recent publication. In particular, we discuss two-dimensional complex scalar field theory at nonzero temperature and chemical potential – a theory with a nontrivial phase diagram. A neural network is successfully trained to recognize the different phases of this system and to predict the value of various observables, based on the field configurations. We analyze a broad range of chemical potentials and find that the network is robust and able to recognize patterns far away from the point where it was trained. Aside from the regressive analysis, which belongs to supervised learning, an unsupervised generative network is proposed to produce new quantum field configurations that follow a specific distribution. An implicit local constraint fulfilled by the physical configurations was found to be automatically captured by our generative model. We elaborate on potential uses of such a generative approach for sampling outside the training region.
 Phys.Rev. D (R)100 (2019) no.1, 011501
(FIAS, Goethe-University Frankfurt am Main)
Studying the parton content of the proton with deep learning models20m
Parton Distribution Functions (PDFs) model the parton content of the proton. Of the many collaborations which focus on PDF determination in the last 20 years, NNPDF was pioneer on the use of Neural Networks to model the probability of finding partons (quarks and gluons) inside the proton with a given energy and momentum.
In this work we introduce state of the art techniques to modernize the NNPDF methodology and study different models and architectures in a systematic way which allows as to assess the quality of the PDF fit: overtraining, model complexity, hyperparameter setup, optimization algorithm...
We show a fully automatized pipeline to achieve a best model setup finding good improvements in both the quality and efficiency of the fits.
DrJuan Manuel Cruz Martínez
(University of Milan)
Machine Learning in Velo LHCb monitoring and calibration in Run I and II.20m
Studies of High energy physics rely on cutting edge technology of particle detectors. These detectors are working under harsh conditions of constant radiations, and their operation must be strictly supervised to achieve maximum efficiency and quality of data taking. The LHCb's Vertex Locator is a strip silicon detector used in Run I and II of Large Hadron Collider. We study the application of machine learning algorithms for some of the monitoring and maintenance tasks. In this work, we present solutions based on data taken during Run I and II, that should inspire similar methods for new VeloPix detector in run III in 2021.
Maciej Witold Majewski
(AGH University of Science and Technology (PL))
Online Estimation of Particle Track Parameters based on Neural Networks for the Belle II Trigger System20m
The Belle II particle accelerator experiment is experiencing substantial background from outside of the interaction point. To avoid taking data representing this background, track parameters are estimated within the pipelined and dead time-free level 1 trigger system of the experiment and used to suppress such events. The estimation of a particle track’s origin with respect to the z-Axis, which is along the beamline, is performed by the neural z-Vertex trigger. This system is estimating the origin or z-Vertex using a trained multilayer perceptron, leveraging the advantages of training to current circumstances of operation. In order fulfil the requirements set by the overall trigger system it has to provide the estimation within an overall latency of 5 µs while matching a refresh rate of up to 31.75 MHz for new track estimations. The focus of this contribution is this system’s current status. For this both implementation and integration into the level 1 trigger will be presented, supported by first data taken during operation as well as figures of merit such as latency and resource consumption. In addition its upgrade plan for the near future will be presented. The center of these is a Hough based track finding approach that uses Bayes theorem for training the weighting of track candidates. Characteristics of this system’s current prototypical implementation on FPGAs as well as present plants towards integration for future operation will be presented.
(Karlsruhe Institute of Technology)
Similarity Hashing and Learning for Tracks Reconstruction20m
At the High Luminosity Large Hadron Collider (HL-LHC), many
proton-proton collisions happen during a single bunch crossing. This leads on average to tens of thousands of particles emerging from the interaction region. Two major factors impede finding charged particle trajectories from measured hits in the tracking detectors. First, deciding whether a given set of hits was produced by a common particle is an under-specified task. State-of-the-art reconstruction models usually tackle this issue via so-called track following only at a later stage after considering many hits. Second, assuming a nearly perfect hit-particle decision function, constructing possible hit combinations to their compatibility using this decision function is a combinatorial problem. Thus, the traditional approach will grow exponentially as the number of simultaneous collisions increase at the HL-LHC and pose a major computational challenge.
We propose a framework for Similarity Hashing and Learning for Track Reconstruction (SHLTR) where multiple small regions of the detector are reconstructed in parallel with minimal fake rate. We use hashing techniques to separate the detector search space into buckets. The particle purity of these buckets, i.e. how many hits from the same particle are contained, is increased using locality sensitivity in feature space where per-hit features beyond just its position are considered. The bucket size is sufficiently small to significantly reduce the complexity of track reconstruction within the buckets or regions.
A neural network selects valid combinations in the buckets and builds up full trajectories by connected components search independently of global positions of the hits and detector geometry. The whole process occurs simultaneously in the multiple regions of the detector and curved particles are found by allowing buckets to overlap. We present first results of such a track reconstruction chain including efficiency, fake estimates, and computational performances in µ=200 datasets.
(Universite de Geneve (CH))
Generative Adversarial Networks for fast simulation: generalisation and distributed training in HPC20m
Deep Learning techniques have are being studied for different applications by the HEP community: in this talk, we discuss the case of detector simulation. The need for simulated events, expected in the future for LHC experiments and their High Luminosity upgrades, is increasing dramatically and requires new fast simulation solutions. We will describe an R&D activity within CERN openlab, aimed at providing a configurable tool capable of training a neural network to reproduce the detector response and replace standard Monte Carlo simulation. This represents a generic approach in the sense that such a network could be designed and trained to simulate any kind of detector in just a small fraction of time. We will present the first application of three-dimensional convolutional Generative Adversarial Networks to the simulation of high granularity electromagnetic calorimeters. We will describe detailed validation studies comparing our results to Geant4 Monte Carlo simulation, showing, in particular, the very good agreement we obtain for high level physics quantities (such as energy shower shapes) and detailed calorimeter response (single cell response). Finally we will show how this tool can easily be generalized to describe a larger class of calorimeters, opening the way to a generic machine learning based fast simulation approach. To achieve generalization we will leverage advanced optimization algorithms (using Bayesian and/or Genetic approach) and apply state of the art data parallel strategies to distribute the training process across multiple nodes in HPC and Cloud environment. Performance of the parallelization of GAN training on HPC clusters will also be discussed in details.
Portraying Double Higgs at the Large Hadron Collider40m
We examine the discovery potential for double Higgs production at the high luminosity LHC in the final state with two b-tagged jets, two leptons and missing transverse momentum. Although this dilepton final state has been considered a difficult channel due to the large backgrounds, we argue that it is possible to obtain sizable signal significance, by adopting a deep learning framework making full use of the relevant kinematics along with the jet images from the Higgs decay. For the relevant number of signal events we obtain a sub- stantial increase in signal sensitivity over existing analyses. We discuss relative improvements at each stage and the correlations among the different input variables for the neutral network. The proposed method can be easily generalized to the semi-leptonic channel of double Higgs production, as well as to other processes with similar final states.
(Universidad Nacional Autonoma (MX))
AI Research at Atos1h
This lecture will cover the AI research challenges currently addressed by Atos research teams over different domains like Cyber Security, Predictive Maintenance of IT and Privacy by Design needs for Video Intelligence solution in the context current European regulations.
Regulating Emerging Technologies: Opportunities and Challenges for Latin America45m
“Can regulators keep up with fintech?” “Your Apps Know Where You Were Last Night, and They’re Not Keeping It Secret.” “Regulators scramble to stay ahead of self-driving cars.” “Digital health dilemma: Regulators struggle to keep pace with health care technology innovation.” Headlines like these capture a central challenge to today’s regulators.
Existing regulatory structures are often slow to adapt to changing social and economic circumstances, and are generally risk-averse. The policy cycle often takes anything from five to 20 years whereas a unicorn startup can develop into a global company in just a few months. Airbnb, for example, went from 21,000 arrivals in 2009 to 80 million in 2016. In the meantime, cities and countries are still trying to figure out how, and if, they can regulate short-term rental markets. Ride-hailing services, such as Uber, have experienced similar hyper-growth as regulations in the space struggle to follow. As new innovation and business models emerge government agencies are challenged with creating, adapting, enforcing and communicating the regulations to the public at a previously undreamed-of pace. And they must do so while working within legacy frameworks and attempting to foster innovation.
This presentation will highlight the unique regulatory challenges posed by digital-age technologies. Then it will describe the four critical questions policymakers and regulators in Latin America must address when it comes to regulating the Fourth Industrial Revolution. Finally, it will provide a set of five principles to guide the future of regulation.
Simulation and generative models in high energy physics30m
High-energy particle physic experiments rely heavily on billions of CPU hours per year for data processing purposes. The production of synthetic data through Geant4-based Monte Carlo simulation describing particle shower developments in the calorimeter is the single most compute-intensive tasks. Fast simulation techniques are already today indispensable. With the upcoming high-luminosity upgrade of the LHC experiments even larger simulated datasets are needed to support physics analyses. Current "classical" fast calorimeter simulation techniques are based on parametrizations of the calorimeter response. Generative models have the potential to learn the appropriate detector output response. The goal is to replace the slow but accurate Geant4-based Monte Carlo simulation by a fast, generic and accurate (enough) generative model. Both studies of Variational Auto-Encoders (VAEs) and Generative Adversarial Networks (GANs) will be presented.
(Universite de Geneve (CH))
Towards inverse design in chemistry: from prediction to deep generative models30m
Many of the challenges of the 21st century, from personalized healthcare to energy production and storage, share a common theme: materials are part of the solution. Groundbreaking advances are likely to come from unexplored regions of chemical space. A central challenge is, how do we design molecules and materials according to a desired functionality?
In this talk I showcase how we can apply machine learning to a variety of chemical problems centered around two main themes: 1) building data-driven models for prediction and interpretation of molecular properties and 2) generating and optimizing molecules according to properties via deep generative models.
Ultra-high energy cosmic rays (UHECRs) are the most energetic particles found in nature and originate from extragalactic sources. These particles induce extensive air showers when propagating within the Earth’s atmosphere. Cosmic-ray bservatories like the Pierre Auger Observatory measure such air showers using large arrays of surface-detector stations and luorescence telescopes. The reconstruction of event-by-event information sensitive to the cosmic-ray mass is a challenging task and so far, mainly based on the shower shape as measured by fluorescence detector observations with their duty cycle of about 15%.
Recently, significant progress has been made in multiple fields of machine learning using deep neural networks and associated techniques. Applying these new techniques to cosmic-ray physics opens up possibilities for improved reconstructions, including an event-by-event estimation of the cosmic-ray mass using the surface-detector exclusively, exploiting the 100% duty cycle.
In this contribution, we present deep learning-based algorithms applied in the context of cosmic-ray physics. Beside supervised learning for event reconstruction, we investigate methods for understanding the decisions of deep networks which are trained on physics-related tasks. Finally, we discuss strategies for reducing differences between data and physics simulations using adversarial frameworks.
In this talk I will describe ongoing efforts to shed light on still-unanswered questions in fundamental physics using cosmological observations. I will explain how we can use measurements of the Supernovae data, Baryon Acoustic Oscillations, Cosmic Microwave Background and the large-scale structure of the universe to reconstruct the detailed physics of the dark universe. Also I will address this inverse-problem reconstruction from a Bayesian and Machine Learning perspectives.
The Turing Machine is the paradigmatic case of computing machines, but there are others, such as Artificial Neural Networks, Table Computing, Relational-Indeterminate Computing and diverse forms of analogical computing, each of which based on a particular underlying intuition of the phenomenon of computing. This variety can be captured in terms of system levels, re-interpreting and generalizing Newell's hierarchy, which includes the knowledge level at the top and the symbol level immediately below it. In this re-interpretation the knowledge level consists of human knowledge and the symbol level is generalized into a new level that here is called The Mode of Computing. Natural computing performed by the brains of humans and non-human animals with a developed enough neural system should be understood in terms of a hierarchy of system levels too. By analogy from standard computing machinery there must be a system level above the neural circuitry levels and directly below the knowledge level that is named here The mode of Natural Computing. A central question for Cognition is the characterization of this mode. The Mode of Computing provides a novel perspective on the phenomena of computing, interpreting, the representational and non-representational views of cognition, and consciousness.
(Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas, UNAM)
Submitted contributions: Session 2
From GigScience to Combined Intelligence of Machine and Crowd20m
Online education captures all new frontiers. Modern platforms are scaled to transfer the necessary information to the student: illustrations of basic concepts, theoretical foundations, etc. At the same time, practical skills are difficult to convey online; they can only be acquired through personal experience. Such difficulties make it difficult to study natural sciences (physics, chemistry, biology, etc.). At the same time, there is a lack of qualification in working with advanced data analysis techniques for solving problems of signal cleaning from noise, optimization of structures, searching for protein structures and hundreds of other tasks. It is seen that, on the one hand, there is an excess of the workforce, of interest in acquiring practice, on the other - a lack of intellectual resources.
This situation leads to an exciting technological win-win project. A possible approach to resolving the contradiction is the creation of a platform to support focal research gig-research teams to work in a competitive-collaborative format. Such working groups consisting of subject matter researchers and online program students will solve both problems: scaling up practical online programs and increasing the level of data analysis technologies used in applied scientific challenges. The story about such a platform is accompanied by a description of examples of practical projects and diverse tasks: from the rules of gamification to the next generation autoML, in the solution of which you can also participate.
(Yandex School of Data Analysis (RU))
A machine learning approach for the feature extraction of pulmonary nodules20m
In recent times, computational studies have emerged as a viable alternative for complementing the efforts of experienced radiologists in disease diagnosis. Computed tomography (CT) studies are a common way of predicting the lung nodule malignancy for the early diagnosis and treatment of lung cancer in patients. Early detection of the type of nodule is the key to determining the appropriate treatment, thus increasing patient survival. Feature extraction is an important stage in classifying benign and malignant nodules in chest CT scans. However, determining the type of nodule in CT scans is a challenge in medical imaging, since CT images cannot be evaluated as an average or generic image. Hence, the study was based on the application of machine learning techniques, for the feature extraction of pulmonary nodules on a public, Lung TIME dataset, which features scans of teenage and adult patients.
(Instituto Tecnológico Superior de Misantla), CECILIA IRENE LOEZA MEJIA
(Instituto Tecnológico Superior de Misantla)
Skin Lesion Detection in Dermatological Images using Deep Learning20m
We demonstrate that it is possible to approach the skin lesion classification problem as a detection problem (multiple localization with classification), a much more complex and interesting problem with satisfactory and promising results. The image dataset used in the experiments comes from the ISIC Dermoscopic Archive, an open-access dermatology repository. In particular, the ISIC 2017 dataset, a subset of the ISIC archive, released for the annual ISIC challenge was used. We show that it is possible to adapt a high quality imaging dataset to the requirements demanded by a deep learning detection architecture such as YOLOv3 . After the training process is completed, the detector is able to identify and locate three different types of skin lesions in real-time. Finally, we describe some of the next steps we are taking to improve the performance of this proposed solution as we seek to deploy in the near future a real-time deep learning solution in a smartphone app that could be used by a medical specialist as well as the general public.
José Carlos Moreno-Tagle
The QUA³CK Machine Learning Development Process and the Laboratory for Applied Machine Learning Approaches (LAMA)20m
According to recent studies, Machine Learning has become a demanded skill in the world of engineering. Therefore, it is possible to find hundreds of online courses offering to teach these abilities. Some companies now call for ‘Machine Learning Engineers’ as a new hiring position. For many engineering students this is a very difficult situation, since taught courses are often focusing on the theoretical backgrounds of machine learning. However, in order to learn the correct use of machine learning methods, students have to apply them to real world problems. Hence, the scientists of the Institute for Information Processing Technologies (ITIV) at the Karlsruhe Institute of Technology (KIT, Germany) created the Laboratory for Applied Machine Learning Approaches (LAMA). We strive for a wide audience of engineering students, independently of their previous machine learning experience. The students go through three phases of learning. First, they get introduced to the theoretical concepts, then they will undertake guided hands-on experiments and lastly there is a creative ‘Into-the-Wild’-part. This approach was very successful and was positively received by the students.
The lab was designed around a novel machine learning process. This methodology is an adaptation of CRISP-DM, which is a unified cross-industry standard. However, it did not find general acceptance in academia. One reason might be the strong emphasis on business understanding.
We developed a simplified and easy-to-remember version of the CRISP-DM. This new machine learning process, which we named QUAAACK or QUA3CK, is an important part of the contribution of this work. It is an acronym for the individual process steps. The method commences with Question and Understanding the data. Followed by the iterative A³-process: Algorithm selection, data Adaption and parameter Adjustment. Based on these we will Conclude and compare the results and finally do the Knowledge transfer into a solution, product or publication. Figure 1 displays the process with precise description of each individual step.
![Figure 1: The LAMA teaches the iterative QUAAACK process as part of the lab.]
The laboratory is composed of two main parts: Guided assignments and the project based ‘Into-the-Wild’ phase. First, we teach students the basic concepts of frequently used machine-learning algorithms in theory and practice. We are using the widely adapted programming language ‘python’ for the guided part of the laboratory. Utilizing ‘jupyter notebooks’ we combine theoretical background documentation and its associated code in an interactive manner.
During the ‘Into-the-Wild’ part, students have the possibility to apply their learned skills to a real world engineering challenge. The aim of this creative part is to apply the QUA3CK process. Groups of two to four students can seek their own challenge or use provided datasets by the institute. In the end, all teams present their results to the other students and tutors. There was a broad variety of students’ projects in summer 2019. Examples were a music composing neural network, the detection of sitting behaviour in order to improve posture and classifying whether bees are affected by toxins.
We can conclude that self-imposed goals lead to a better intrinsic motivation and a higher identification with the project.
(Karlsruhe Institute of Technology)
Randomness Characterization through Bayesian Model Selection20m
Random number generation currently plays a fundamental role due to its several applications in probabilistic algorithms (e.g. Monte Carlo methods, stochastic gradient descent, etc.), but mainly for its importance in cryptography. The most common methods for characterizing random numbers generators (RNG) either lack formality (e.g. the battery of tests provided by the NIST) or are not generally applicable, even in principle (e.g. the characterization developed by the Algorithmic Information Theory). In this work we present a method based on Model Selection using Bayesian Inference which turns out to be both rigorous and effective in assessing the randomness of bits sequences. We are able to obtain analytic expressions for a model's likelihood and our results shows that this new method is more stringent than both the NIST's set of tests and the Borel's Normality criterion. Additionally, given that Bayesian Inference entails the generalizability feature for the selected model, our scheme transcends single sequence analysis and provides a characterization of the source acting as a RNG. (More details and and experimental case are presented in Ref. .)
Rafael Diaz Hernandez Rojas
(Sapienza University of Rome)
Digital pulse shape analysis (DPSA) is becoming an essential tool to extract relevant information from waveforms arising from different source. For instance, in the particle detector field, digital techniques are competing very favorable against the traditional analog way to extract the information contained in the pulses coming from particle detectors. Nevertheless, the extraction of the information contained in these digitized pulses requires powerful methods. One can visualize this extracting procedure as a pattern recognition problem. To approach this problem one can use different alternatives. One very popular alternative is to use an artificial neural network (ANN) as a pattern identifier. When using an ANN, it is common to introduce a regularization method in order to get rid or at least to reduce the effects of overfitting and overtraining. In addition, another option that helps to solve these problems is to use a large training dataset to train the ANN. In this paper, we make an intercomparison of the advantage of regularization methods vs large training datasets when used as methods to reduce the overtraining and overfitting effects when training an ANN.
DrJuan Jaime Vega Castro
(Instituto Nacional de Investigaciones Nucleares), Juan Jaime Vega Castro
Andreas Hartl, Head of Division on Artificial Intelligence and Data Economy, German Ministry of Economy - responsible for the implementation of the German AI strategy, and drafting a Blockchain strategy.
Ana Valcarel Orti, INRIA and Project Manager, Co-ordination of the French national plan of research in AI
Cristina Martínez Pinto, AI for Good Lab Director at CMINDS (NGO) - a future AI strategy in Mexico
Estelle Parker, Australian Embassy in Mexico - AI ethics in Australia
Ashley Casovan, Executive Director of AI Global, formerly Director of Data Architecture and Innovation at the Government of Canada - Canadian AI strategy
OECD: Panel 2: New policy ideas - how can we harness AI for societal progress?
Julien Chiaroni from General Secretariat for Investment (Office of the Prime Minister, France) - developing “trustable” AI.
Mirjana Stankovic, Vice-President, Emerging Technologies at Tambourine ventures - what can be done when policies develop more slowly than technology?
Ashley Casovan, Executive Director of AI Global, a non-profit dedicated to responsible implementation of AI will talk about responsibility and ethics with AI, as well as future certification of AI technologies.
Cédric Bourrasset, Artificial Intelligence product manager at ATOS will speak about ethical choices in videosurveillance applications
Prof. Andrea Bertolini, Dirpolis Institute and University of Pisa will speak about regulation in AI and robotics
Yan Zhenfang, General Director of Intelligent Computing Dept. of Huawei will speak about the experience of Huawei
Andrea Escobedo Lastiri, IBM, will talk about policy issues and AI ethics as practiced by IBM
DrGergely Gabor Barnafoldi
(Wigner RCP Hungarian Academy of Sciences (HU))
Leveraging AI Communities as Innovation Engines in Life Sciences45m
The generation of large-scale biomedical data is creating unprecedented opportunities for applications of AI solutions. Typically, the data producers develop initial predictions using AI, but it is very likely that the higher performing AI methods may reside with other groups. Crowdsourcing the analysis of complex and massive data has emerged as a framework to find robust methodologies in healthcare and drug discovery. When the crowdsourcing is done in the form of collaborative scientific competitions, known as Challenges, the validation of the AI methods is inherently addressed. Challenges also encourage open innovation, create collaborative communities to solve diverse and important biomedical problems, and foster the creation and dissemination of well-curated data repositories.
(IBM, Thomas J. Watson Research Center)
Upcoming Big Data and AI Challenges in Astronomy45m
With the advent of deep learning-based techniques, new approaches and novel architectures are proposed every day, improving the state of the art in almost every technical discipline, including medical imaging. Autoencoders are one of the most popular unsupervised deep learning techniques. Thanks to their unsupervised nature, autoencoders, and its several variants, such as variational autoencoders, provide high-capacity, unbiased, multi-dimensional vector representation of information, proving effective in many applications, including feature extraction, diagnosis, and image reconstruction. In this presentation, we will review some of the most recent advances of deep convolutional autoencoders, but always with a special focus on their practical applications in medical imaging, including prenatal and cardiac imaging.
(Accenture Iberia - Artificial Intelligence Group)
Huawei - Introduction to LATAM Intelligent Computing Business and CERN Case Study30m
Christian Eduardo Hernandez Garcia
(Huawei), Su Quing
Submitted contributions: Session 3
Network science to analyse insecurity perturbations that undermine Mexican tourism industry20m
The increasing availability of data due to effective and fast sharing methods offered by technological advances has catalyzed new approaches like network science to process large data sets and transcend traditional statistical tools analysis; hence we introduce the tripartite: destinations-rates-advisories network to make connections between complex non-linear interrelations that affect Mexican tourism industry. In the construction of this network the first set of nodes are 70 main Mexican tourist destinations; the second set corresponds to occupancy rates each destination had in 2017 and 2018; and the third set captures travel advisories emitted in 2017 and 2018 for some Mexican tourist destinations. Our network analysis is an attempt to extract useful information from a large amount of data, identifying occupancy rates prevalence, variation and their correlation with travel advisories; pertinent to develop deep understanding of tourism encoded interactions and to lead informed strategic decision making and operational coordination to amplify range of responses to adverse scenarios in tourism industry.
(Centro de Ciencias de la Complejidad ), Tanya Arenas
AI Applications In Oil Extraction, Review and Case of Study: Convolutional Neural Networks Applied to Seismic Image Analysis20m
Application of AI methods in the energy industry has been increasing in the last years. The introduction of AI in the different segments of the energy industry has turned out to be revolutionary. Some of these segments are considered as fundamental for the economic development of countries. This means, increasing their efficiency can highlight the difference between conservatory and vanguardist economies. The ability of AI methods to find patterns in data makes AI a strong tool in decision making. In the energy sector, some of the decisions that must be taken has shown a high level of complexity due to the quantity of features and data that needs to be considered. In such cases, the decisions that are taken by hand, have the risk not be optimum and that the economic waste of the implementation be too high. That is why methods in Deep Learning (DL) and Machine Learning (ML) are being introduced in the decision making process and energetic analysis in different parts of the world. These methods can be applied, for instance, in the prediction of the projections in the different energy markets, the analysis of geophysical data in order to maximize the natural resources and the optimization of the logistic in the energy exploitation, among others. In this work, we want to present a review of some AI methods in the oil industry, and an application example. In this study case, we analyze seismic images applying convolutional neural networks (CNN) to them in order to localize and segment salt bodies, which are of great importance while we are analyzing oil yields. We acknowledge funding from the European Union’s Horizon 2020 Programme under the ENERXICO Project, grant agreement No. 828947 and under the Mexican CONACYT-SENER-Hidrocarburos grant agreement No. B-S-69926.
Artificial Intelligence and Machine Learnig for Human Mobility in Megacities20m
Human mobility in megacities is a fundamental problem to address and one of the most pressing societal challenges nowadays. Fortunately, we have now at our disposal a vast set of data, through mobile devices and geolocalized social networks, that allow us to explore, using Data Science, the patterns of mobility of tens of millions of people on a daily basis. We present here recent result for 415 cities in 77 countries using a big data set of the social network Foursquare. Mining this data set we obtained patterns of visitation of attractive sites in megacities that work as a proxy for mobility. We explore these patterns using Artificial Intelligence and Machine Learning to classify and predict statistically mobility patterns according to socio-economic and cultural variations in hundreds of cities around the world.
Prof.Jose L. Mateos
(Instituto de Fisica, UNAM)
Machine Ethics Frameworks and Practical Implications for AI & DS20m
Talk given at SXSW 2019 in Austin, TX.
Contrary to what tech news outlets like Mashable, Techcrunch, and Wired tell us about the future of work and the purported "rise of the machines", we're seeing that Machine Learning developments are still far away from general AI, and hence far away from becoming real problem solvers. However, it is precisely because of this early stage of development that engineers and mathematicians need to talk about the real, down-to-earth implications of imbuing their algorithms and models with ethical considerations. We will discuss these, specifically for supervised, unsupervised and reinforcement learning, and will share some flops and bloopers from both sides of the border. We will then finish this talk with 3 frameworks for ethical modelling.
(Sociedad.Mexicana de Ciencia de Datos)
Arhuaco: Deep Learning and Isolation Based Intrusion Detection in High Energy Physics20m
High Energy Physics utilizes powerful distributed computational networks called grids, to process and analyze scientific data. Monitoring the security of these networks is a challenging task. Arbitrary and not-trusted applications can be executed inside the grid worker nodes by the scientists. Innovative methods and tools are required to reduce the risk associated with the
execution of users' software (also called jobs), to identify cyber-security incidents and to perform autonomous responses. The isolation and monitoring of job payload activity are necessary in order to protect the computational infrastructure and to find evidence of malicious behavior. We describe a security architecture that integrates Linux containers to safely execute grid jobs with behavior monitoring powered by deep learning methods for the analysis of real-time data to detect and prevent intrusions. A generative method with recurrent neural networks is utilized to improve the detection performance. We describe how these methods aim to increase the security of computational grids, improving existing solutions. We present Arhuaco, a proof-of-concept implementation and provide an evaluation for the ALICE collaboration grid at CERN.
Andres Gomez Ramirez
(Johann-Wolfgang-Goethe Univ. (DE))
(Universite de Geneve (CH))
Large-Scale Scientific endeavours: the production and dissemination of advance computer science's knowledge20m
Perform big data analysis and visualisation on your own computer? Yes, you can! Commodity computers are now very powerful in comparison to only a few years ago. On top of that, the performance of today's software and data development techniques facilitates complex computation with fewer resources. Cloud computing is not always the solution, and reliability or even privacy is regularly a concern. While the Infrastructure as a Service (IaaS) and Software as a Service (SaaS) philosophies are a key part of current scientific endeavours, there is a misleading feeling that we need to have remote computers to do any kind of data analysis. One of the aims of the ATLAS Open Data project is to provide resources — data, software and documents — that can be stored and executed in computers with minimal or non-internet access, and in as many different operating systems as possible. This approach is viewed as complementary to the IaaS/SaaS approach, where local university, students and trainers' resources can be used in an effective and reproducible way — making the HEP and Computer Sciences fields accessible to more people. We present the latest developments in the production and use of local Virtual Machines and Docker Containers for the development of physics data analysis. We also discuss example software and Jupyter notebooks, which are in constant development for use in classrooms, and students’ and teachers' computers around the world. Finally, we showcase how those tools are developed and in use by international efforts (IPPOG, ICTP-PWF and CEVALE2VE.org) in Latin America and Europe to bring knowledge in countries like Mexico, Venezuela, Colombia, Peru, Ecuador, Argentina and Uruguay with support of the ICTP, CERN and many universities in the region.
Machine Learning-Based System for the Availability and Reliability Assessment and Management of Critical Infrastructures20m
A critical infrastructure is a complex interconnected system of systems providing basic and essential services to support the operation of particle accelerators but also industries and households for which they must guarantee high reliability of critical functions.
Model-based approaches are usually adopted to provide an early identiﬁcation of failures and to reveal hidden dependencies among subsystems. System models are complex and require constant updating to be reactive to system changes and real operating conditions, wear and aging. The interconnections between the different systems and the functional dependencies between their components are in many cases modified at both physical and functional levels while their degraded performances impact the overall system availability and reliability.
A novel approach is proposed which combines model-based and Big Data analytics by machine learning techniques to extract descriptive and predictive models directly from data. The objective is to foresee and react in time to failures to reduce downtimes as well as to optimize maintenance and operation costs.
The Computer-Aided System for critical infrastructure Operation (CASO) is designed to significantly and efficiently enhance the quality, safety, reliability and availability of critical infrastructures.
We report on the design of CASO, its implementation and on the preliminary results inferred on historical and live stream data recorded from CERN’s technical infrastructure. Proposal for the full deployment and expected long-term capabilities will also be discussed.
An assessment of an ConvNet Applied to Classify Animals in the Wild20m
The objective of this work is to show an analysis of the performance of a very well known Convolutional Neural Network applied to the classification of animals in the wild.
The interesting aspect of this application is that the set of images used have diverse characteristics from the training set of the ConvNet.
The analysis goes beyond the typical display of error rates, precision and recall, by being focused on describing situations where the behaviour of the ConvNet is more erratic than expected. For instance, whereas in some cases a misclassification can be explained by a lack of visual information or by the presence of confusing information, we show errors in situations where a correct classification is easily made by a human being.
The aim is to contribute to the discussion of whether current Artificial Neural Networks technologies are suitable to be used in applications with very open and noisy domains.
MrJosé Luis Gordillo
(Center for Complexity Sciences, UNAM), José Luis Gordillo
Social events: Conference banquet
The dinner will be held on Thursday, October 24, at the Antigua Hacienda de Tlalpan (http://www.aht.mx/). We will provide transport for the non-local participants from the meeting venue to the restaurant and from the restaurant to the hotels.
The busses will leave the conference venue at 19:00.
In this talk it is presented the semantic-reasoning module of VIRBOT, our proposed architecture for service robots.
We show that by combining symbolic AI with digital-signal processing techniques this module achieves competitive performance.
Our system translates a voice command into an unambiguous representation that helps an inference engine, built around an expert system, to perform action and motion planning.
First, in the natural-language interpretation process, the system generates two outputs: (1) conceptual dependence,
expressing the linguistic meaning of the statement, and (2) verbal confirmation, a paraphrase in natural language that is
repeated to the user to confirm that the command has been correctly understood.
Then, a conceptual-dependency interpreter extracts semantic role structures from the input sentence and looks for such
structures in a set of known interpretation patterns.
We evaluate this approach in a series of skill-specific semantic-reasoning experiments.
Finally, we demonstrate our system in the general-purpose service robot test of the RoboCup-at-Home international competition,where incomplete information is given to a robot and the robot must recognize and request the missing information, and we compare our results with a series of baselines from the competition where our proposal performed best.
Atos Quantum Computing: Future Proofing What Lies Beyond Super-Computing45m
Plenary: Plenaries 5
Machine learning in accelerator physics: applications at the CERN Large Hadron Collider30m
With the advent of machine learning a few decades ago, Science and Engineering have had new powerful tools at their disposal. Particularly in the domain of particle physics, machine learning techniques have become an essential part in the analysis of data from particle collisions. Accelerator physics, however, only recently discovered the possibilities of using these tools to improve its analysis. In different laboratories worldwide, several activities are being carried out, typically in view of providing new insights to beam dynamics in circular accelerators. This is, for instance, the case for the CERN Large Hadron Collider, where since a few years exploratory studies are being carried out, covering a broad range of topics. These include the optimisation of the collimation system, the anomaly detection of beam position monitors, analysis of optimal correction tools for linear optics, lifetime and performance optimisation, and detection of hidden correlations in the huge data set of beam dynamics observables collected during the LHC Run 2. Furthermore, very recently, machine learning techniques are being scrutinised for the advanced analysis of numerical simulations data, in view of improving our models of dynamic aperture evolution.
Frederik Van Der Veken
(University of Malta (MT))
(Universidad Nacional Autonoma (MX)), Lukas Nellen
(Universidad Nacional Autonoma (MX))
Organization: Machine learning tutorial
(Universidad Nacional Autonoma (MX))
Neural Network optimisation3h
Modern neural network architecture reflects the complexity of the problem. So those may become quite complex and computationally heavy. Usually, there are plenty of different meta-parameters to tune: number of layers, activation function, number of neurons per layer, drop-out rate, etc. There many different methods and tools that aimed at tuning those parameters for various reasons - accuracy, memory footprint or inference rate. This mini-course will cover the basics approaches for neural networks optimizing including hyperparameter optimization, network architecture search and Bayesian Neural Network perspective. Practical hands-on sessions will follow the theoretical introduction.
(Yandex School of Data Analysis (RU))