Third HSE-Yandex autumn school on generative models

Europe/Moscow
Princeton (YSDA Moscow)

Princeton

YSDA Moscow

Moscow, Timura Frunze st, 11 bld 2
Alexey Naumov, Dmitriy Vetrov (HSE), Andrey Ustyuzhanin (Yandex School of Data Analysis (RU)), Natalia Talaikova
Description

HSE University and Yandex invite you to the joint third autumn school on generative models aimed at young undergraduate/graduate students planning to develop a research career. This intense four-day workshop will consist of 3 interdisciplinary mini-courses, a list of invited talks, and a lively discussions. Key topics of this school are:

  • Generative models basics
  • Advanced generative models 
  • Efficient sampling for generative models

Due to the pandemic, the school will run in the mixed format: in addition to a zoom meeting room, a physical room is booked to allow participants to follow the event from the YSDA campus.

The school is organised by three HSE laboratories:

  • LAMBDA (Laboratory of Methods for Big Data Analysis)
  • DeepBayes (Centre of Deep Learning and Bayesian Methods)
  • HDI Lab (International laboratory of stochastic algorithms and high-dimensional inference)

Yandex is an industry partner of the school.

Pre-requisites:

  • Statistics 101
  • machine learning 101
  • neural networks basic level
  • PyTorch

Mini-courses:

Invited talks (TBC):

  • Dmitry Vetrov (HSE)
  • Alexey Naumov (HSE)
  • Denis Derkach (HSE)
  • Sergey Shirobokov (ICL, Twitter)
  • Maxim Panov (Skoltech)
  • Mikhail Burtsev (MIPT)
  • Mikhail Lazarev (HSE)
  • Nikita Kazeev (HSE)

Registration:

Please register for the school using this form.

School materials:

Please note that most of the materials (slides) will be in English, although the primary language of the school will be Russian. All materials of the school are available via this page timetable. It contains slides of the lectures and talks. All seminar materials are available via the github repository.

This school is supported by the RSF grant N19-71-30020, "Applications of probabilistic artificial neural generative models to the development of digital twin technology for Non-linear stochastic systems."

For any questions, please feel free to contact Natalia Talaikova
    • 09:45 10:00
      Welcome 15m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Andrey Ustyuzhanin (Yandex School of Data Analysis (RU))
    • 10:00 11:20
      Generative Models Intro 1h 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Denis Derkach (National Research University Higher School of Economics (RU))
    • 11:20 11:40
      Coffee Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 11:40 13:00
      LAMBDA lab: Introduction to Generative Models. Practice Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Artem Ryzhikov
    • 13:00 14:00
      Lunch 1h
    • 14:00 15:20
      GANs Introduction 1h 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Denis Derkach (National Research University Higher School of Economics (RU))
    • 15:20 15:40
      Tea Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 15:40 17:00
      LAMBDA lab: Generative Adversarial Networks. Practice Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Artem Ryzhikov
    • 10:00 11:20
      Deep Bayes: Diffusion Generative Models-1 Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Dmitriy Vetrov (HSE)
    • 11:20 11:40
      Coffee Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 11:40 13:00
      Deep Bayes: Diffusion Generative Models-1. Practice Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Mr Victor Oganesyan (HSE University)
    • 13:00 14:00
      Lunch 1h
    • 14:00 15:20
      Deep Bayes: Diffusion Generative Models-2 Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Dmitriy Vetrov (HSE)
      • 14:00
        Generative adversarial networks 1h R608 (NRU HSE)

        R608

        NRU HSE

        Moscow, 109028, 11, Pokrovsky blvrd
        Speaker: Mr Artem Maevskiy (HSE University)
    • 15:20 15:40
      Tea Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 15:40 17:00
      Deep Bayes: Diffusion Generative Models-2. Practice Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Victor Oganesyan (HSE University)
    • 10:00 11:20
      HDI lab: Efficient sampling methods-1 Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Alexey Naumov
    • 11:20 11:40
      Coffee Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 11:40 13:00
      HDI lab: Efficient sampling methods-2 Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Alexey Naumov
    • 13:00 14:00
      Lunch 1h
    • 14:00 15:20
      HDI lab: Efficient sampling methods-3 Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Mr Sergey Samsonov (HSE)
    • 15:20 15:40
      Tea Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 15:40 17:00
      HDI lab: Efficient sampling methods-4 Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Convener: Mr Sergey Samsonov (HSE)
    • 10:00 10:30
      Monte Carlo Variational Auto-Encoders 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Panov Maxim
    • 10:30 11:00
      Survey of methods of k-means clustering with optimal transport 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2

      TBA

      Speaker: Suvorikova Alexandra
    • 11:00 11:30
      RICH GAN 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Mokhnenko Sergey
    • 11:30 12:10
      The Robustness of Deep Networks: A Geometrical Perspective 40m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Evgenii Burnaev
    • 12:10 12:30
      Coffee Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 12:30 13:00
      Resolution-robust Large Mask Inpainting with Fourier Convolutions 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Roman Suvorov, Aleksei Silvestrov
    • 13:00 13:30
      Problems with Deep Learning: new AI winter or new synthesis? 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Mikhail Burtsev
    • 13:30 14:30
      Lunch 1h
    • 14:30 15:00
      Material generation with AI 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Lazarev Mikhail
    • 15:00 15:30
      Uncertainty of Generative models 30m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Agatha Shishigina
    • 15:30 15:50
      Tea Break 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 15:50 16:10
      Black-Box Optimization with Local Generative Surrogates 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Sergey Shirobokov
    • 16:10 16:30
      Weather tuning via surrogates 20m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
      Speaker: Sergey Popov
    • 16:30 17:10
      Tackling the Challenge of Uncertainty Estimation and Robustness to Distributional Shift in Real-World applications 40m Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2

      While much research has been done on developing methods for improving robustness to distributional shift and uncertainty estimation, most of these methods were developed only for small-scale regression or image classification tasks. Limited work has examined developing standard datasets and benchmarks for assessing these approaches. Furthermore, many tasks of practical interest have different modalities, such as tabular data, audio, text, or sensor data, which offer significant challenges involving regression and discrete or continuous structured prediction. In this work, we propose the Shifts Dataset for evaluation of uncertainty estimates and robustness to distributional shift. The dataset, which has been collected from industrial sources and services, is composed of three tasks, with each corresponding to a particular data modality: tabular weather prediction, machine translation, and self-driving car (SDC) vehicle motion prediction. All of these data modalities and tasks are affected by real, `in-the-wild' distributional shifts and pose interesting challenges with respect to uncertainty estimation. We hope that this dataset will enable researchers to meaningfully evaluate the plethora of recently developed uncertainty quantification methods, assessment criteria and baselines, and accelerate the development of safe and reliable machine learning in real-world risk-critical applications.

      An additional challenge to uncertainty estimation in real world tasks is that standard approaches, such as model ensembles, are computationally expensive. Ensemble Distribution Distillation (EnDD) is an approach that allows a single model to efficiently capture both the predictive performance and uncertainty estimates of an ensemble. Although theoretically principled, this work shows that the original Dirichlet log-likelihood criterion for EnDD exhibits poor convergence when applied to large-scale tasks where the number of classes is very high. Specifically, we show that in such conditions the original criterion focuses on the distribution of the ensemble tail-class probabilities rather than the probability of the correct and closely related classes. We propose a new training objective which resolves the gradient issues of EnDD and enables its application to tasks with many classes, as we demonstrate on the ImageNet, LibriSpeech, and WMT17 En-De datasets containing 1000, 5000, and 40,000 classes, respectively.

      Speaker: Andrey Malinin
    • 17:10 17:20
      Organization: The end Princeton

      Princeton

      YSDA Moscow

      Moscow, Timura Frunze st, 11 bld 2
    • 18:00 20:30
      Evening session: School dinner Izya Grill

      Izya Grill