## Choose timezone

Your profile timezone:

Indico has been updated to v3.3. See our blog post for details on this release. (OTG0146394)

string_data_2021

from -
Monday, December 13, 20214:00 PM What is string data? - Michael Douglas (Stony Brook University)What is string data?
- Michael Douglas (Stony Brook University)

4:00 PM - 4:30 PMMachine learning critically depends on high quality datasets. In a theoretical subject like string theory, we can generate datasets, but what sort of data should we generate and study? We discuss this question from several perspectives: mathematical (generating solutions), statistical (getting representative samples), and methodological (improving access to prior work).4:30 PM The world in a grain of sand - Yang-Hui He (London Institute, Royal Institution)The world in a grain of sand- Yang-Hui He (London Institute, Royal Institution)

4:30 PM - 5:00 PMWe propose a novel approach toward the vacuum degeneracy problem of the string landscape, using few-shot machine-learning, and by finding an efficient measure of similarity amongst compactification scenarios. Using a class of some one million Calabi-Yau manifolds as concrete examples, the paradigm of few-shot machine-learning and Siamese Neural Networks represents them as points in R^3. Using these methods, we can compress the search space for exceedingly rare manifolds to within one percent of the original data by training on only a few hundred data points. We also demonstrate how these methods may be applied to characterize ‘typicality’ for vacuum representatives. Joint work with Shailesh Lal and Zaid Zaz.5:00 PM ML to identify symmetries and integrability of physical systems - Sven Krippendorf (LMU Munich)ML to identify symmetries and integrability of physical systems- Sven Krippendorf (LMU Munich)

5:00 PM - 5:30 PMIn the first part I discuss methods on how to identify symmetries of a system without requiring knowledge about such symmetries. In the second part I discuss how to find a Lax pair/connection associated with integrable systems.6:00 PM Permutation invariant random matrix theory and natural language data - Sanjaye Ramgoolam (Queen Mary, University of London)Permutation invariant random matrix theory and natural language data- Sanjaye Ramgoolam (Queen Mary, University of London)

6:00 PM - 6:30 PMI give an introduction to the Linguistic Matrix Theory programme, where permutation invariant random matrix theory is developed for applications to matrix data arising in compositional distributional semantics. Techniques from distributional semantics produce ensembles of matrices and it is argued that the relevant semantic information has an invariance under permutations. The general 13-parameter permutation invariant Gaussian matrix models are described. Techniques from symmetric group representation theory and quantum field theory allow the computation of expectation values of permutation invariant matrix observables (PIMOs). This is used to give evidence for approximate Gaussianity in the ensembles of matrices in compositional distributional semantics. Statistical distributions of PIMOs are applied to natural language tasks involving synonyms, antonyms, hypernyms and hyponyms.6:30 PM Machine learning Calabi–Yau hypersurfaces - Edward Hirst (City, University of London)Machine learning Calabi–Yau hypersurfaces- Edward Hirst (City, University of London)

6:30 PM - 7:00 PMWe examine the classic datasbase of Calabi-Yau hypersurfaces in weighted P4s with tools from supervised and unsupervised machine-learning. Surprising linear behaviour arises with respect to the Calabi-Yau’s Hodge numbers, with a natural clustering of the spaces. In addition, simple supervised methods learn to identify weights which produce spaces with CY hypersurfaces, with improved performance where the Hodge numbers take extreme values.7:00 PM DiscussionDiscussion7:00 PM - 8:00 PM -
Tuesday, December 14, 20214:00 PM Deep learning and holographic QCD - Koji Hashimoto (Kyoto University)Deep learning and holographic QCD
- Koji Hashimoto (Kyoto University)

4:00 PM - 4:30 PMBulk reconstruction in AdS/CFT correspondence is a key idea revealing the mechanism of it, and various methods were proposed to solve the inverse problem. We use deep learning and identify the neural network as the emergent geometry, to reconstruct the bulk. The lattice QCD data such as chiral condensate, hadron spectra or Wilson loop is used as input data to reconstruct the emergent geometry of the bulk. The requirement that the bulk geometry is a consistent solution of an Einstein-dilaton system determines the bulk dilaton potential backwards, to complete the reconstruction program. We demonstrate the determination of the bulk system from QCD lattice/experiment data.4:30 PM The principles of deep learning theory - Daniel Roberts (MIT)The principles of deep learning theory- Daniel Roberts (MIT)

4:30 PM - 5:00 PMDeep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will encompass both understanding the statistics of initialized deep networks and determining the training dynamics of such an ensemble when learning from data. This talk is based on a book, “The Principles of Deep Learning Theory,” co-authored with Sho Yaida and based on research also in collaboration with Boris Hanin. It will be published next year by Cambridge University Press.5:00 PM Building quantum field theories out of neurons - James Halverson (Northeastern University)Building quantum field theories out of neurons- James Halverson (Northeastern University)

5:00 PM - 5:30 PMIn this talk I'll discuss aspects of using neural networks to design and define quantum field theories. As this approach is generally non-Lagrangian, it requires rethinking some things. Interactions will arise from breaking the Central Limit Theorem, i.e. from both 1/N-corrections and non-independence of neurons. Symmetries will play a crucial role, a duality will arise, and Gaussian Lorentz-invariant quantum field theories will be constructed.6:00 PM A tale of symmetry and duality in neural networks - Anindita Maiti (Northeastern University)A tale of symmetry and duality in neural networks- Anindita Maiti (Northeastern University)

6:00 PM - 6:30 PMWe use a duality between parameter space and function space to study ensembles of Neural Networks. Symmetries of the NN action can be inferred from invariance of its correlation functions, computed in parameter space. This mechanism, which we call ‘symmetry-via-duality,’ utilizes a judicious choice of architecture and parameter distribution to ensure invariant network actions, even when their forms are unknown. Symmetries of input and output layers are analogous to space-time and internal symmetries, respectively, in quantum field theory. In simple experiments we find a correlation between symmetry breaking and training accuracy.6:30 PM Renormalizing the optimal hyperparameters of a neural network - Greg Yang (Microsoft Research)Renormalizing the optimal hyperparameters of a neural network- Greg Yang (Microsoft Research)

6:30 PM - 7:00 PMHyperparameter tuning in deep learning is an expensive process, prohibitively so for neural networks (NNs) with billions of parameters that often can only be trained once. We show that, in the recently discovered Maximal Update Parametrization (µP), many optimal hyperparameters remain stable even as model size changes. Using this insight, for example, we are able to re-tune the 6.7-billion-parameter model of GPT-3 and obtain performance comparable to the 13-billionparameter model of GPT-3, effectively doubling the model size. In this context, there is a rich analogy we can make to Wilsonian effective field theory. For example, if “coupling constants” in physics correspond to “optimal hyperparameters” in deep learning and “cutoff scale” corresponds to “model size”, then we can say “µP” is a renormalizable theory of neural networks.” We explore this analogy further in the talk and leave open the question whether methods from effective field theory itself can make advances in tuning hyperparameters.7:00 PM DiscussionDiscussion7:00 PM - 8:00 PM -
Wednesday, December 15, 20214:00 PM Machine learning the Kitaev honeycomb model - Babak Haghighat (Tsinghua University)Machine learning the Kitaev honeycomb model
- Babak Haghighat (Tsinghua University)

4:00 PM - 4:30 PMIn this talk, we present recent results about the capability of restricted Boltzmann machines (RBMs) to find solutions for the Kitaev honeycomb model with periodic boundary conditions. We start with a review of non-abelian topological phases of matter and their importance for a scheme of quantum computation known as “topological quantum computation”. We then proceed to introduce the Kitaev Honeycomb model and our method for finding representations of its ground and excited states using RBMs. Furthermore, the possibility of realizing anyons in the RBM is discussed and an algorithm is given to build these anyonic excitations and braid them for possible future applications in quantum computation. Using the correspondence between topological field theories in (2 + 1)d and 2d CFTs, we propose an identification between our RBM states with the Moore-Read state and conformal blocks of the 2d Ising model.4:30 PM Towards solving CFTs with reinforcement learning - Costis Papageorgakis (Queen Mary, University of London)Towards solving CFTs with reinforcement learning- Costis Papageorgakis (Queen Mary, University of London)

4:30 PM - 5:00 PMI will introduce a novel numerical approach for solving the conformal-bootstrap equations with Reinforcement Learning. I will apply this to the case of two-dimensional CFTs, successfully identifying well-known theories like the 2D Ising model and the 2D CFT of a compact scalar, but the method can be used to study arbitrary (unitary or non-unitary) CFTs in any spacetime dimension.5:00 PM Non-perturbative renormalization for the neural network-QFT correspondence - Harold Erbin (MIT)Non-perturbative renormalization for the neural network-QFT correspondence- Harold Erbin (MIT)

5:00 PM - 5:30 PMIn a recent work, Halverson, Maiti and Stoner proposed a description of neural networks in terms of a quantum field theory (dubbed NN-QFT correspondence). The infinite-width limit is mapped to a free field theory while finite N corrections are taken into account by interactions. In this talk, after reviewing the correspondence, I will derive non-perturbative renormalization group equations. An important difference with the usual analysis is that the effective (IR) 2-point function is known, while the microscopic (UV) 2-point function is not, which requires setting the problem with care. Finally, I will discuss preliminary numerical results for translation-invariant kernels. A major result is that changing the standard deviation of the neural network weight distribution can be interpreted as a renormalization flow in the space of networks.6:00 PM Using generative adversarial networks to produce knots with specified invariants - Mark Hughes (Brigham Young University)Using generative adversarial networks to produce knots with specified invariants- Mark Hughes (Brigham Young University)

6:00 PM - 6:30 PMKnots in 3-dimensional space form an infinite dataset whose structure is not yet well understood. Recently techniques from machine learning have been applied to knots in an effort to better understand their topology, however so far these approaches have mainly involved techniques from supervised and reinforcement learning. In this talk I will outline an approach to using generative adversarial networks (GAN) to produce knots with specified invariant values. In particular, we show how to construct a GAN which takes as input information from the Jones polynomial, and outputs a knot with specified invariants. This is joint work with Amy Eubanks, Jared Slone, and Dan Ventura.6:30 PM Deep learning knot invariants and gauge theory - Arjun Kar (University of British Columbia)Deep learning knot invariants and gauge theory- Arjun Kar (University of British Columbia)

6:30 PM - 7:00 PMWe discuss correlations between the Jones polynomial and other knot invariants uncovered using deep neural networks. Some of these correlations are explainable in the sense that there are ideas in knot or gauge theory which, together with interpretable machine learning techniques, can be used to reverse-engineer the function computed by the network. After briefly reviewing a correlation between the Jones polynomial and hyperbolic volume that is explained by the volume conjecture and analytic continuation of Chern-Simons theory, we present new (as yet) unexplained correlations between the Jones polynomial, Rasmussen s-invariant, and slice genus. We speculate on a gauge theory explanation of these correlations using the fivebrane construction of Khovanov homology.7:00 PM DiscussionDiscussion7:00 PM - 8:00 PM -
Thursday, December 16, 20214:00 PM Metrics and Machine Learning - Challenger Mishra (Cambridge University)Metrics and Machine Learning
- Challenger Mishra (Cambridge University)

4:00 PM - 4:30 PM4:30 PM SU(3) holonomy and SU(3) structure metrics and stable bundles - Lara Anderson (Virginia Tech)SU(3) holonomy and SU(3) structure metrics and stable bundles- Lara Anderson (Virginia Tech)

4:30 PM - 5:00 PM5:00 PM SU(3) holonomy/structure metrics for CICYs and toric varieties - Fabian Ruehle (Northeastern University)SU(3) holonomy/structure metrics for CICYs and toric varieties- Fabian Ruehle (Northeastern University)

5:00 PM - 5:30 PMI will introduce a Tensorflow package for sampling points and computing metrics of string compactification spaces of SU(3) holonomy or SU(3) structure. We vastly extended previous work in this area, allowing the methods to be applied to any Kreuzer-Skarke (KS) Calabi-Yau or CICY. While extensions to CICYs are rather straight-forward, toric varieties require more work. I will first explain how to obtain the (non-Ricci-flat) analog of the Fubini-Study metric for KS models, and then how to sample points uniformly from these spaces using a powerful mathematical theorem.6:00 PM Calabi–Yau metrics, CFTs, and random matrices - Anthony Ashmore (University of Chicago)Calabi–Yau metrics, CFTs, and random matrices- Anthony Ashmore (University of Chicago)

6:00 PM - 6:30 PMCalabi-Yau manifolds have played a key role in both mathematics and physics, and are particularly important for deriving realistic models of particle physics from string theory. Without the explicit metrics on these spaces, we have resorted to numerical methods, and now have a variety of techniques to find approximate metrics. I will present recent work on what one can do with these numerical metrics, focusing on the “data” of the spectrum of the Laplacian. Computing this for many different points in complex structure moduli space, we will see that the spectrum displays random matrix statistics, suggesting that certain 2d SCFTs are chaotic.6:30 PM The string genome project - Gary Shiu (University of Wisconsin)The string genome project- Gary Shiu (University of Wisconsin)

6:30 PM - 7:00 PMIn this talk, I’ll discuss how genetic algorithms and reinforcement learning can be used as complementary approaches to search for optimal solutions in string theory and to discover structure of the string landscape, based on 1907.10072 and 2111.11466. I’ll also present ongoing work with Gregory Loges on breeding realistic D-brane models.7:00 PM DiscussionDiscussion7:00 PM - 8:00 PM -
Friday, December 17, 20214:00 PM Machine learning for computing in quantum field theory - Miranda Cheng (University of Amsterdam)Machine learning for computing in quantum field theory
- Miranda Cheng (University of Amsterdam)

4:00 PM - 4:30 PMMachine learning has the potential of becoming, among other things, a major computational tools in physics, making possible what was not. In this talk I focus on one specific example of using this new tool on concrete computational problems. I will summarise my recent paper (2110.02673) with de Haan, Rainone, and Bondesan, where we use a continuous flow model to help ameliorate the numerical difficulties in sampling in lattice field theories, which for instance hampers high-precision computations in LQCD.4:30 PM Hands-on Ricci-flat metrics with \texttt{cymetric} - Robin Schneider (Uppsala University)Hands-on Ricci-flat metrics with \texttt{cymetric}- Robin Schneider (Uppsala University)

4:30 PM - 5:00 PMIn this talk I'll demonstrate how to find numerical approximations of the unique Ricci-flat metric on Calabi-Yau manifolds with the new open source package \texttt{cymetric}. In a first step points and their integration weights are sampled from some Calabi-Yau manifold. In a second step a neural network will be trained to learn the exact correction to the Fubini-Study metric. A jupyter notebook will be used to visualize this process.5:00 PM On machine learning Kreuzer–Skarke Calabi–Yau manifolds - Per Berglund (University of New Hampshire)On machine learning Kreuzer–Skarke Calabi–Yau manifolds- Per Berglund (University of New Hampshire)

5:00 PM - 5:30 PM6:00 PM Exploring heterotic models with reinforcement learning and genetic algorithms - Andre Lukas (Oxford University)Exploring heterotic models with reinforcement learning and genetic algorithms- Andre Lukas (Oxford University)

6:00 PM - 6:30 PMWe present work applying reinforcement learning and genetic algorithms to string model building, specifically to heterotic Calabi-Yau models with monad bundles. Both methods are found to be highly efficient in identifying phenomenologically attractive three-family models, in cases where systematic scans are not feasible. For monads on the bi-cubic Calabi-Yau either method facilitates a complete search of the environment and leads to similar sets of previously unknown three-family models.6:30 PM BI for AI - Eva Silverstein (Stanford University)BI for AI- Eva Silverstein (Stanford University)

6:30 PM - 7:00 PMWe introduce a novel framework for optimization based on energy-conserving Hamiltonian dynamics in a strongly mixing (chaotic) regime. The prototype is a discretization of Born-Infeld dynamics, with a relativistic speed limit given by the objective (loss) function. This class of frictionless, energy-conserving optimizers proceeds unobstructed until stopping at the desired minimum, whose basin of attraction contains the parametrically dominant contribution to the phase space volume of the system. Building from mathematical and physical studies of chaotic systems such as dynamical billiards, we formulate a specific algorithm with good performance on machine learning and PDE-solving tasks. We analyze this and the effects of noise both theoretically and experimentally, performing preliminary comparisons with other optimization algorithms containing friction such as Adam and stochastic gradient descent.7:00 PM DiscussionDiscussion7:00 PM - 8:00 PM