# Brummer & Partners MathDataLab

Welcome to the home page of the Brummer & Partners MathDataLab. Here you can find information about the Lab and its activities.

## Upcoming events

### Conference on the Mathematics of Complex Data, June 13-16, 2022.

Finally, the conference will take place (after a long period of Covid19-related restrictions) and we invite you to join us in Stockholm.

The Brummer & Partners MathDataLab at the Department of Mathematics at KTH Royal Institute of Technology in Stockholm is organizing an international conference on the Mathematics for Complex Data . The aim of the conference is to bring together leading experts in the field and highlight recent advances in mathematics related to data-analysis and artificial intelligence. The conference is intended to cover a broad range of topics with emphasis on data analysis, including algebra and topology, probability and statistics, applied harmonic analysis, and optimization.

**Plenary speakers**

Anna Gilbert (University of Michigan), Kathryn Hess Bellwald (EPFL), Mauro Maggioni (Johns Hopkins Universiy), Eric Moulines (Ecole Polytechnique), Bernd Sturmfels (Berkeley/MPI Leipzig), Joel Tropp (Caltech), Sara van de Geer (ETH, Zurich)

The conference webpage can be found here .

## Past events

### Seminar with Samuel Cohen, Oxford University, Nov 29

Monday November 22, 15.15-16.15.

Zoom: kth-se.zoom.us/j/63635413752

**Title: **Arbitrage-free Neural-SDE Market models

Abstract: Modelling joint dynamics of liquid vanilla options is crucial for arbitrage-free pricing of illiquid derivatives and managing risks of option trade books. In this talk we will develop a nonparametric model for the European options book respecting underlying financial constraints and while being practically implementable. In particular, we will consider a state space for prices which are free from static (or model-independent) arbitrage and study the inference problem where a model is learnt from discrete time series data of stock and option prices. We use neural networks as function approximators for the drift and diffusion of the modelled SDE system, and impose constraints on the neural nets such that no-arbitrage conditions are preserved. In particular, we give methods to calibrate *neural SDE* models which are guaranteed to satisfy a set of linear inequalities. We validate our approach with numerical experiments using data generated from a Heston stochastic volatility model, and with observed market data.

### Seminar with Jonas Adler, DeepMind, Nov 22

Monday November 22, 15.15-16.15.

Room 3721, Lindstedtsvägen 25, 7th floor.

**Title: **Highly accurate protein structure prediction with AlphaFold

Abstract: Predicting a protein’s structure from its primary sequence has been a grand challenge in biology for the past 50 years, holding the promise to bridge the gap between the pace of genomics discovery and resulting structural characterization. In this talk, we will describe work at DeepMind to develop AlphaFold, a new deep learning-based system for structure prediction that achieves high accuracy across a wide range of targets. We demonstrated our system in the 14th biennial Critical Assessment of Protein Structure Prediction (CASP14) across a wide range of difficult targets, where the assessors judged our predictions to be at an accuracy “competitive with experiment” for approximately 2/3rds of proteins. The talk will cover both the underlying machine learning ideas and the implications for biological research as well as some promising further work.

### Mini-course: Understanding nonlinear dynamics with finite data, March 17, March 18, and March 19.

Speaker: Konstantin Mischaikow, Rutgers University.

Lecture 1. Zoom, Wednesday March 17, 15.15-16.15. kth-se.zoom.us/j/63389722507

Lecture 2. Zoom, Thursday March 18, 15.15-16.15. kth-se.zoom.us/j/63389722507

Lecture 3. Zoom, Friday March 19, 15.15-16.15. kth-se.zoom.us/j/63389722507

Abstract: This set of tutorial lectures will focus on five topics:

1. I will provide an extremely a brief description of the subject of dynamical systems and argue that as currently framed is too rich a subject in an area of data driven science and engineering.

2. I will describe a courser approach to dynamics that is based on combinatorics, order theory, and algebraic topology.

3. I will discuss computational aspects and challenges associated with this new approach.

4. I will describe how some results from classical dynamics can obtain via this approach.

5. I will present some applications.

### Zoom seminar with Martin Larsson, Carnegie Mellon University, Jan 25

Zoom, Monday January 25, 15.15-16.00.

Zoom Meeting ID: 657 1501 8037,
kth-se.zoom.us/j/65715018037

**Title: **Finance and Statistics: Trading Analogies for Sequential Learning

Abstract: The goal of sequential learning is to draw inference from data that is gathered gradually through time. This is a typical situation in many applications, including finance. A sequential inference procedure is `anytime-valid’ if the decision to stop or continue an experiment can depend on anything that has been observed so far, without compromising statistical error guarantees. A recent approach to anytime-valid inference views a test statistic as a bet against the null hypothesis. These bets are constrained to be supermartingales - hence unprofitable - under the null, but designed to be profitable under the relevant alternative hypotheses. This perspective opens the door to tools from financial mathematics. In this talk I will discuss how notions such as supermartingale measures, log-optimality, and the optional decomposition theorem shed new light on anytime-valid sequential learning. (This talk is based on joint work with Wouter Koolen (CWI), Aaditya Ramdas (CMU) and Johannes Ruf (LSE).)

### Mini-course: Sampling problems in computational statistical physics, Jan 18, Jan 19, and Jan 21.

Speaker: Tony Lelièvre, Ecole des Ponts ParisTech.

Lecture 1. Zoom, Monday January 18, 15.15-16.15. kth-se.zoom.us/j/63389722507

- Slides (pdf 4.1 MB)
- Video

Lecture 2. Zoom, Tuesday January 19, 15.15-16.15. kth-se.zoom.us/j/63389722507

- Slides (pdf 3.9 MB)
- Video

Lecture 3. Zoom, Thursday January 21, 15.15-16.15. kth-se.zoom.us/j/63389722507

- Slides (pdf 2.4 MB)
- Video

Abstract: Computational statistical physics is typically a domain where efficient sampling methods are crucial. The objective is indeed to obtain macroscopic properties of materials starting from a microscopic description at the molecular level, using ensemble averages (thermodynamic properties) or averages over paths (dynamical properties). Applications are numerous in very different scientific fields such as molecular biology, chemistry or materials science. The objective of these lectures will be, starting from some prototypical sampling problems raised in statistical physics, to introduce general purpose algorithms which are useful to sample multimodal distributions, rare events and metastable trajectories. More precisely, the first lecture will be devoted to free energy adaptive biasing techniques, and their analysis using entropy techniques or standard methods to prove the convergence of stochastic algorithms. In the second lecture, we will present rare event sampling techniques in order to sample reactive trajectories, namely the pieces of trajectory between metastable states. Finally, the third lecture will be devoted to a discussion of the link between metastable dynamics and jump Markov processes, using the notion of quasi-stationary distribution, and associated algorithms (accelerated dynamics methods).

### Mini-course: Randomized methods in linear algebra, with applications in data science, Nov 17-19

Speaker: Per-Gunnar Martinsson, University of Texas at Austin.

Lecture 1. Zoom, Tuesday November 17, 15.15-16.00.
kth-se.zoom.us/j/61670083105

Lecture 2. Zoom, Wednesday November 18, 15.15-16.00.
kth-se.zoom.us/j/61670083105

Lecture 3. Zoom, Thursday November 19, 15.15-16.00.
kth-se.zoom.us/j/61670083105

Abstract: The lectures will describe a set of recently developed randomized algorithms for accelerating matrix computations and the analysis of large data sets. A recurring theme will be the use of randomized embeddings that reduce the effective dimensionality of data sets while in certain respects preserving their geometric properties. We will describe how to use the methods in practice, and how their performance can be analyzed mathematically.

For a detailed description, see the
flyer
.

Slides will be posted here:
link to slides
.

Recordings:
Lecture 1
,
Lecture 2
,
Lecture 3

### Zoom seminar with Henrik Hult, KTH, Jun 15

Zoom, Monday June 15, 15.15-16.00.

Zoom Meeting ID: 650 4456 0258,
Zoom link

**Title: **Estimates of the proportion of SARS-CoV-2 infected individuals in Sweden

Abstract: In this talk a Bayesian SEIR model is studied to estimate the proportion of the population infected with SARS-CoV-2, the virus responsible for COVID-19. To capture heterogeneity in the population and the effect of interventions to reduce the rate of epidemic spread, the model uses a time-varying contact rate, whose logarithm has a Gaussian process prior. A Poisson point process is used to model the occurrence of deaths due to COVID-19 and the model is calibrated using data of daily death counts in combination with a snapshot of the the proportion of individuals with an active infection, performed in Stockholm in late March. The methodology is applied to regions in Sweden. The results show that the estimated proportion of the population who has been infected is around 13:5% in Stockholm, by 2020-05-15, and ranges between 2:5%-15:6% in the other investigated regions. In Stockholm where the peak of daily death counts is likely behind us, parameter uncertainty does not heavily influence the expected daily number of deaths, nor the expected cumulative number of deaths. It does, however, impact the estimated cumulative number of infected individuals. In the other regions, where random sampling of the number of active infections is not available, parameter sharing is used to improve estimates, but the parameter uncertainty remains substantial. This is joint work with Martina Favero. The report is available at arxiv.org/abs/2005.13519 .

### Study group on Machine Learning and Numerical Analysis

With this study group we wish to explore some topics in machine learning with focus on numerical analysis. The aim is to increase knowledge and familiarity with various machine learning tools. There will be three seminars during the autumn:

- October 4, 14:15-16:00, Room F11. Anders Szepessy, KTH: "Convergence for learning neural networks by the stochastic gradient method".
- December 4, 13:15-15:00, Room F11, Richard Tsai, University of Texas at Austin. Title: Time-parallelized computation of the second order wave equation aided by deep learning and comparison of point clouds by random rays.
- December 11, 13.15-15, Room F11, Ozan Öktem, KTH: Title: Bayesian inversion and Deep Learning

Welcome!

Organizers: Anna Persson & Patrick Henning

### Seminar with Nicola Quercioli, University of Bologna, Dec 2

Room F11, Monday Dec 2, 15.15-16.15

**Title: **Group equivariant non-expansive operators for data analysis and machine learning

Abstract: In this talk we will introduce a general mathematical framework for group equivariance in the machine learning context. The framework builds on a synergy between persistent homology and the theory of group actions. Our focus will be on showing the topological and metric properties of the space of group equivariant non-expansive operators (GENEOs), which are maps between function spaces associated with groups of transformations.

### Seminar with Richard Davis, Columbia University, Oct 14 (joint with Mathematical Statistics)

Room F11, Monday Oct 14, 15.15-16.15

**Title: **The Use of Shape Constraints for Modeling Time Series of Counts

Abstract: For many formulations of models for time series of counts, the specification of a family of probability mass functions relating the observation Y_{t} at time t to a state variable X_{t} must be explicitly specified. Typical choices are the Poisson and negative binomial distributions. One of the principal goals of this research is to relax this parametric framework and assume that the requisite pmf is a one-parameter exponential family in which the reference distribution is unknown but log-concave. This class of distributions includes many of the commonly used pmfs. The serial dependence in the model is governed by specifying the evolution of the conditional mean process. The particular link function used in the exponential family model depends on the specification of the reference distribution. Using this semi-parametric model formulation, we are able to extend the class of observation-driven models studied in Davis and Liu (2016). In particular, we show there exists a stationary and ergodic solution to the state-space model. In this new semi-parametric framework, we compute and maximize the likelihood function over both the parameters associated with the mean function and the reference measure subject to a concavity constraint. On top of this we can “smooth” the pmf using the Skellam distribution in order to obtain an estimated distribution defined on all the non-negative integers. In general, the smooth version has better performance than existing methods. The estimator of the mean function and the conditional distribution are shown to be consistent and perform well compared to a full parametric model specification. Further limit theory in other situations will be described. The finite sample behavior of the estimators are studied via simulation and empirical examples are provided to illustrate the methodology. This is joint work with Jing Zhang and Thibault Vatter.

### Seminar with Sid Resnick, Cornell University, Sep 16 (joint with Mathematical Statistics)

Room F11, Monday Sep 16, 15.15-16.15

**Title: **Exploring Dependence in Multivariate Heavy Tailed Data

Abstract: We review a framework for considering multivariate data that could plausibly come from a multivariate power law. The framework is flexible enough to allow for multiple (even infinite) heavy tail regimes depending on the choice of scaling and the definition of an extreme region. We use the framework to explore extremal dependence between components using both graphical and analytical means. An example using market returns is given for what we call "strong dependence" and an exploratory graph technique can highlight the most dependent components.

### Seminar with Paul Jenkins, University of Warwick, Sep 12

Room F11, Thursday Sep 12, 15.15-16.15

**Title: **Asymptotic genealogies of interacting particle systems

Abstract: Interacting particle systems are a broad class of stochastic models for phenomena in disciplines including physics, engineering, biology, and finance. A prominent example in statistics is the particle filter, which features prominently in numerical approximation schemes for observations from hidden Markov models. A particle filter proceeds by evolving a discrete-time, weighted population of particles whose empirical measure approximates the distribution of the hidden process. In this talk I discuss how to characterise the genealogy underlying this evolving particle system. More precisely, under certain conditions we can show that the genealogy converges (as the number of particles grows) to Kingman's coalescent, a stochastic tree-valued process widely studied in population genetics. This makes explicit the analogy between a particle filter and an evolving biological population.

## Workshop on Mathematics for Complex Data, June 24-26, 2019

The purpose of this workshop is to bring together researchers interested in the mathematics of complex data. There will be talks on mathematical theory and methods related to data analysis and artificial intelligence.

### Schedule

*Monday, June 24*

10.30-11.00 Registration

11.00-11.10 Welcome

11.10-11.55 Daniel Persson, Chalmers

11.55-14.00 Lunch

14.00-14.45 Thomas Schön, Uppsala

14.45-15.30 Natasa Sladoje, Uppsala

15.30-16.00 Coffee

16.00-16.45 Henrik Hult, KTH

*Tuesday, June 25*

9.15-10.00 Guo-Jhen Wu, Brown

10.00-10.15 Coffee

10.15-11.00 Konstantin Mischaikow, Rutgers

11.00-11.45 Wojciech Chacholski, KTH

11.45-14.00 Lunch

14.00-14.45 Martina Scolamiero, KTH

14.45-15.30 Annika Lang, Chalmers

15.30-16.00 Coffee

16.00-16.45 Anna Persson, KTH

Wednesday, June 26

8.30-9.15 Liam Solus, KTH

9.15-10.00 Tatyana Turova, Lund

10.00-10.15 Coffee

10.15-11.00 Joel Larsson, Warwick

11.00-11.45 Johan Jonasson, Chalmers

### Location

Room U31 , Brinellvägen 28A, KTH Campus, Valhallavägen, Stockholm.

### Talks

**Daniel Persson, Chalmers**

**Title:** Quantum deep learning

**Abstract:** Despite the overwhelming success of deep neural networks we are still at a loss for explaining exactly how deep learning works, and why it works as well as it does. What are the fundamental principles underlying deep learning? In recent years there have been various intriguing proposals for how results from physics and mathematics may shed light on these principles. In this talk I will describe some of these exciting developments, including topics such as renormalization, quantum entanglement and gauge equivariant convolutional neural networks.

**Thomas Schön, Uppsala**

**Title:** Probabilistic modelling - driven by data, guided by physics

**Abstract:** In this talk we want to show that the combined use of data-driven modelling and existing scientific knowledge can be quite rewarding. We illustrate this using concrete examples from physics, including modelling the ambient magnetic field, neutron diffraction experiments aiming to reconstruct the strain field, and computed tomographic (CT) reconstruction. These are all concrete examples where physics provide us with linear operator constraints that needs to be fulfilled (first example) or alternatively measurements constituted by line integrals (two last). The Gaussian process is one of the most widely used models within machine learning, even though it has to some extent been overshadowed by the deep neural networks in recent years. The reason for the usefulness of the Gaussian process is that it offers a probabilistic and non-parametric model of nonlinear functions. When these properties are combined with basic existing scientific knowledge of the phenomenon under study we have a useful mathematical tool capable of fusing existing knowledge with new measured data. Besides briefly introducing the Gaussian process we will also provide some insights into how we can adapt it so that it obeys linear operator constraints (including ODEs, PDEs and integrals), motivated for example by the specific examples above. Towards the end we will also (very briefly) sketch how the Gaussian process can incorporate deep neural networks to further enhance its flexibility. These developments opens up for the use of basic scientific knowledge within one of our classic machine learning models.

**Natasa Sladoje, Uppsala**

**Title:** Distance functions for robust image registration: Theory and applications

**Abstract: **Analysis of visual data - digital images - provides numerous challenges. At the same time, the amount of such data to be analysed nowadays increases rapidly and overwhelmingly in a range of application fields. This talk will focus on a few fundamental tasks in image processing which require comparison of images, our recently proposed solutions which rely on robust and efficient image distance functions, underlying mathematical tools, as well as some successful applications, primarily within biomedicine.

**Henrik Hult, KTH**

**Title:** Latent variable generative models

**Abstract:** Latent variable models use hidden variables to describe unobserved abstractions of data. I will discuss a variety of latent variable models and their applications in population genetics and machine learning. Emphasis will be on asymptotic properties related to sampling, interpolation, and goodness-of-fit.

**Guo-Jhen Wu, Brown**

**Title:** Optimal temperature selection for infinite swapping in the low temperature limit

**Abstract:** Parallel tempering, also known as replica exchange, is an algorithm used to speed up the convergence of slowly converging Markov processes (corresponding to lower temperatures for models from the physical sciences). By constructing other processes with higher temperature and allowing Metropolis type swaps between the different processes, the original process is able to explore the state space more efficiently via the swapping mechanism. It has been proven that by sending the swap rate to infinity, the sampling properties reach optimality in a certain sense. Moreover, this "infinite swapping limit" is attained by process with symmetrized dynamics, which when combined with a weighted empirical measure provide approximations to the original problem. After discussing the construction, we focus on optimizing variance with respect to selection of the temperatures. As will be discussed, there are two main contributions of variance reduction. The first one comes from a lowering of energy barriers and consequent improved communication properties. The second and less obvious source is because of the weights appearing in the weighted empirical measure. These two variance reduction mechanisms behave in opposite ways as the temperatures vary. Based on an extension of Freidlin-Wentzell theory, we are able to identify the best temperature sequence for certain models when the lowest temperature is sent to zero, i.e., when sampling is most difficult.

**Konstantin Mischaikow, Rutgers**

**Title:** Conley Theory, Dynamics, and Data

**Abstract:** In this talk I will provide a brief review of Conley theory from the perspective of it being a combinatorial and topological extension of Morse theory. The focus of the lecture will be on how Conley theory can be used to derive conclusions about the structure of continuous nonlinear dynamical systems from finite data.

**Wojciech Chacholski, KTH**

**Title:** What is persistence?

**Abstract:** How to give a machine a sense of geometry? There are two aspects of what a sense is: technical tool and ability to learn to use it. This learning ability is essential. For example we are born with technical ability to detect smells and through our lives we develop it, depending on needs and environment around us. In my talk I will describe how to use homology to give a machine a sense of geometry.

**Martina Scolamiero, KTH**

**Title:** Multi-parameter persistence and noise

**Abstract:** Topology has recently been used to study spaces arising from data, leading to important results in several fields such as neuroscience, cancer research, and material science. By using a method called multi-parameter persistence one can define new topological signatures that represent correlation patterns among different distances on a dataset. In this talk, I will describe our strategy to define such signatures for data and show that they are robust to noise and amenable to statistics. I will then focus on the computational challenges associated with the multi-parameter setting and outline directions for further research.

**Annika Lang, Chalmers**

**Title**: Deep learning and stochastic partial differential equations: a possible connection

**Abstract:** Neural networks are getting deeper and deeper. Letting the number of layers going to infinity, we can interpret them as discretizations of a time continuous problem. In this talk, I will discuss this idea and give a connection to the simulation of stochastic partial differential equations.

**Anna Persson, KTH**

**Title:** A multiscale method for Maxwell's equations

**Abstract:** In this talk I want to introduce you to so called multiscale problems and the numerical challenges we face when modeling them. Problems of multiscale type typically appear when simulating physical behavior in heterogeneous media, such as composite materials and porous structures. This results in partial differential equations with highly varying and oscillating coefficients. It is well known that classical polynomial based finite element methods fail to approximate the solution to such equations well, unless the mesh width is sufficiently small to resolve the data. This leads to issues with computational cost and available memory, which calls for new approaches and methods. In the first part of this talk I will present a multiscale method based on localized orthogonal decomposition, which uses information from the data to enrich the finite element space. In the second part of the talk I will explain how to apply this method to Maxwell's equations. These equations are of interest when simulating electromagnetic wave propagation in heterogeneous materials, such as photonic crystals and other metamaterials.

**Liam Solus, KTH**

**Title:** Discrete Geometry in Model Discovery

**Abstract:** In the today's world, where data is an abundant resource, there is a strong interest in data-driven algorithms for model discovery that are both efficient and reliable. Of particular interest are such algorithms for learning probabilistic, or even causal, DAG models. Historically, the combinatorics of graphs has played a central role in the development of DAG model learning algorithms. However, ideas from contemporary geometric and algebraic combinatorics were previously uninvolved. In this talk, we will discuss a new method for probabilistic DAG model discovery. This method arises as a simplex-type algorithm over convex polytopes known as generalized permutohedra, which are central to the field of algebraic and geometric combinatorics. We will see that, when compared with the state-of-the-art, these methods perform competitively, are provably more reliable, and even shed some restricting parametric assumptions. We will then discuss how these methods extend to give the first ever consistent greedy algorithms for learning causal DAG models, and examine their performance on real data coming from genomics and computational biology.

**Tatyana Turova, Lund**

**Title:** Random graph models in Neuroscience

**Abstract:** Nearly all contemporary analyses of empirical data on the neuronal networks appeal to the graph theory for its very relevant terminology and methodology. Neuroscience gave in turn rise to new

models in statistical physics, helping to understand better the issue of phase transitions in large networks.

We provide an overview of the graph theory results relevant to the analysis and modelling of neuronal networks and their functioning. To begin with the structure we discuss models of growing random graphs and closely related models of geometric random graphs. To address the issue of functioning of neuronal networks, we consider bootstrap percolation processes. In particular, we show that the threshold of connectivity for an auto-associative memory in a Hopfield model on a random graph

coincides with the threshold for the bootstrap percolation on the same random graph. We shall argue that this coincidence reflects the relations between the auto-associative memory mechanism and the properties of the underlying random network structure.

**Joel Larsson, Warwick**

**Title:** Biased and polarized random k-SAT

**Abstract:** Random SAT problems were first studied as a means to understand the average-case complexity of finding solutions to a boolean formula in n variables, but has since then become one of the most intensively studied intersections of combinatorics, computer science, and physics. Such formulae typically exhibit a threshold phenomenon: as more (random) constraints are added, a formula rapidly switches from being satisfiable to unsatisfiable. This switch occurs at some critical density of constraints per variable, and much work has been devoted to nailing down such critical densities for various models. In the classic random k-SAT model there is a balance between TRUE and FALSE, in that each variable is equally likely to be assigned either value. This is not necessarily true of real-world SAT instances, and we study the critical densities in two models where this balance is broken: biased random k-SAT and polarized random k-SAT. Joint work with Klas Markström

**Johan Jonasson, Chalmers**

**Title:** Mislabeling in statistical learning

**Abstract:** Mislabeling is a notorious problem in statistical learning and can have a great impact on the generalization performance of neural network classifiers. We propose a novel double regularization of the neural network training loss that combines a penalty on the complexity of the network and an optimal regularized reweighting of training observations. We show that there is a synergy between these two penalties such that our combined penalty outperforms the standard regularized neural network. Moreover, we show that the combined penalty improves the generalization classification performance even in the absence of mislabeling. We provide a mathematical proof to explain the observed synergy between the two regularizations, derived for a simple two-layer network; the combined penalty method strictly outperforms a standard regularized neural net, both in terms of finding a true structure in data space and as avoiding to erroneously identify a structure where there is none.

We demonstrate our double regularized neural network, DRNet, on synthetic data of varying levels of difficulty. We also illustrate DRNet performance on the FMNIST data set with simulated mislabeling. These demonstrations do not only show that there indeed is a synergy between the two penalties but also that the relationship is dependent on the level of contamination on the training labels. We observe that observation reweighting through penalization provides classification models that are more robust against overfitting as well as less sensitive with respect to the selection of hyper-parameters. This last finding provides strong support for DRNet as a practical off-the-shelf classifier since costly hyper-parameter tuning can be much reduced.

### Seminar with Michael Lesnick, SUNY, May 28

Room 3418, Tuesday May 28, 11.15-12.15

**Title: **Interleavings and Multi-parameter Persistent Homology

Abstract: In topological data analysis (TDA), we associate to data a diagram of topological spaces, which we then study using algebraic topology. Topologists have been studying diagrams of topological spaces for decades; mathematically, what sets TDA apart from classical work is that in TDA, we are interested primarily in "approximate relations" between diagrams of spaces and their invariants, rather than in exact relations. For example, we are typically more interested in whether two diagrams of spaces are close to one another in some suitably chosen metric than whether they are isomorphic (or weakly equivalent) on the nose. Much of my recent work has focused on *interleavings*, which have emerged as the formal language of choice in TDA for expressing such approximate relations. I've been especially interested in interleavings in the setting of multi-parameter persistent homology, where they can be used to formulate multi-parameter versions of fundamental stability and inference results in TDA. In this talk, I'll introduce interleavings and multi-parameter persistent homology, and discuss some recent results about these.

### Seminar with Stephan Zhechev, IST Austria, May 22 (joint with combinatorics seminar)

Room 3418, Wednesday May 22, 11.15-12.15

**Title: **Embeddability is undecidable outside the meta-stable range

Abstract: We will prove that the following question is algorithmically undecidable for k+3 < d < 3(k+1)/2 , and k>4, which covers essentially everything outside the meta-stable range: Given a finite simplicial complex K of dimension k and an embedding f : L -> R^d of a subcomplex L of K, can f be extended to an embedding F : K -> R^d of the whole complex? Here, we assume that the given embedding f of the subcomplex L is linear (on each simplex of L) whereas the desired embedding F of the whole complex is allowed to be piecewise-linear (i.e., linear on an arbitrarily fine subdivision of K); moreover F is not required to be part of the output.

More generally, we prove that the question of deciding embeddability, which is the special case of the question above when we set L to be empty, is also undecidable for k+3 < d < 3(k+1)/2 , and k>4.

The strategy of our proof is to construct a reduction from Hilbert’s tenth problem to both the embeddability and extension of embeddings problems. More specifically, for particular types of systems of quadratic Diophantine equations, we will show how to construct corresponding instances of the two problems we consider, so that an embedding or extension exists if and only if the original system of equations has an integer solution. This is a joint work with Uli Wagner and Marek Filakovsky.

### Study group on Topological Data Analysis and Machine Learning

The aim of this study group is to explore and discuss links between these three topics through recent research papers. We will present state of the art techniques and focus on the approach of the TDA group at KTH. This study group will take place in room 3418 at the Mathematics department with the following schedule:

### Workshop on Deep Learning and Inverse Problems, Jan 21-25, 2019

DLIP2019 is a one week workshop for researchers and practitioners working on deep learning techniques for inverse problems. The objective is to enable open discussions on both practical and theoretical aspects, and give researchers time to discuss these problems in depth.

The workshop will feature some invited talks, but we hope that most attendants will also contribute with their own knowledge.

*Invited Speakers*: Ozan Öktem (KTH), Andreas Hauptmann (UCL), Sebastian Lunz (Cambridge).

See the workshop webpage for more information.

### Seminar with Josef Teichmann, ETH Zurich, Jan 18

Room F11, Friday Jan 18, 14.15-15.15

**Title: **Machine Learning in Finance

Abstract: We show three instances of machine learning in finance: deep hedging, deep calibration and deep simulation. The first two applications are direct application of universal approximation theorems, in contrast to deep simulation where Johnson-Lindenstrauss random projection are used to obtain expressive but tractable sets of trajectories.

### Seminar with Per-Gunnar Martinsson, University of Texas, Nov 26

Room F11, Monday Oct 26, 15.15-16.15

**Title: **Fast Direct Solvers for Elliptic PDEs

Abstract: That the linear systems arising upon the discretization of elliptic PDEs can be solved very efficiently is well-known, and many successful iterative solvers with linear complexity have been constructed (multigrid, Krylov methods, etc). Interestingly, it has recently been demonstrated that it is often possible to directly compute an approximate inverse to the coefficient matrix in linear (or close to linear) time. The talk will survey some recent work in the field and will argue that direct solvers have several advantages, including improved stability and robustness, and dramatic improvements in speed in certain environments. Moreover, the direct solvers being proposed have low communication costs, and are very well suited to parallel implementations.

### Seminar with Phyllis Wan, Rotterdam University, Oct 22

Room F11, Monday Oct 22, 15.15-16.15

**Title: **Modeling social networks through linear preferential attachment

Abstract: Preferential attachment is an appealing mechanism for modeling power-law behavior of degree distributions in social networks. In this talk, we consider fitting a directed linear preferential attachment model to network data under three data scenarios: 1) When the full history of the network growth is given, MLE of the parameter vector and its asymptotic properties are derived. 2) When only a single-time snapshot of the network is available, an estimation method combining method of moments with an approximation to the likelihood is proposed. 3) When the data are believed to have come from a misspecified model or have been corrupted, a semi-parametric approach to model heavy-tailed features of the degree distributions is presented, using ideas from extreme value theory. We illustrate these estimation procedures and explore the usage of this model through simulated and real data examples. This is a joint work with Tiandong Wang (Cornell), Richard Davis (Columbia) and Sid Resnick (Cornell).

### Seminar with Jonas Peters, University of Copenhagen, Oct 15

Room F11, Monday Oct 15, 15.15-16.15

**Title: Causality and data**

Abstract: Causality enters data science in different ways. The goal of causal discovery is to learn causal structure from observational data, an important but difficult problem. Several methods rely on testing for conditional independence. We prove that, statistically, this is fundamentally harder than testing for unconditional independence; solving it requires carefully chosen assumptions on the data generating process. In many practical problems, the focus may lie on prediction, and it is not necessary to solve (full) causal discovery. It might still be beneficial, however, to apply causality related ideas. In particular, interpolating between causality and predictability enables us to infer models that yield more robust prediction with respect to changes in the test set. We illustrate this idea for ODE based systems considering artificial and real data sets. The talk does not require any prior knowledge in causal inference. It contains joint work with Stefan Bauer, Niklas Pfister, and Rajen Shah.

### Workshop on Mathematics for Complex Data, May 30-31, 2018

The purpose of this workshop is to bring together researchers interested in the mathematics of complex data. There will be talks on mathematical methods for data analysis as well as presentations of complex data in applications.

See here for details.

### Two lectures by Scott Baden, Lawrence Berkeley National Laboratory and University of California, San Diego, May 2-3, 2018

Lecture 1: Room F11, Wednesday, May 2, 11.15-12.00

Lecture 2: Room F11, Thursday, May 3, 11.15-12.00

**Title: Scalable memory machines**

Abstract: Distributed memory computers provide scalable memory and - hopefully - scalable performance. Over two lectures, I'll present the principles

and practice of applying scalable memory machines to solve scientific problems and describe my current research in addressing the challenges

entailed in highly scalable computing.

Bio: Prof. Baden received his M.S and PhD in Computer Science from UC Berkeley in 1982 and 1987. He is also Adjunct Professor in the Department of Computer Science and Engineering at UCSD, where he was a faculty member for 27 years. His research interests are in high performance and scientific computation: domain specific translation, abstraction mechanisms, run times, and irregular problems. He has taught parallel programming at both the graduate and undergraduate level at UCSD and at the PDC Summer School.

### Seminar with Jeffrey Herschel Giansiracusa

Room F11, Friday Feb 16, 9:00-10:00.

**Title: A tour of some applications of persistent homology**

Abstract: I will give an overview of persistent homology - how it is constructed and how we use it as a tool in data analysis. Originally it was popularised as a way of producing a description of the shape of a data set, but more recently it has taken on an alternative role as a component in functional data analysis pipelines where each element in a data set represents a complicated geometric object and persistent homology provides a way of comparing the topology and geometry of different elements, and potentially feeding the topology directly into statistical learning methods. I will describe how this works in some examples.

### Seminar with Caroline Uhler, MIT

Room F11, Wednesday Feb 7, 13.15-14.15

**Title: Your dreams may come true with MTP2**

Abstract: We study probability distributions that are multivariate totally positive of order two (MTP2). Such distributions appear in various applications from ferromagnetism to Brownian tree models used in phylogenetics. We first describe some of the intriguing properties of such distributions with respect to conditional independence and graphical models. In the Gaussian setting, these translate into new statements about M-matrices that may be of independent interest to algebraists. We then consider the problem of nonparametric density estimation under MTP2. This requires us to develop new results in geometric combinatorics. In particular, we introduce bimonotone subdivisions of polytopes and show that the maximum likelihood estimator under MTP2 is a piecewise linear function that induces bimonotone subdivisions. In summary, MTP2 distributions not only have broad applications for data analysis, but also leads to interesting new problems in combinatorics, geometry, and algebra.

Carolone Uhler joined the MIT faculty ni 2015 as the Henry L. and Grace Doherty assistant professor in EECS and IDSS. She is a member of the LIDS, the Center for Statistics, Machine Learning at MIT, and the ORC. She holds a PhD in statistics from UC Berkely. Her research focuses on mathematical statistics and computational biology, in particular on graphical models, causal inference and algebraic statistics, and on applications to learning gene regulatory networks and the development of geometric models for the organization of chromosomes. |

### Opening workshop on Nov 17, 2017.

For information, see

- Monday 28 January 10.00-12.00, talk by Martina Scolamiero, KTH
- Monday 4 February 10.00-12.00, talk by Steve Oudot, INRIA
- Monday 11 February 10.00-12.00, talk by Wojciech Chacholski, KTH
- Monday 18 February 10.00-12.00, talk by Oliver Gävfert, KTH.