BEGIN:VCALENDAR
VERSION:2.0
PRODID:researchseminars.org
CALSCALE:GREGORIAN
X-WR-CALNAME:researchseminars.org
BEGIN:VEVENT
SUMMARY:Cláudia Soares (Instituto Superior Técnico and ISR)
DTSTART:20200514T163000Z
DTEND:20200514T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/1
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 1/">The learning machine and beyond: a tour for the curious</a>\nby Cláud
 ia Soares (Instituto Superior Técnico and ISR) as part of Mathematics\, P
 hysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThis talk will d
 raw a few perspectives on the broad topic of Machine Learning\, with non-s
 pecialists in mind. We will go through major subfields like supervised\, u
 nsupervised\, or active learning\, never forgetting the emergent reinforce
 ment learning. We will cover a few different trends over recent years\, li
 ke the mathematically inclined Support Vector Machine\, or the empirical D
 eep Learning.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/1/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Afonso Bandeira (ETH Zurich)
DTSTART:20200604T163000Z
DTEND:20200604T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/2
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 2/">Computation\, statistics\, and optimization of random functions</a>\nb
 y Afonso Bandeira (ETH Zurich) as part of Mathematics\, Physics and Machin
 e Learning (IST\, Lisbon)\n\n\nAbstract\nWhen faced with a data analysis\,
  learning\, or statistical inference problem\, the amount and quality of d
 ata available fundamentally determines whether such tasks can be performed
  with certain levels of accuracy. Indeed\, many theoretical disciplines st
 udy limits of such tasks by investigating whether a dataset effectively co
 ntains the information of interest. With the growing size of datasets howe
 ver\, it is crucial not only that the underlying statistical task is possi
 ble\, but also that is doable by means of efficient algorithms. In this ta
 lk we will discuss methods aiming to establish limits of when statistical 
 tasks are possible with computationally efficient methods or when there is
  a fundamental Statistical-to-Computational gap in which an inference task
  is statistically possible but inherently computationally hard.\n\nThis is
  intimately related to understanding the geometry of random functions\, wi
 th connections to statistical physics\, study of spin glasses\, random geo
 metry\; and in an important example\, algebraic invariant theory.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/2/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Marcelo Pereyra (Heriot-Watt University)
DTSTART:20200611T163000Z
DTEND:20200611T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/3
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 3/">Efficient Bayesian computation by proximal Markov chain Monte Carlo: w
 hen Langevin meets Moreau</a>\nby Marcelo Pereyra (Heriot-Watt University)
  as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\
 nAbstract\nThis talk summarises some new developments in Bayesian statisti
 cal methodology for performing inference in high-dimensional inverse probl
 ems with an underlying convex geometry. We pay particular attention to pro
 blems related to imaging sciences and to new stochastic computation method
 s that tightly combine proximal convex optimisation and Markov chain Monte
  Carlo sampling techniques. The new computation methods are illustrated wi
 th a range of imaging experiments\, where they are used to perform uncerta
 inty quantification analyses\, automatically adjust regularisation paramet
 ers\, and objectively compare alternative models in the absence of ground 
 truth.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/3/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Csaba Szepesvári (University of Alberta and DeepMind)
DTSTART:20200625T163000Z
DTEND:20200625T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/5
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 5/">Confident Off-Policy Evaluation and Selection through Self-Normalized 
 Importance Weighting</a>\nby Csaba Szepesvári (University of Alberta and 
 DeepMind) as part of Mathematics\, Physics and Machine Learning (IST\, Lis
 bon)\n\n\nAbstract\nOff-policy evaluation is the problem of predicting the
  value of a policy given some batch of data. In the language of statistics
 \, this is also called counterfactual estimation. Batch policy optimizatio
 n refers to the problem of finding a good policy\, again\, given some logg
 ed data.\nIn this talk\, I will consider the case of contextual bandits\, 
 give a brief (and incomplete) review of the approaches proposed in the lit
 erature and explain why this problem is difficult. Then\, I will describe 
 a new approach based on self-normalized importance weighting. In this appr
 oach\, a semi-empirical Efron-Stein concentration inequality is combined w
 ith Harris' inequality to arrive at non-vacuous high-probability value low
 er bounds\, which can then be used in a policy selection phase. On a numbe
 r of synthetic and real datasets this new approach is found to be signific
 antly superior than its main competitors\, both in terms of tightness of t
 he confidence intervals and the quality of the policies chosen. \n\nThe ta
 lk is based on joint work with Ilja Kuzborskij\, Claire Vernade and Andras
  Gyorgy.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/5/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Kyle Cranmer (NYU)
DTSTART:20200702T163000Z
DTEND:20200702T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/6
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 6/">On the Interplay between Physics and Deep Learning.</a>\nby Kyle Cranm
 er (NYU) as part of Mathematics\, Physics and Machine Learning (IST\, Lisb
 on)\n\n\nAbstract\nThe interplay between physics and deep learning is typi
 cally divided into two themes.\nThe first is “physics for deep learning
 ”\, where techniques from physics are brought to bear on understanding d
 ynamics of learning. The second is “deep learning for physics\,” which
  focuses on application of deep learning techniques to physics problems. I
  will present a more nuanced view of this interplay with examples of how t
 he structure of physics problems have inspired advances in deep learning a
 nd how it yields insights on topics such as inductive bias\, interpretabil
 ity\, and causality.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/6/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Hilbert Johan Kappen (Donder Institute\, Radboud University Nijmeg
 en\, the Netherlands)
DTSTART:20200528T163000Z
DTEND:20200528T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/7
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 7/">Path integral control theory</a>\nby Hilbert Johan Kappen (Donder Inst
 itute\, Radboud University Nijmegen\, the Netherlands) as part of Mathemat
 ics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nStochasti
 c optimal control theory deals with the problem to compute an optimal set 
 of actions to attain some future goal. Examples are found in many contexts
  such as motor control tasks for robotics\, planning and scheduling tasks 
 or managing a financial portfolio. The computation of the optimal control 
 is typically very difficult due to the size of the state space and the sto
 chastic nature of the problem. Special cases for which the computation is 
 tractable are linear dynamical systems with quadratic cost and determinist
 ic control problems. For a special class of non-linear stochastic control 
 problems\, the solution can be mapped onto a statistical inference problem
 . For these so-called path integral control problems the optimal cost-to-g
 o solution of the Bellman equation is given by the minimum of a free energ
 y. I will give a high level introduction to the underlying theory and illu
 strate with some examples from robotics and other areas.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/7/
END:VEVENT
BEGIN:VEVENT
SUMMARY:André David Mendes (CERN)
DTSTART:20200521T163000Z
DTEND:20200521T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/8
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 8/">How we discovered the Higgs ahead of schedule - ML's role in unveiling
  the keystone of elementary particle physics</a>\nby André David Mendes (
 CERN) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)
 \n\n\nAbstract\nn 2010\, when the LHC started colliding proton pairs in ea
 rnest\, multi-variate analyses were newfangled methods starting to make in
 roads in experimental particle physics. These methods faced widespread ske
 pticism as to their performance and biases\, reflecting a winter of suspic
 ion over overtrained neural networks that set in in the late 1990s. Thanks
  to more robust techniques\, like boosted decision trees\, it became possi
 ble to make better and more extensive use of the full information recorded
  in particle collisions at the Tevatron and LHC colliders.\n\nThe Higgs bo
 son discovery by the CMS and ATLAS collaborations in 2012 was only possibl
 e because of the use of multi-variate techniques that enhanced the sensiti
 vity by up to the equivalent of having 50% more collision data available f
 or analysis.\n\nWe will review the use of classification and regression in
  the Higgs to diphoton search and subsequent discovery\, a concrete exampl
 e of a decade-old ML-based analysis in high-energy particle physics. Parti
 cular emphasis will be placed in the modular design of the analysis and th
 e inherent explainability advantages\, used to great effect in assuaging c
 oncerns raised by hundreds of initially-skeptical colleagues in the CMS co
 llaboration.\nFinally\, we'll quickly highlight some particle physics chal
 lenges that have contributed to\, and made use of\, the last decade of gra
 ph\, adversarial\, and deep ML developments.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/8/
END:VEVENT
BEGIN:VEVENT
SUMMARY:João Miranda Lemos (Instituto Superior Técnico and INESC-ID)
DTSTART:20200716T163000Z
DTEND:20200716T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/9
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 9/">Reinforcement learning and adaptive control</a>\nby João Miranda Lemo
 s (Instituto Superior Técnico and INESC-ID) as part of Mathematics\, Phys
 ics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe aim of this sem
 inar is to explain\, to a wide audience\, how to combine optimal control t
 echniques with reinforcement learning\, by using approximate dynamic progr
 amming\, and artificial neural networks\, to obtain adaptive optimal contr
 ollers. Although with roots since the end of the XX century\, this problem
  has been the subject of an increasing attention. In addition to the promi
 sing tools that it offers to tackle difficult nonlinear problems with majo
 r engineering importance (ranging from robotics to biomedical engineering 
 and beyhond)\, it has the charm of creating a meeting point between the co
 ntrol and machine learning research communities.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/9/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Francisco C. Santos (Instituto Superior Técnico and INESC-ID)
DTSTART:20200709T163000Z
DTEND:20200709T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/10
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 10/">Climate action and cooperation dynamics under uncertainty</a>\nby Fra
 ncisco C. Santos (Instituto Superior Técnico and INESC-ID) as part of Mat
 hematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWhen
  attempting to avoid global warming\, individuals often face a social dile
 mma in which\, besides securing future benefits\, it is also necessary to 
 reduce the chances of future losses. In this talk\, I will resort to game 
 theory and populations of adaptive agents to offer a theoretical analysis 
 of this type of dilemmas\, in which the risk of failure plays a central ro
 le in individual decisions. I will discuss both deterministic dynamics in 
 large populations\, and stochastic social learning dynamics in finite popu
 lations. This class of models can be shown to capture some of the essentia
 l features discovered in recent key experiments while allowing one to exte
 nd in non-trivial ways the experimental conditions to regions of practical
  interest. Moreover\, this approach leads us to identify useful parallels 
 between ecological and socio-economic systems\, particularly in what conce
 rns the evolution and self-organization of their institutions. Particularl
 y\, our results suggest that global coordination for a common good should 
 be attempted through a polycentric structure of multiple small-scale agree
 ments\, in which perception of risk is high and uncertainty in collective 
 goals is minimized. Whenever the perception of risk is low\, our results i
 ndicate that sanctioning institutions may significantly enhance the chance
 s of coordinating to tame the planet's climate\, as long as they are imple
 mented in a bottom-up manner.  I will discuss the impact on public goods d
 ilemmas of heterogeneous political networks and wealth inequality\, includ
 ing distribution of wealth representative of existing inequalities among n
 ations. Finally\, I will briefly discuss the impact of scientific uncertai
 nty — both in what concerns the collective targets and the time window a
 vailable for action — on individuals' strategies and polarization of pre
 ferences.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/10/
END:VEVENT
BEGIN:VEVENT
SUMMARY:João Xavier (Instituto Superior Técnico and ISR)
DTSTART:20200618T163000Z
DTEND:20200618T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/11
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 11/">Learning from distributed datasets: an introduction with two examples
 </a>\nby João Xavier (Instituto Superior Técnico and ISR) as part of Mat
 hematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nData
  are increasingly measured\, in ever tinier minutiae\, by networks of spat
 ially distributed agents. Illustrative examples include a team of robots s
 earching a large region\, a collection of sensors overseeing a critical in
 fra-structure\, or a swarm of drones policing a wide area.\n\nHow to learn
  from these large\, spatially distributed datasets? In the centralized app
 roach each agent forwards its dataset to a fusion center\, which then carr
 ies out the learning from the pile of amassed datasets. This approach\, ho
 wever\, prevents the number of agents to scale up: as more and more agents
  ship data to the center\, not only the communication channels near the ce
 nter quickly swell to congestion\, but also the computational power of the
  center is rapidly outpaced.\n\nIn this seminar\, I describe the alternati
 ve approach of distributed learning. Here\, no fusion center exists\, and 
 the agents themselves recreate the centralized computation by exchanging s
 hort messages (not data) between network neighbors. To illustrate\, I desc
 ribe two learning algorithms: one solves convex learning problems via a to
 ken that randomly roams through the network\, and the other solves a class
 ification problem via random meetings between agents (e.g.\, gossip)\, eac
 h agent measuring only its own stream of features.\n\nThis seminar is aime
 d at non-specialists. Rather than trying to impart the latest developments
  of the field\, I hope to open a welcoming door to those wishing to have a
  peek at this bubbling field of research\, where optimization\, control\, 
 probability\, and machine learning mingle happily.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/11/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Marylou Gabrié (Center for Data Science\, NYU and Flatiron Instit
 ute\, CCM)
DTSTART:20200723T163000Z
DTEND:20200723T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/12
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 12/">Progress and hurdles in the statistical mechanics of deep learning</a
 >\nby Marylou Gabrié (Center for Data Science\, NYU and Flatiron Institut
 e\, CCM) as part of Mathematics\, Physics and Machine Learning (IST\, Lisb
 on)\n\n\nAbstract\nUnderstanding the great performances of deep neural net
 works is a very active direction of research with contributions coming fro
 m a wide variety of fields. The statistical mechanics of learning is a the
 oretical framework dating back to the 80s studying learning problems from 
 a physicist viewpoint and using tools from the physics of disordered syste
 ms. In this talk\, I will first go over this traditional framework\, which
  relies on the teacher-student scenario\, bayesian analysis and mean-field
  approximations. Then I will discuss some recent advances in the correspon
 ding analysis of modern deep neural network\, and highlight remaining chal
 lenges.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/12/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Masoud Mohseni (Google Quantum Artificial Intelligence Laboratory)
DTSTART:20200730T163000Z
DTEND:20200730T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/13
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 13/">TensorFlow Quantum: An open source framework for hybrid quantum-class
 ical machine learning.</a>\nby Masoud Mohseni (Google Quantum Artificial I
 ntelligence Laboratory) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nIn this talk\, I introduce TensorFlow Qu
 antum (TFQ)\, an open source library that was launched by Google in March 
 2020\, for the rapid prototyping of hybrid quantum-classical models for cl
 assical or quantum data.This framework offers high-level abstractions for 
 the design\, training\, and testing of both discriminative and generative 
 quantum models under TensorFlow and supports high-performance quantum circ
 uit simulators. I provide an overview of the software architecture and bui
 lding blocks through several examples and illustrate TFQ functionalities v
 ia constructing hybrid quantum-classical convolutional neural networks for
  quantum state classification.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/13/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gunnar Carlsson (Stanford University)
DTSTART:20200930T170000Z
DTEND:20200930T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/14
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 14/">Topological Data Analysis and Deep Learning</a>\nby Gunnar Carlsson (
 Stanford University) as part of Mathematics\, Physics and Machine Learning
  (IST\, Lisbon)\n\n\nAbstract\nDeep Learning is a powerful collection of t
 echniques for statistical learning\, which has shown dramatic applications
  in many different directions\, including including the study of data sets
  of images\, text\, and time series. It uses neural networks\, specificall
 y convolutional neural networks (CNN's)\, to produce these results. What w
 e have observed recently is that methods of topology can contribute to thi
 s effort\, in diagnosing behavior within the CNN's\, in the design of neur
 al networks with excellent computational properties\, and in improving gen
 eralization\, i.e. the transfer of results of one neural network from one 
 data set to another of similar type. We'll discuss topological methods in 
 data science\, as well as there application to this interesting set of tec
 hniques.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/14/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Lindsey Gray (Fermi National Accelerator Laboratory)
DTSTART:20201014T170000Z
DTEND:20201014T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/15
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 15/">Graph Neural Networks for Pattern Recognition in Particle Physics</a>
 \nby Lindsey Gray (Fermi National Accelerator Laboratory) as part of Mathe
 matics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nModern
  particle physics detectors generate copious amounts of data packed with m
 eaning that provides the means for high-quality measurements in demanding 
 experimental environments. To achieve these measurements there is a trend 
 towards finer granularity in these detectors and that implies the data rea
 d out has less intrinsic structure. Accurate pattern recognition is requir
 ed to define the signatures of particles within those detectors and simult
 aneously extract physical parameters for the particles. Typically\, algori
 thms to achieve these goals are written using well known unsupervised algo
 rithms\, but recent advances in machine learning on graph structures\, "Gr
 aph Neural Networks" (GNNs)\, provide powerful new methodologies for desig
 ning pattern recognition algorithms. In particular\, methodologies for pre
 dicting the link structure between pieces of data from detectors are well 
 suited to the particle physics pattern recognition task. Furthermore\, the
 re are interesting avenues for enforcing known symmetries of the data into
  the output of such networks and there is ongoing research in this directi
 on. This talk will discuss the challenges of pattern recognition\, the adv
 ent of GNNs and the connections to particle physics\, and the paths of res
 earch ahead for fully utilizing this powerful new tool.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/15/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Carola-Bibiane Schönlieb (DAMTP\, University of Cambridge)
DTSTART:20201120T150000Z
DTEND:20201120T160000Z
DTSTAMP:20260404T094311Z
UID:MPML/16
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 16/">Combining knowledge and data driven methods for solving inverse imagi
 ng problems - getting the best from both worlds</a>\nby Carola-Bibiane Sch
 önlieb (DAMTP\, University of Cambridge) as part of Mathematics\, Physics
  and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nInverse problems in im
 aging range from tomographic reconstruction (CT\, MRI\, etc) to image deco
 nvolution\, segmentation\, and classification\, just to name a few. In thi
 s talk I will discuss\napproaches to inverse imaging problems which have b
 oth a mathematical modelling (knowledge driven) and a machine learning (da
 ta-driven) component. Mathematical modelling is crucial in the presence of
  ill-posedness\, making use of information about the imaging data\, for na
 rrowing down the search space. Such an approach results in highly generali
 zable reconstruction and analysis methods which come with desirable soluti
 ons guarantees. Machine learning on the other hand is a powerful tool for 
 customising methods to individual data sets. Highly parametrised models su
 ch as deep neural networks in particular\, are powerful tools for accurate
 ly modelling prior information about solutions. The combination of these t
 wo paradigms\, getting the best from both of these worlds\, is the topic o
 f this talk\, furnished with examples for image classification under minim
 al supervision and for tomographic image reconstruction.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/16/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Weinan E (Princeton University)
DTSTART:20201007T100000Z
DTEND:20201007T110000Z
DTSTAMP:20260404T094311Z
UID:MPML/17
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 17/">Machine Learning and Scientific Computing</a>\nby Weinan E (Princeton
  University) as part of Mathematics\, Physics and Machine Learning (IST\, 
 Lisbon)\n\n\nAbstract\nNeural network-based deep learning is capable of ap
 proximating functions in very high dimension with unprecedented efficiency
  and accuracy. This has opened up many exciting new possibilities\, not ju
 st in traditional areas of artificial intelligence\, but also in scientifi
 c computing and computational science. At the same time\, deep learning ha
 s also acquired the reputation of being a set of “black box” type of t
 ricks\, without fundamental principles. This has been a real obstacle for 
 making further progress in machine learning.\n\nIn this talk\, I will try 
 to address the following two questions:\n\n1. How machine learning will im
 pact computational mathematics and computational science?\n\n2. How comput
 ational mathematics\, particularly numerical analysis\, can impact machine
  learning? We describe some of the most important progresses that have bee
 n made on these issues so far.\n\nOur hope is to put things into a perspec
 tive that will help to integrate machine learning with computational scien
 ce.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/17/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gitta Kutyniok (Mathematical Institute of the University of Munich
 )
DTSTART:20201202T180000Z
DTEND:20201202T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/18
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 18/">Deep Learning meets Physics: Taking the Best out of Both Worlds in Im
 aging Science</a>\nby Gitta Kutyniok (Mathematical Institute of the Univer
 sity of Munich) as part of Mathematics\, Physics and Machine Learning (IST
 \, Lisbon)\n\n\nAbstract\nPure model-based approaches are today often insu
 fficient for solving complex inverse problems in imaging. At the same time
 \, we witness the tremendous success of data-based methodologies\, in part
 icular\, deep neural networks for such problems. However\, pure deep learn
 ing approaches often neglect known and valuable information from physics.\
 n\nIn this talk\, we will provide an introduction to this problem complex 
 and then discuss a general conceptual approach to inverse problems in imag
 ing\, which combines deep learning and physics. This hybrid approach is ba
 sed on shearlet-based sparse regularization and deep learning and is guide
 d by a microlocal analysis viewpoint to pay particular attention to the si
 ngularity structures of the data. Finally\, we will present several applic
 ations such as tomographic reconstruction and show that our approach outpe
 rforms previous methodologies\, including methods entirely based on deep l
 earning.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/18/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Tommaso Dorigo (Italian Institute for Nuclear Physics)
DTSTART:20201125T180000Z
DTEND:20201125T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/19
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 19/">Dealing with Systematic Uncertainties in HEP Analysis with Machine Le
 arning Methods</a>\nby Tommaso Dorigo (Italian Institute for Nuclear Physi
 cs) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n
 \n\nAbstract\nI will discuss the impact of nuisance parameters on the effe
 ctiveness of supervised classification in high energy physics problems\, a
 nd techniques that may mitigate or remove their effect in the search for o
 ptimal selection criteria and variable transformations. The approaches dis
 cussed include nuisance parametrized models\, modified or adversary losses
 \, semi supervised learning approaches and inference-aware techniques.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/19/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Florent Krzakala (EPFL)
DTSTART:20201028T180000Z
DTEND:20201028T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/20
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 20/">Some exactly solvable models for statistical machine learning</a>\nby
  Florent Krzakala (EPFL) as part of Mathematics\, Physics and Machine Lear
 ning (IST\, Lisbon)\n\n\nAbstract\nThe increasing dimensionality of data i
 n the modern machine learning age presents new challenges and opportunitie
 s. The high-dimensional settings allow one to use powerful asymptotic meth
 ods from probability theory and statistical physics to obtain precise char
 acterizations and develop new algorithmic approaches. There is indeed a de
 cades-long tradition in statistical physics with building and solving such
  simplified models of neural networks.\n\nI will give examples of recent w
 orks that build on powerful methods of physics of disordered systems to an
 alyze different problems in machine learning and neural networks\, includi
 ng overparameterization\, kernel methods\, and the gradient descent algori
 thm in a high dimensional non-convex setting.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/20/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Joan Bruna (Courant Institute and Center for Data Science\, NYU)
DTSTART:20201104T180000Z
DTEND:20201104T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/21
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 21/">Mathematical aspects of neural network learning through measure dynam
 ics</a>\nby Joan Bruna (Courant Institute and Center for Data Science\, NY
 U) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\
 n\nAbstract\nHigh-dimensional learning remains an outstanding phenomena wh
 ere experimental evidence outpaces our current mathematical understanding\
 , mostly due to the recent empirical successes of Deep Learning algorithms
 . Neural Networks provide a rich yet intricate class of functions with sta
 tistical abilities to break the curse of dimensionality\, and where physic
 al priors can be tightly integrated into the architecture to improve sampl
 e efficiency. Despite these advantages\, an outstanding theoretical challe
 nge in these models is computational\, ie providing an analysis that expla
 ins successful optimization and generalization in the face of existing wor
 st-case computational hardness results.\n\nIn this talk\, I will focus on 
 the framework that lifts parameter optimization to an appropriate measure 
 space. I will cover existing results that guarantee global convergence of 
 the resulting Wasserstein gradient flows\, as well as recent results that 
 study typical fluctuations of the dynamics around their mean field evoluti
 on. We will also discuss extensions of this framework beyond vanilla super
 vised learning\, to account for symmetries in the function\, as well as fo
 r competitive optimization.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/21/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Bin Dong (BICMR\, Peking University)
DTSTART:20201111T110000Z
DTEND:20201111T120000Z
DTSTAMP:20260404T094311Z
UID:MPML/22
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 22/">Learning and Learning to Solve PDEs</a>\nby Bin Dong (BICMR\, Peking 
 University) as part of Mathematics\, Physics and Machine Learning (IST\, L
 isbon)\n\n\nAbstract\nDeep learning continues to dominate machine learning
  and has been successful in computer vision\, natural language processing\
 , etc. Its impact has now expanded to many research areas in science and e
 ngineering. In this talk\, I will mainly focus on some recent impact of de
 ep learning on computational mathematics. I will present our recent work o
 n bridging deep neural networks with numerical differential equations. On 
 the one hand\, I will show how to design transparent deep convolutional ne
 tworks to uncover hidden PDE models from observed dynamical data. On the o
 ther hand\, I will present our preliminary attempt to establish a deep rei
 nforcement learning based framework to solve 1D scalar conservation laws\,
  and a meta-learning approach for solving linear parameterized PDEs based 
 on the multigrid method.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/22/
END:VEVENT
BEGIN:VEVENT
SUMMARY:René Vidal (Mathematical Institute for Data Science\, Johns Hopki
 ns University)
DTSTART:20201216T180000Z
DTEND:20201216T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/23
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 23/">From Optimization Algorithms to Dynamical Systems and Back</a>\nby Re
 né Vidal (Mathematical Institute for Data Science\, Johns Hopkins Univers
 ity) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\
 n\n\nAbstract\nRecent work has shown that tools from dynamical systems can
  be used to analyze accelerated optimization algorithms. For example\, it 
 has been shown that the continuous limit of Nesterov’s accelerated gradi
 ent (NAG) gives an ODE whose convergence rate matches that of NAG for conv
 ex\, unconstrained\, and smooth problems. Conversely\, it has been shown t
 hat NAG can be obtained as the discretization of an ODE\, however since di
 fferent discretizations lead to different algorithms\, the choice of the d
 iscretization becomes important. The first part of this talk will extend t
 his type of analysis to convex\, constrained and non-smooth problems by us
 ing Lyapunov stability theory to analyze continuous limits of the Alternat
 ing Direction Method of Multipliers (ADMM). The second part of this talk w
 ill show that many existing and new optimization algorithms can be obtaine
 d by suitably discretizing a dissipative Hamiltonian. As an example\, we w
 ill present a new method called Relativistic Gradient Descent (RGD)\, whic
 h empirically outperforms momentum\, RMSprop\, Adam and AdaGrad on several
  non-convex\nproblems.\n\nThis is joint work with Guilherme Franca\, Danie
 l Robinson and Jeremias Sulam.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/23/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mauro Maggioni (Johns Hopkins University)
DTSTART:20201021T170000Z
DTEND:20201021T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/24
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 24/">Learning Interaction laws in particle- and agent-based systems</a>\nb
 y Mauro Maggioni (Johns Hopkins University) as part of Mathematics\, Physi
 cs and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nInteracting agent-ba
 sed systems are ubiquitous in science\, from modeling of particles in Phys
 ics to prey-predator and colony models in Biology\, to opinion dynamics in
  economics and social sciences. Oftentimes the laws of interactions betwee
 n the agents are quite simple\, for example they depend only on pairwise i
 nteractions\, and only on pairwise distance in each interaction. We consid
 er the following inference problem for a system of interacting particles o
 r agents: given only observed trajectories of the agents in the system\, c
 an we learn what the laws of interactions are? We would like to do this wi
 thout assuming any particular form for the interaction laws\, i.e. they mi
 ght be "any" function of pairwise distances. We consider this problem both
  the mean-field limit (i.e. the number of particles going to infinity) and
  in the case of a finite number of agents\, with an increasing number of o
 bservations\, albeit in this talk we will mostly focus on the latter case.
  We cast this as an inverse problem\, and study it in the case where the i
 nteraction is governed by an (unknown) function of pairwise distances. We 
 discuss when this problem is well-posed\, and we construct estimators for 
 the interaction kernels with provably good statistically and computational
  properties. We measure their performance on various examples\, that inclu
 de extensions to agent systems with different types of agents\, second-ord
 er systems\, and families of systems with parametric interaction kernels. 
 We also conduct numerical experiments to test the large time behavior of t
 hese systems\, especially in the cases where they exhibit emergent behavio
 r.\n\nThis is joint work with F. Lu\, J.Miller\, S. Tang and M. Zhong.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/24/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Xavier Bresson (Nanyang Technological University)
DTSTART:20210127T110000Z
DTEND:20210127T120000Z
DTSTAMP:20260404T094311Z
UID:MPML/25
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 25/">Benchmarking Graph Neural Networks</a>\nby Xavier Bresson (Nanyang Te
 chnological University) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nGraph neural networks (GNNs) have become
  the standard toolkit for analyzing and learning from data on graphs. As t
 he field grows\, it becomes critical to identify key architectures and val
 idate new ideas that generalize to larger\, more complex datasets. Unfortu
 nately\, it has been increasingly difficult to gauge the effectiveness of 
 new models in the absence of a standardized benchmark with consistent expe
 rimental settings. In this work\, we introduce a reproducible GNN benchmar
 king framework\, with the facility for researchers to add new models conve
 niently for arbitrary datasets. We demonstrate the usefulness of our frame
 work by presenting a principled investigation into the recent Weisfeiler-L
 ehman GNNs (WL-GNNs) compared to message passing-based graph convolutional
  networks (GCNs) for a variety of graph tasks with medium-scale datasets.\
 n
LOCATION:https://stable.researchseminars.org/talk/MPML/25/
END:VEVENT
BEGIN:VEVENT
SUMMARY:James Halverson (Northeastern University)
DTSTART:20210120T180000Z
DTEND:20210120T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/26
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 26/">Neural Networks and Quantum Field Theory</a>\nby James Halverson (Nor
 theastern University) as part of Mathematics\, Physics and Machine Learnin
 g (IST\, Lisbon)\n\n\nAbstract\nIn this talk I will review essentials of q
 uantum field theory (QFT) and demonstrate how the function-space distribut
 ion of many neural networks (NNs) shares similar properties. This allows\,
  for instance\, computation of correlators of neural network outputs in te
 rms of Feynman diagrams and a direct analogy between non-Gaussian correcti
 ons in NN distributions and particle interactions. Some cases yield diverg
 ences in perturbation theory\, requiring the introduction of regularizatio
 n and renormalization. Potential advantages of this perspective will be di
 scussed\, including a duality between function-space and parameter-space d
 escriptions of neural networks.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/26/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Anna C. Gilbert (Yale University)
DTSTART:20210113T180000Z
DTEND:20210113T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/27
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 27/">Metric representations: Algorithms and Geometry</a>\nby Anna C. Gilbe
 rt (Yale University) as part of Mathematics\, Physics and Machine Learning
  (IST\, Lisbon)\n\n\nAbstract\nGiven a set of distances amongst points\, d
 etermining what metric representation is most "consistent" with the input 
 distances or the metric that best captures the relevant geometric features
  of the data is a key step in many machine learning algorithms. In this ta
 lk\, we focus on 3 specific metric constrained problems\, a class of optim
 ization problems with metric constraints: metric nearness (Brickell et al.
  (2008))\, weighted correlation clustering on general graphs (Bansal et al
 . (2004))\, and metric learning (Bellet et al. (2013)\; Davis et al. (2007
 )).\n\nBecause of the large number of constraints in these problems\, howe
 ver\, these and other researchers have been forced to restrict either the 
 kinds of metrics learned or the size of the problem that can be solved. We
  provide an algorithm\, PROJECT AND FORGET\, that uses Bregman projections
  with cutting planes\, to solve metric constrained problems with many (pos
 sibly exponentially) inequality constraints. We also prove that our algori
 thm converges to the global optimal solution. Additionally\, we show that 
 the optimality error decays asymptotically at an exponential rate. We show
  that using our method we can solve large problem instances of three types
  of metric constrained problems\, out-performing all state of the art meth
 ods with respect to CPU times and problem sizes.\n\nFinally\, we discuss t
 he adaptation of PROJECT AND FORGET to specific types of metric constraint
 s\, namely tree and hyperbolic metrics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/27/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Caroline Uhler (MIT and Institute for Data\, Systems and Society)
DTSTART:20210210T180000Z
DTEND:20210210T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/28
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 28/">Causal Inference and Overparameterized Autoencoders in the Light of D
 rug Repurposing for SARS-CoV-2</a>\nby Caroline Uhler (MIT and Institute f
 or Data\, Systems and Society) as part of Mathematics\, Physics and Machin
 e Learning (IST\, Lisbon)\n\n\nAbstract\nMassive data collection holds the
  promise of a better understanding of complex phenomena and ultimately\, o
 f better decisions. An exciting opportunity in this regard stems from the 
 growing availability of perturbation / intervention data (drugs\, knockout
 s\, overexpression\,\netc.) in biology. In order to obtain mechanistic ins
 ights from such data\, a major challenge is the development of a framework
  that integrates observational and interventional data and allows predicti
 ng the effect of yet unseen interventions or transporting the effect of in
 terventions observed in one context to another. I will present a framework
  for causal structure discovery based on such data and highlight the role 
 of overparameterized autoencoders. We end by demonstrating how these ideas
  can be applied for drug repurposing in the current SARS-CoV-2 crisis.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/28/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Samantha Kleinberg (Stevens Institute of Technology)
DTSTART:20201209T180000Z
DTEND:20201209T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/29
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 29/">Data\, Decisions\, and You: Making Causality Useful and Usable in a C
 omplex World</a>\nby Samantha Kleinberg (Stevens Institute of Technology) 
 as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\n
 Abstract\nThe collection of massive observational datasets has led to unpr
 ecedented opportunities for causal inference\, such as using electronic he
 alth records to identify risk factors for disease. However\, our ability t
 o understand these complex data sets has not grown the same pace as our ab
 ility to collect them. While causal inference has traditionally focused on
  pairwise relationships between variables\, biological systems are highly 
 complex and knowing when events may happen is often as important as knowin
 g whether they will. In the first half of this talk I discuss new methods 
 that allow causal relationships to be reliably inferred from complex obser
 vational data\, motivated by analysis of intensive care unit and other med
 ical data. Causes are useful because they allow us to take action\, but ho
 w there is a gap between the output of machine learning and what helps peo
 ple make decisions. In the second part of this talk I discuss our recent f
 indings in testing just how people fare when using the output of machine l
 earning and how we can go from data to knowledge to decisions.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/29/
END:VEVENT
BEGIN:VEVENT
SUMMARY:A. Pedro Aguiar (Faculdade de Engenharia\, Universidade do Porto)
DTSTART:20210303T180000Z
DTEND:20210303T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/31
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 31/">Model based control design combining Lyapunov and optimization tools:
  Examples in the area of motion control of autonomous robotic vehicles</a>
 \nby A. Pedro Aguiar (Faculdade de Engenharia\, Universidade do Porto) as 
 part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbs
 tract\nThe past few decades have witnessed a significant research effort i
 n the field of Lyapunov model based control design. In parallel\, optimal 
 control and optimization model based design have also expanded their range
  of applications\, and nowadays\, receding horizon approaches can be consi
 dered a mature field for particular classes of control systems.\nIn this t
 alk\, I will argue that Lyapunov based techniques play an important role f
 or analysis of model based optimization methodologies and moreover\, both 
 approaches can be combined for control design resulting in powerful framew
 orks with formal guarantees of robustness\, stability\, performance\, and 
 safety. Illustrative examples in the area of motion control of autonomous 
 robotic vehicles will be presented for Autonomous Underwater Vehicles (AUV
 s)\, Autonomous Surface Vehicles (ASVs) and Unmanned Aerial Vehicles (UAVs
 ).\n
LOCATION:https://stable.researchseminars.org/talk/MPML/31/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sanjeev Arora (Computer Science Department\, Princeton University)
DTSTART:20210106T180000Z
DTEND:20210106T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/32
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 32/">The quest for mathematical understanding of deep learning</a>\nby San
 jeev Arora (Computer Science Department\, Princeton University) as part of
  Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n
 Deep learning has transformed Machine Learning and Artificial Intelligence
  in the past decade. It raises fundamental questions for mathematics and t
 heory of computer science\, since it relies upon solving large-scale nonco
 nvex problems via gradient descent and its variants. This talk will be an 
 introduction to mathematical questions raised by deep learning\, and some 
 partial understanding obtained in recent years with respect to optimizatio
 n\, generalization\, self-supervised learning\, privacy etc.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/32/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Jan Peters (Technische Universitaet Darmstadt)
DTSTART:20210423T130000Z
DTEND:20210423T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/33
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 33/">Robot Learning - Quo Vadis?</a>\nby Jan Peters (Technische Universita
 et Darmstadt) as part of Mathematics\, Physics and Machine Learning (IST\,
  Lisbon)\n\n\nAbstract\nAutonomous robots that can assist humans in situat
 ions of daily life have been a long standing vision of robotics\, artifici
 al intelligence\, and cognitive sciences. A first step towards this goal i
 s to create robots that can learn tasks triggered by environmental context
  or higher level instruction. However\, learning techniques have yet to li
 ve up to this promise as only few methods manage to scale to high-dimensio
 nal manipulator or humanoid robots. In this talk\, we investigate a genera
 l framework suitable for learning motor skills in robotics which is based 
 on the principles behind many analytical robotics approaches. It involves 
 generating a representation of motor skills by parameterized motor primiti
 ve policies acting as building blocks of movement generation\, and a learn
 ed task module that transforms these movements into motor commands. We dis
 cuss learning on three different levels of abstraction\, i.e.\, learning f
 or accurate control is needed to execute\, learning of motor primitives is
  needed to acquire simple movements\, and learning of the task-dependent 
 „hyperparameters“ of these motor primitives allows learning complex ta
 sks. We discuss task-appropriate learning approaches for imitation learnin
 g\, model learning and reinforcement learning for robots with many degrees
  of freedom. Empirical evaluations on a several robot systems illustrate t
 he effectiveness and applicability to learning control on an anthropomorph
 ic robot arm. These robot motor skills range from toy examples (e.g.\, pad
 dling a ball\, ball-in-a-cup\, juggling) to playing robot table tennis aga
 inst a human being and manipulation of various objects.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/33/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Miguel Couceiro (Université de Lorraine)
DTSTART:20210203T180000Z
DTEND:20210203T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/34
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 34/">Making ML Models fairer through explanations\, feature dropout\, and 
 aggregation</a>\nby Miguel Couceiro (Université de Lorraine) as part of M
 athematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nAl
 gorithmic decisions are now being used on a daily basis\, and based on Mac
 hine Learning (ML) processes that may be complex and biased. This raises s
 everal concerns given the critical impact that biased decisions may have o
 n individuals or on society as a whole. Not\nonly unfair outcomes affect h
 uman rights\, they also undermine public trust in ML and AI. In this talk\
 , we will address fairness issues of ML models based on decision outcomes\
 , and we will show how the simple idea of "feature dropout" followed by an
  "ensemble approach" can improve model fairness without compromising its a
 ccuracy. To illustrate we will present a general workflow that relies on e
 xplainers to tackle "process fairness"\, which essentially measures a mode
 l's reliance on sensitive or discriminatory features. We will present diff
 erent applications and empirical settings that show improvements not only 
 with respect to process fairness but also other fairness metrics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/34/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Steve Brunton (University of Washington)
DTSTART:20210331T170000Z
DTEND:20210331T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/35
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 35/">Machine learning for Fluid Mechanics</a>\nby Steve Brunton (Universit
 y of Washington) as part of Mathematics\, Physics and Machine Learning (IS
 T\, Lisbon)\n\n\nAbstract\nMany tasks in fluid mechanics\, such as design 
 optimization and control\, are challenging because fluids are nonlinear an
 d exhibit a large range of scales in both space and time. This range of sc
 ales necessitates exceedingly high-dimensional measurements and computatio
 nal discretization to resolve all relevant features\, resulting in vast da
 ta sets and time-intensive computations. Indeed\, fluid dynamics is one of
  the original big data fields\, and many high-performance computing archit
 ectures\, experimental measurement techniques\, and advanced data processi
 ng and visualization algorithms were driven by decades of research in flui
 d mechanics. Machine learning constitutes a growing set of powerful techni
 ques to extract patterns and build models from this data\, complementing t
 he existing theoretical\, numerical\, and experimental efforts in fluid me
 chanics. In this talk\, we will explore current goals and opportunities fo
 r machine learning in fluid mechanics\, and we will highlight a number of 
 recent technical advances. Because fluid dynamics is central to transporta
 tion\, health\, and defense systems\, we will emphasize the importance of 
 machine learning solutions that are interpretable\, explainable\, generali
 zable\, and that respect known physics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/35/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Hsin Yuan Huang\, (Robert) (Caltech)
DTSTART:20210317T180000Z
DTEND:20210317T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/36
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 36/">Information-theoretic bounds on quantum advantage in machine learning
 </a>\nby Hsin Yuan Huang\, (Robert) (Caltech) as part of Mathematics\, Phy
 sics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWe compare the com
 plexity of training classical and quantum machine learning (ML) models for
  predicting outcomes of physical experiments. The experiments depend on an
  input parameter x and involve the execution of a (possibly unknown) quant
 um process $E$. Our figure of merit is the number of runs of $E$ needed du
 ring training\, disregarding other measures of complexity. A classical ML 
 performs a measurement and records the classical outcome after each run of
  $E$\, while a quantum ML can access $E$ coherently to acquire quantum dat
 a\; the classical or quantum data is then used to predict outcomes of futu
 re experiments. We prove that\, for any input distribution $D(x)$\, a clas
 sical ML can provide accurate predictions on average by accessing $E$ a nu
 mber of times comparable to the optimal quantum ML. In contrast\, for achi
 eving accurate prediction on all inputs\, we show that exponential quantum
  advantage exists in certain tasks. For example\, to predict expectation v
 alues of all Pauli observables in an $n-$qubit system\, we present a quant
 um ML using only $O(n)$ data and prove that a classical ML requires $2^{\\
 Omega(n)}$ data.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/36/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mário Figueiredo (Instituto Superior Técnico and IT)
DTSTART:20210217T180000Z
DTEND:20210217T190000Z
DTSTAMP:20260404T094311Z
UID:MPML/37
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 37/">Dealing with Correlated Variables in Supervised Learning</a>\nby Már
 io Figueiredo (Instituto Superior Técnico and IT) as part of Mathematics\
 , Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nLinear (and g
 eneralized linear) regression (LR) is an old\, but still essential\, stati
 stical tool: its goal is to learn to predict a (response) variable from a 
 linear combination of other (explanatory) variables. A central problem in 
 LR is the selection of relevant variables\, because using fewer variables 
 tends to yield better generalization and because this identification may b
 e meaningful (e.g.\, which genes are relevant to predict a certain disease
 ). In the past quarter-century\, variable selection (VS) based on sparsity
 -inducing regularizers has been a central paradigm\, the most famous examp
 le being the LASSO\, which has been intensively studied\,\nextended\, and 
 applied.\n\nIn many contexts\, it is natural to have highly-correlated var
 iables (e.g.\, several genes that are strongly co-regulated)\, thus simult
 aneously relevant as predictors. In this case\, sparsity-based VS may fail
 : it may select an arbitrary subset of these variables and it is unstable.
  Moreover\, it is often desirable to identify all the relevant variables\,
  not just an arbitrary subset thereof\, a goal for which several approache
 s have been proposed. This talk will be devoted to a recent class of such 
 approaches\, called ordered weighted l1 (OWL). The key feature of OWL is t
 hat it is provably able to explicitly identify (i.e. cluster) sufficiently
 -correlated features\, without having to compute these correlations. Sever
 al theoretical results characterizing OWL will be presented\, including co
 nnections to the mathematics of economic inequality. Computational and opt
 imization aspects will also be addressed\, as well as recent applications 
 in subspace clustering\, learning Gaussian graphical models\, and deep neu
 ral networks.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/37/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Maciej Koch-J8anusz (University of Zurich)
DTSTART:20210222T170000Z
DTEND:20210222T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/38
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 38/">Statistical physics through the lens of real-space mutual information
 </a>\nby Maciej Koch-J8anusz (University of Zurich) as part of Mathematics
 \, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIdentifying 
 the relevant coarse-grained degrees of freedom in a complex physical syste
 m is a key stage in developing effective theories. The renormalization gro
 up (RG) provides a framework for this task\, but its practical execution i
 n unfamiliar systems is fraught with ad hoc choices. Machine learning appr
 oaches\, on the other hand\, though promising\, often lack formal interpre
 tability: it is unclear what relation\, if any\, the architecture- and tra
 ining-dependent learned "relevant" features bear to standard objects of ph
 ysical theory.\n\nI will present recent results addressing both issues. We
  develop a fast algorithm\, the RSMI-NE\, employing state-of-art results i
 n machine-learning-based estimation of information-theoretic quantities to
  construct the optimal coarse-graining. We use it to develop a new approac
 h to identifying the most relevant field theory operators describing a sta
 tistical system\, which we validate on the example of interacting dimer mo
 del. I will also discuss formal results underlying the method: we establis
 h equivalence between the information-theoretic notion of relevance define
 d in the Information Bottleneck (IB) formalism of compression theory\, and
  the field-theoretic relevance of the RG. We show analytically that for st
 atistical physical systems the "relevant" degrees of freedom found using I
 B compression indeed correspond to operators with the lowest scaling dimen
 sions\, providing a dictionary connecting two distinct theoretical toolbox
 es.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/38/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Markus Heyl (Max-Planck Institute for the Physics of Complex Syste
 ms\, Dresden)
DTSTART:20210322T170000Z
DTEND:20210322T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/39
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 39/">Quantum many-body dynamics in two dimensions with artificial neural n
 etworks</a>\nby Markus Heyl (Max-Planck Institute for the Physics of Compl
 ex Systems\, Dresden) as part of Mathematics\, Physics and Machine Learnin
 g (IST\, Lisbon)\n\n\nAbstract\nIn the last two decades the field of noneq
 uilibrium quantum many-body physics has seen a rapid development driven\, 
 in particular\, by the remarkable progress in quantum simulators\, which t
 oday provide access to dynamics in quantum matter with an unprecedented co
 ntrol. However\, the efficient numerical simulation of nonequilibrium real
 -time evolution in isolated quantum matter still remains a key challenge f
 or current computational methods especially beyond one spatial dimension. 
 In this talk I will present a versatile and efficient machine learning ins
 pired approach. I will first introduce the general idea of encoding quantu
 m many-body wave functions into artificial neural networks. I will then id
 entify and resolve key challenges for the simulation of real-time evolutio
 n\, which previously imposed significant limitations on the accurate descr
 iption of large systems and long-time dynamics. As a concrete example\, I 
 will consider the dynamics of the paradigmatic two-dimensional transverse 
 field Ising model\, where we observe collapse and revival oscillations of 
 ferromagnetic order and demonstrate that the reached time scales are compa
 rable to or exceed the capabilities of state-of-the-art tensor network met
 hods.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/39/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Pedro A. Santos (Instituto Superior Técnico and INESC-ID)
DTSTART:20210409T130000Z
DTEND:20210409T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/40
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 40/">Two-time scale stochastic approximation for reinforcement learning wi
 th linear function approximation</a>\nby Pedro A. Santos (Instituto Superi
 or Técnico and INESC-ID) as part of Mathematics\, Physics and Machine Lea
 rning (IST\, Lisbon)\n\n\nAbstract\nIn this presentation\, I will introduc
 e some traditional Reinforcement Learning problems and algorithms\, and an
 alyze how some problems can be avoided and convergence results obtained us
 ing a two-time scale variation of the usual stochastic approximation appro
 ach.\n\nThis variation was inspired by the practical successes of Deep Q-L
 earning in attaining superhuman performance at some classical Atari games 
 by Deepmind's research team in 2015. Machine Learning practical successes 
 like this often have no corresponding explaining theory. The work that wil
 l be presented intends to contribute to that goal.\n\nJoint work with Diog
 o Carvalho and Francisco Melo from INESC-ID.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/40/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rebecca Willett (University of Chicago)
DTSTART:20210507T130000Z
DTEND:20210507T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/41
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 41/">Machine Learning and Inverse Problems: Deeper and More Robust</a>\nby
  Rebecca Willett (University of Chicago) as part of Mathematics\, Physics 
 and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMany challenging image 
 processing tasks can be described by an ill-posed linear inverse problem: 
 deblurring\, deconvolution\, inpainting\, compressed sensing\, and superre
 solution all lie in this framework. Recent advances in machine learning an
 d image processing have illustrated that it is often possible to learn a r
 egularizer from training data that can outperform more traditional approac
 hes by large margins. In this talk\, I will describe the central prevailin
 g themes of this emerging area and a taxonomy that can be used to categori
 ze different problems and reconstruction methods. We will also explore mec
 hanisms for model adaptation\; that is\, given a network trained to solve 
 an initial inverse problem with a known forward model\, we propose novel p
 rocedures that adapt the network to a perturbed forward model\, even witho
 ut full knowledge of the perturbation. Finally\, I will describe a new cla
 ss of approaches based on "infinite-depth networks" that can yield up to a
  4dB PSNR improvement in reconstruction accuracy above state-of-the-art al
 ternatives and where the computational budget can be selected at test time
  to optimize context-dependent trade-offs between accuracy and computation
 .\n
LOCATION:https://stable.researchseminars.org/talk/MPML/41/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mikhail Belkin (Halicioğlu Data Science Institute\, University of
  California San Diego)
DTSTART:20210428T170000Z
DTEND:20210428T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/42
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 42/">Two mathematical lessons of deep learning</a>\nby Mikhail Belkin (Hal
 icioğlu Data Science Institute\, University of California San Diego) as p
 art of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbst
 ract\nRecent empirical successes of deep learning have exposed significant
  gaps in our fundamental understanding of learning and optimization mechan
 isms. Modern best practices for model selection are in direct contradictio
 n to the methodologies suggested by classical analyses. Similarly\, the ef
 ficiency of SGD-based local methods used in training modern models\, appea
 red at odds with the standard intuitions on optimization.\n\nFirst\, I wil
 l present evidence\, empirical and mathematical\, that necessitates revisi
 ting classical notions\, such as over-fitting. I will continue to discuss 
 the emerging understanding of generalization\, and\, in particular\, the "
 double descent" risk curve\, which extends the classical U-shaped generali
 zation curve beyond the point of interpolation.\n\nSecond\, I will discuss
  why the landscapes of over-parameterized neural networks are generically 
 never convex\, even locally. Instead\, as I will argue\, they satisfy the 
 Polyak-Lojasiewicz condition across most of the parameter space instead\, 
 which allows SGD-type methods to converge to a global minimum.\n\nA key pi
 ece of the puzzle remains - how does optimization align with statistics to
  form the complete mathematical picture of modern ML?\n
LOCATION:https://stable.researchseminars.org/talk/MPML/42/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gabriel Peyré (École Normale Supérieure)
DTSTART:20210414T170000Z
DTEND:20210414T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/43
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 43/">Scaling Optimal Transport for High dimensional Learning</a>\nby Gabri
 el Peyré (École Normale Supérieure) as part of Mathematics\, Physics an
 d Machine Learning (IST\, Lisbon)\n\n\nAbstract\nOptimal transport (OT) ha
 s recently gained lot of interest in machine learning. It is a natural too
 l to compare in a geometrically faithful way probability distributions. It
  finds applications in both supervised learning (using geometric loss func
 tions) and unsupervised learning (to perform generative model fitting). OT
  is however plagued by the curse of dimensionality\, since it might requir
 e a number of samples which grows exponentially with the dimension. In thi
 s talk\, I will explain how to leverage entropic regularization methods to
  define computationally efficient loss functions\, approximating OT with a
  better sample complexity.\n\nMore information and references can be found
  on the website of our book\n"Computational Optimal Transport"\, https://o
 ptimaltransport.github.io/\n
LOCATION:https://stable.researchseminars.org/talk/MPML/43/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Kyriakos Vamvoudakis (Georgia Institute of Technology)
DTSTART:20210521T130000Z
DTEND:20210521T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/44
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 44/">Learning-Based Actuator Placement and Receding Horizon Control for Se
 curity against Actuation Attacks</a>\nby Kyriakos Vamvoudakis (Georgia Ins
 titute of Technology) as part of Mathematics\, Physics and Machine Learnin
 g (IST\, Lisbon)\n\n\nAbstract\nCyber-physical systems (CPS) comprise inte
 racting digital\, analog\, physical\, and human components engineered for 
 function through integrated physics and logic. Incorporating intelligence 
 in CPS\, however\, makes their physical components more exposed to adversa
 ries that can potentially cause failure or malfunction through actuation a
 ttacks. As a result\, augmenting CPS with resilient control and design met
 hods is of grave significance\, especially if an actuation attack is steal
 thy. Towards this end\, in the first part of the talk\, I will present a r
 eceding horizon controller\, which can deal with undetectable actuation at
 tacks by solving a game in a moving horizon fashion. In fact\, this contro
 ller can guarantee stability of the equilibrium point of the CPS\, even if
  the attackers have an information advantage. The case where the attackers
  are not aware of the decision-making mechanism of one another is also con
 sidered\, by exploiting the theory of bounded rationality. In the second p
 art of the talk\, and for CPS that have partially unknown dynamics\, I wil
 l present an online actuator placement algorithm\, which chooses the actua
 tors of the CPS that maximize an attack security metric. It can be proved 
 that the maximizing set of actuators is found in finite time\, despite the
  CPS having uncertain dynamics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/44/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gustau Camps-Valls (Universitat de València)
DTSTART:20210528T130000Z
DTEND:20210528T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/45
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 45/">Physics Aware Machine Learning for the Earth Sciences</a>\nby Gustau 
 Camps-Valls (Universitat de València) as part of Mathematics\, Physics an
 d Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMost problems in Earth sc
 iences aim to do inferences about the system\, where accurate predictions 
 are just a tiny part of the whole problem. Inferences mean understanding v
 ariables relations\, deriving models that are physically interpretable\, t
 hat are simple parsimonious\, and mathematically tractable. Machine learni
 ng models alone are excellent approximators\, but very often do not respec
 t the most elementary laws of physics\, like mass or energy conservation\,
  so consistency and confidence are compromised. I will review the main cha
 llenges ahead in the field\, and introduce several ways to live in the Phy
 sics and machine learning interplay that allows us (1) to encode different
 ial equations from data\, (2) constrain data-driven models with physics-pr
 iors and dependence constraints\, (3) improve parameterizations\, (4) emul
 ate physical models\, and (5) blend data-driven and process-based models. 
 This is a collective long-term AI agenda towards developing and applying a
 lgorithms capable of discovering knowledge in the Earth system.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/45/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ulugbek Kamilov (University of Washington)
DTSTART:20210611T130000Z
DTEND:20210611T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/46
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 46/">Computational Imaging: Reconciling Physical and Learned Models</a>\nb
 y Ulugbek Kamilov (University of Washington) as part of Mathematics\, Phys
 ics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p class="western" 
 style="text-align:justify"><span style="line-height:100%"><font face="Cali
 bri\, serif"><span style="font-size:11pt">Computational imaging is a rapid
 ly growing area that seeks to enhance the capabilities of imaging instrume
 nts by viewing imaging as an inverse problem. There are currently two dist
 inct approaches for designing computational imaging methods: model-based a
 nd learning-based. Model-based methods leverage analytical signal properti
 es and often come with theoretical guarantees and insights. Learning-based
  methods leverage data-driven representations for best empirical performan
 ce through training on large datasets. This talk presents Regularization b
 y Artifact Removal (RARE)\, as a framework for reconciling both viewpoints
  by providing a learning-based extension to the classical theory. RARE rel
 ies on pre-trained “artifact-removing deep neural nets” for infusing l
 earned prior knowledge into an inverse problem\, while maintaining a clear
  separation between the prior and physics-based acquisition model. O</span
 ></font><font face="Calibri\, serif"><span style="font-size:11pt">ur resul
 ts indicate that RARE can achieve state-of-the-art performance in differen
 t computational imaging tasks\, while also being amenable to rigorous theo
 retical analysis. We will focus on the applications of RARE in biomedical 
 imaging\, including magnetic resonance and tomographic imaging.</span></fo
 nt></span></p>\n\n<p class="western" style="text-align:justify"><span styl
 e="line-height:100%"><font face="Calibri\, serif"><span style="font-size:1
 1pt"><b>This talk will be based on the following references</b></span></fo
 nt></span></p>\n\n<ol>\n	<li class="western"><span style="line-height:100%
 "><font face="Calibri\, serif"><span style="font-size:11pt">J. Liu\, Y. Su
 n\, C. Eldeniz\, W. Gan\, H. An\, and U. S. Kamilov\, “<a href="https://
 arxiv.org/abs/1912.05854">RARE: Image Reconstruction using Deep Priors Lea
 rned without Ground Truth\,</a>” IEEE J. Sel. Topics Signal Process.\, v
 ol. 14\, no. 6\, pp. 1088-1099\, October 2020.</span></font></span></li>\n
 	<li class="western" style="text-align: justify\;"><span style="line-heigh
 t:100%"><font face="Calibri\, serif"><span style="font-size:11pt">Z. Wu\, 
 Y. Sun\, A. Matlock\, J. Liu\, L. Tian\, and U. S. Kamilov\, “<a href="h
 ttps://arxiv.org/abs/1911.13241">SIMBA: Scalable Inversion in Optical Tomo
 graphy using Deep Denoising Priors</a>\,” IEEE J. Sel. Topics Signal Pro
 cess.\, vol. 14\, no. 6\, pp. 1163-1175\, October 2020.</span></font></spa
 n></li>\n	<li class="western" style="text-align: justify\;"><span style="l
 ine-height:100%"><font face="Calibri\, serif"><span style="font-size:11pt"
 >J. Liu\, Y. Sun\, W. Gan\, X. Xu\, B. Wohlberg\, and U. S. Kamilov\, “<
 a href="https://arxiv.org/abs/2101.09379">SGD-Net: Efficient Model-Based D
 eep Learning with Theoretical Guarantees</a>\,” IEEE Trans. Comput. Imag
 .\, in press.</span></font></span></li>\n</ol>\n\n<p class="western" style
 ="text-align:justify">&nbsp\;</p>\n
LOCATION:https://stable.researchseminars.org/talk/MPML/46/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ruth Misener (Imperial College London)
DTSTART:20210618T130000Z
DTEND:20210618T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/47
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 47/">Partition-based formulations for mixed-integer optimization of traine
 d ReLU neural networks</a>\nby Ruth Misener (Imperial College London) as p
 art of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbst
 ract\nThis work develops a class of relaxations in between the big-M and c
 onvex hull formulations of disjunctions\, drawing advantages from both. We
  show that this class leads to mixed-integer formulations for trained ReLU
  neural networks. The approach balances model size and tightness by partit
 ioning node inputs into a number of groups and forming the convex hull ove
 r the partitions via disjunctive programming. At one extreme\, one partiti
 on per input recovers the convex hull of a node\, i.e.\, the tightest poss
 ible formulation for each node. For fewer partitions\, we develop smaller 
 relaxations that approximate the convex hull\, and show that they outperfo
 rm existing formulations. Specifically\, we propose strategies for partiti
 oning variables based on theoretical motivations and validate these strate
 gies using extensive computational experiments. Furthermore\, the proposed
  scheme complements known algorithmic approaches\, e.g.\, optimization-bas
 ed bound tightening captures dependencies within a partition.\n\nThis join
 t work with Calvin Tsay\, Jan Kronqvist\, Alexander Thebelt is based on tw
 o papers: https://arxiv.org/abs/2102.04373  & https://arxiv.org/abs/2101.1
 2708\n
LOCATION:https://stable.researchseminars.org/talk/MPML/47/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mathieu Blondel (Google Research\, Brain team\, Paris)
DTSTART:20210604T130000Z
DTEND:20210604T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/48
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 48/">Efficient and Modular Implicit Differentiation</a>\nby Mathieu Blonde
 l (Google Research\, Brain team\, Paris) as part of Mathematics\, Physics 
 and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nAutomatic differentiati
 on (autodiff) has revolutionized machine learning. It allows expressing co
 mplex computations by composing elementary ones in creative ways and remov
 es the burden of computing their derivatives by hand. More recently\, diff
 erentiation of optimization problem solutions has attracted widespread att
 ention with applications such as optimization as a layer\, and in bi-level
  problems such as hyper-parameter optimization and meta-learning. However\
 , the formulas for these derivatives often involve case-by-case tedious ma
 thematical derivations. In this work\, we propose a unified\, efficient an
 d modular approach for implicit differentiation of optimization problems. 
 In our approach\, the user defines (in Python in the case of our implement
 ation) a function F capturing the optimality conditions of the problem to 
 be differentiated. Once this is done\, we leverage autodiff of F and impli
 cit differentiation to automatically differentiate the optimization proble
 m. Our approach thus combines the benefits of implicit differentiation and
  autodiff. It is efficient as it can be added on top of any state-of-the-a
 rt solver and modular as the optimality condition specification is decoupl
 ed from the implicit differentiation mechanism. We show that seemingly sim
 ple principles allow to recover many recently proposed implicit differenti
 ation methods and create new ones easily. We demonstrate the ease of formu
 lating and solving bi-level optimization problems using our framework. We 
 also showcase an application to the sensitivity analysis of molecular dyna
 mics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/48/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Yuejie Chi (Carnegie Mellon University)
DTSTART:20210625T130000Z
DTEND:20210625T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/49
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 49/">Policy Optimization in Reinforcement Learning: A Tale of Precondition
 ing and Regularization</a>\nby Yuejie Chi (Carnegie Mellon University) as 
 part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbs
 tract\nPolicy optimization\, which learns the policy of interest by maximi
 zing the value function via large-scale optimization techniques\, lies at 
 the heart of modern reinforcement learning (RL). In addition to value maxi
 mization\, other practical considerations arise commonly as well\, includi
 ng the need of encouraging exploration\, and that of ensuring certain stru
 ctural properties of the learned policy due to safety\, resource and opera
 tional constraints. These considerations can often be accounted for by res
 orting to regularized RL\, which augments the target value function with a
  structure-promoting regularization term\, such as Shannon entropy\, Tsall
 is entropy\, and log-barrier functions. Focusing on an infinite-horizon di
 scounted Markov decision process\, this talk first shows that entropy-regu
 larized natural policy gradient methods converge globally at a linear conv
 ergence that is near independent of the dimension of the state-action spac
 e. Next\, a generalized policy mirror descent algorithm is proposed to acc
 ommodate a general class of convex regularizers beyond Shannon entropy. En
 couragingly\, this general algorithm inherits similar convergence guarante
 es\, even when the regularizer lacks strong convexity and smoothness. Our 
 results accommodate a wide range of learning rates\, and shed light upon t
 he role of regularization in enabling fast convergence in RL.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/49/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ard Louis (University of Oxford)
DTSTART:20210702T130000Z
DTEND:20210702T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/50
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 50/">Deep neural networks have an inbuilt Occam's razor</a>\nby Ard Louis 
 (University of Oxford) as part of Mathematics\, Physics and Machine Learni
 ng (IST\, Lisbon)\n\n\nAbstract\nOne of the most surprising properties of 
 deep neural networks (DNNs) is that they perform best in the overparameter
 ized regime. We are taught early on that having more parameters than data 
 points is a terrible idea. So why do DNNs work so well in a regime where c
 lassical learning theory predicts they should heavily overfit? By adapting
  the coding theorem from algorithmic information theory (which every physi
 cist should learn about!) we show that DNNs are exponentially biased at in
 itialisation to functions that have low descriptional (Kolmogorov) complex
 ity. In other words\, DNNs have an inbuilt Occam's razor\, a bias towards 
 simple functions. We next show that stochastic gradient descent (SGD)\, th
 e most popular optimisation method for DNNs\, follows the same bias\, and 
 so does not itself explain the good generalisation of DNNs. Our approach n
 aturally leads to a marginal-likelihood PAC-Bayes generalisation bound whi
 ch performs better than any other bounds on the market. Finally\, we discu
 ss why this bias towards simplicity allows DNNs to perform so well\, and s
 peculate on what this may tell us about the natural world.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/50/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Usman Khan (Tufts University)
DTSTART:20210709T130000Z
DTEND:20210709T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/51
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 51/">Distributed ML: Optimal algorithms for distributed stochastic non-con
 vex optimization</a>\nby Usman Khan (Tufts University) as part of Mathemat
 ics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIn many e
 merging applications\, it is of paramount interest to learn hidden paramet
 ers from data. For example\, self-driving cars may use onboard cameras to 
 identify pedestrians\, highway lanes\, or traffic signs in various light a
 nd weather conditions. Problems such as these can be framed as classificat
 ion\, regression\, or risk minimization in general\, at the heart of which
  lies stochastic optimization and machine learning. In many practical scen
 arios\, distributed and decentralized learning methods are preferable as t
 hey benefit from a divide-and-conquer approach towards data at the expense
  of local (short-range) communication. In this talk\, I will present our r
 ecent work that develops a novel algorithmic framework to address various 
 aspects of decentralized stochastic first-order optimization methods for n
 on-convex problems. A major focus will be to characterize regimes where de
 centralized solutions outperform their centralized counterparts and lead t
 o optimal convergence guarantees. Moreover\, I will characterize certain d
 esirable attributes of decentralized methods in the context of linear spee
 dup and networkindependent convergence rates. Throughout the talk\, I will
  demonstrate such key aspects of the proposed methods with the help of pro
 vable theoretical results and numerical experiments on real data.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/51/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Simon Du (University of Washington)
DTSTART:20210728T160000Z
DTEND:20210728T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/52
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 52/">Provable Representation Learning</a>\nby Simon Du (University of Wash
 ington) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbo
 n)\n\n\nAbstract\nRepresentation learning has been widely used in many app
 lications. In this talk\, I will present our work\, which uncovers when an
 d why representation learning provably improves the sample efficiency\, fr
 om a statistical learning point of view. I will show 1) the existence of a
  good representation among all tasks\, and 2) the diversity of tasks are k
 ey conditions that permit improved statistical efficiency via multi-task r
 epresentation learning. These conditions provably improve the sample effic
 iency for functions with certain complexity measures as the representation
 . If time permits\, I will also talk about leveraging the theoretical insi
 ghts to improve practical performance.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/52/
END:VEVENT
BEGIN:VEVENT
SUMMARY:J. Nathan Kutz (University of Washington)
DTSTART:20210916T160000Z
DTEND:20210916T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/53
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 53/">Deep learning for the discovery of parsimonious physics models</a>\nb
 y J. Nathan Kutz (University of Washington) as part of Mathematics\, Physi
 cs and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nA major challenge in
  the study of dynamical systems is that of model discovery: turning data i
 nto reduced order models that are not just predictive\, but provide insigh
 t into the nature of the underlying dynamical system that generated the da
 ta. We introduce a number of data-driven strategies for discovering nonlin
 ear multiscale dynamical systems and their embeddings from data. We consid
 er two canonical cases: (i) systems for which we have full measurements of
  the governing variables\, and (ii) systems for which we have incomplete m
 easurements. For systems with full state measurements\, we show that the r
 ecent sparse identification of nonlinear dynamical systems (SINDy) method 
 can discover governing equations with relatively little data and introduce
  a sampling method that allows SINDy to scale efficiently to problems with
  multiple time scales\, noise and parametric dependencies.   For systems w
 ith incomplete observations\, we show that the Hankel alternative view of 
 Koopman (HAVOK) method\, based on time-delay embedding coordinates and the
  dynamic mode decomposition\, can be used to obtain a linear models and Ko
 opman invariant measurement systems that nearly perfectly captures the dyn
 amics of nonlinear quasiperiodic systems. Neural networks are used in targ
 eted ways to aid in the model reduction process. Together\, these approach
 es provide a suite of mathematical strategies for reducing the data requir
 ed to discover and model nonlinear multiscale systems.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/53/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Leong Chuan Kwek (Nanyang Technological University\, Singapore)
DTSTART:20210923T090000Z
DTEND:20210923T100000Z
DTSTAMP:20260404T094311Z
UID:MPML/54
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 54/">Machine Learning and Quantum Technology</a>\nby Leong Chuan Kwek (Nan
 yang Technological University\, Singapore) as part of Mathematics\, Physic
 s and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe rise of machine l
 earning in recent times has remarkably transformed science and society. Th
 e goal of machine learning is to get computers to act without being explic
 itly programmed. Machine learning with deep reinforcement learning (RL) wa
 s recently recognized as a powerful tool to engineer dynamics in quantum s
 ystem. Also\, recently there has been some interest to exploit and leverag
 e the limited available quantum resources for performing classically chall
 enging tasks with noisy intermediate-scale quantum (NISQ) computers. Here\
 , we discuss some of our recent work on the applications of machine learni
 ng to quantum systems.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/54/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Constantino Tsallis (Group of Statistical Physics\, CBPF and Santa
  Fe Institute)
DTSTART:20211021T160000Z
DTEND:20211021T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/55
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 55/">Statistical mechanics for complex systems</a>\nby Constantino Tsallis
  (Group of Statistical Physics\, CBPF and Santa Fe Institute) as part of M
 athematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nTo
 gether with Newtonian mechanics\, Maxwell electromagnetism\, Einstein rela
 tivity and quantum mechanics\, Boltzmann-Gibbs (BG) statistical mechanics 
 constitutes one of the pillars of contemporary theoretical physics\, with 
 uncountable applications in science and technology. This theory applies fo
 rmidably well to a plethora of physical systems. Still\, it fails in the r
 ealm of complex systems\, characterized by generically strong space-time e
 ntanglement of their elements. On the basis of a nonadditive entropy (defi
 ned by an index q\, which recovers\, for q=1\, the celebrated Boltzmann-Gi
 bbs-von Neumann-Shannon entropy)\, it is possible to generalize the BG the
 ory. We will briefly review the foundations and applications in natural\, 
 artificial and social systems.\n\nA Bibliography is available at http://ts
 allis.cat.cbpf.br/biblio.htm\n
LOCATION:https://stable.researchseminars.org/talk/MPML/55/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Volkan Cevher (Laboratory for Information and Inference Systems 
 – LIONS\, EPFL)
DTSTART:20210930T160000Z
DTEND:20210930T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/56
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 56/">Optimization Challenges in Adversarial Machine Learning</a>\nby Volka
 n Cevher (Laboratory for Information and Inference Systems – LIONS\, EPF
 L) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\
 n\nAbstract\nThanks to neural networks (NNs)\, faster computation\, and ma
 ssive datasets\, machine learning (ML) is under increasing pressure to pro
 vide automated solutions to even harder real-world tasks beyond human perf
 ormance with ever faster response times due to potentially huge technologi
 cal and societal benefits. Unsurprisingly\, the NN learning formulations p
 resent a fundamental challenge to the back-end learning algorithms despite
  their scalability\, in particular due to the existence of traps in the no
 n-convex optimization landscape\, such as saddle points\, that can prevent
  algorithms from obtaining “good” solutions.\n\nIn this talk\, we desc
 ribe our recent research that has demonstrated that the non-convex optimiz
 ation dogma is false by showing that scalable stochastic optimization algo
 rithms can avoid traps and rapidly obtain locally optimal solutions. Coupl
 ed with the progress in representation learning\, such as over-parameteriz
 ed neural networks\, such local solutions can be globally optimal.\n\nUnfo
 rtunately\, this talk will also demonstrate that the central min-max optim
 ization problems in ML\, such as generative adversarial networks (GANs)\, 
 robust reinforcement learning (RL)\, and\ndistributionally robust ML\, con
 tain spurious attractors that do not include any stationary points of the 
 original learning formulation. Indeed\, we will describe how algorithms ar
 e subject to a grander challenge\, including unavoidable convergence failu
 res\, which could explain the stagnation in their progress despite the imp
 ressive earlier demonstrations.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/56/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Clément Hongler (EPFL)
DTSTART:20211014T160000Z
DTEND:20211014T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/57
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 57/">Neural Tangent Kernel</a>\nby Clément Hongler (EPFL) as part of Math
 ematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe N
 eural Tangent Kernel is a new way to understand the gradient descent in de
 ep neural networks\, connecting them with kernel methods. In this talk\, I
 'll introduce this formalism and give a number of results on the Neural Ta
 ngent Kernel and explain how they give us insight into the dynamics of neu
 ral networks during training and into their generalization features.\n\nBa
 sed off joint works with Arthur Jacot and Franck Gabriel.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/57/
END:VEVENT
BEGIN:VEVENT
SUMMARY:George Em Karniadakis (Brown University)
DTSTART:20211104T170000Z
DTEND:20211104T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/58
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 58/">Operator regression via DeepOnet: Theory\, Algorithms and Application
 s</a>\nby George Em Karniadakis (Brown University) as part of Mathematics\
 , Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWe will revie
 w physics-informed neural network and summarize available theoretical resu
 lts. We will also introduce new NNs that learn functionals and nonlinear o
 perators from functions and corresponding responses for system identificat
 ion. The universal approximation theorem of operators is suggestive of the
  potential of NNs in learning from scattered data any continuous operator 
 or complex system. We first generalize the theorem to deep neural networks
 \, and subsequently we apply it to design a new composite NN with small ge
 neralization error\, the deep operator network (DeepONet)\, consisting of 
 a NN for encoding the discrete input function space (branch net) and anoth
 er NN for encoding the domain of the output functions (trunk net). We demo
 nstrate that DeepONet can learn various explicit operators\, e.g.\, integr
 als\, Laplace transforms and fractional Laplacians\, as well as implicit o
 perators that represent deterministic and stochastic differential equation
 s. More generally\, DeepOnet can learn multiscale operators spanning acros
 s many scales and trained by diverse sources of data simultaneously.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/58/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Michael Arbel (INRIA Grenoble Rhône-Alpes)
DTSTART:20211111T170000Z
DTEND:20211111T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/59
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 59/">Annealed Flow Transport Monte Carlo</a>\nby Michael Arbel (INRIA Gren
 oble Rhône-Alpes) as part of Mathematics\, Physics and Machine Learning (
 IST\, Lisbon)\n\n\nAbstract\nAnnealed Importance Sampling (AIS) and its Se
 quential Monte Carlo (SMC) extensions are state-of-the-art methods for est
 imating normalizing constants of probability distributions. We propose her
 e a novel Monte Carlo algorithm\, Annealed Flow Transport (AFT)\, that bui
 lds upon AIS and SMC and combines them with normalizing flows (NF) for imp
 roved performance. This method transports a set of particles using not onl
 y importance sampling (IS)\, Markov chain Monte Carlo (MCMC) and resamplin
 g steps - as in SMC\, but also relies on NF which are learned sequentially
  to push particles towards the successive annealed targets. We provide lim
 it theorems for the resulting Monte Carlo estimates of the normalizing con
 stant and expectations with respect to the target distribution. Additional
 ly\, we show that a continuous-time scaling limit of the population versio
 n of AFT is given by a Feynman--Kac measure which simplifies to the law of
  a controlled diffusion for expressive NF. We demonstrate experimentally t
 he benefits and limitations of our methodology on a variety of application
 s.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/59/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Soledad Villar (Mathematical Institute for Data Science at Johns H
 opkins University)
DTSTART:20211202T170000Z
DTEND:20211202T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/60
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 60/">Equivariant machine learning structure like classical physics</a>\nby
  Soledad Villar (Mathematical Institute for Data Science at Johns Hopkins 
 University) as part of Mathematics\, Physics and Machine Learning (IST\, L
 isbon)\n\n\nAbstract\nThere has been enormous progress in the last few yea
 rs in designing neural networks that respect the fundamental symmetries an
 d coordinate freedoms of physical law. Some of these frameworks make use o
 f irreducible representations\, some make use of high-order tensor objects
 \, and some apply symmetry-enforcing constraints. Different physical laws 
 obey different combinations of fundamental symmetries\, but a large fracti
 on (possibly all) of classical physics is equivariant to translation\, rot
 ation\, reflection (parity)\, boost (relativity)\, and permutations. Here 
 we show that it is simple to parameterize universally approximating polyno
 mial functions that are equivariant under these symmetries\, or under the 
 Euclidean\, Lorentz\, and Poincare groups\, at any dimensionality d. The k
 ey observation is that nonlinear O(d)-equivariant (and related-group-equiv
 ariant) functions can be expressed in terms of a lightweight collection of
  scalars---scalar products and scalar contractions of the scalar\, vector\
 , and tensor inputs. These results demonstrate theoretically that gauge-in
 variant deep learning models for classical physics with good scaling for l
 arge problems are feasible right now.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/60/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Pier Luigi Dragotti (Department of Electrical and Electronic Engin
 eering\, Imperial College\, London)
DTSTART:20211209T170000Z
DTEND:20211209T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/61
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 61/">Computational Imaging for Art investigation and for Neuroscience</a>\
 nby Pier Luigi Dragotti (Department of Electrical and Electronic Engineeri
 ng\, Imperial College\, London) as part of Mathematics\, Physics and Machi
 ne Learning (IST\, Lisbon)\n\n\nAbstract\nThe revolution in sensing\, with
  the emergence of many new imagingtechniques\, offers the possibility of g
 aining unprecedented access tothe physical world\, but this revolution can
  only bear fruit through the skilful interplay between the physical and co
 mputational worlds. This is the domain of computational imaging which advo
 cates that\, to develop effective imaging systems\, it will be necessary t
 o go beyond the traditional decoupled imaging pipeline where device physic
 s\, image processing and the end-user application are considered separatel
 y. Instead\, we need to rethink imaging as an integrated sensing and infer
 ence model. In this talk we cover two research areas where computational i
 maging is likely to have an impact.\n\nWe first focus on the heritage sect
 or which is experiencing a digital revolution driven in part by the increa
 sing use of non-invasive\, non-destructive imaging techniques. These new i
 maging methods provide a way to capture information about an entire painti
 ng and can give us information about features at or below the surface of t
 he painting. We focus on Macro X-Ray Fluorescence (XRF) scanning which is 
 a technique for the mapping of chemical elements in paintings. After descr
 ibing in broad terms the working of this device\, a method that can proces
 s XRF scanning data from paintings is introduced. The method is based on c
 onnecting the problem of extracting elemental maps in XRF data to Prony's 
 method\, a technique broadly used in engineering to estimate frequencies o
 f a sum of sinusoids. The results presented show the ability of our method
  to detect and separate weak signals related to hidden chemical elements i
 n the paintings. We then discuss results on the Leonardo's "The Virgin of 
 the Rocks" and show that our algorithm is able to reveal\, more clearly th
 an ever before\, the hidden drawings of a previous composition that Leonar
 do then abandoned for the painting that we can now see.\n\nIn the second p
 art of the talk\, we focus on two-photon microscopy and neuroscience. To u
 nderstand how networks of neurons process information\, it is essential to
  monitor their activity in living tissue. Multi-photon microscopy is unpar
 alleled in its ability to image cellular activity and neural circuits\, de
 ep in living tissue\, at single-cell resolution. However\, in order to ach
 ieve step changes in our understanding of brain function\, large-scale ima
 ging studies of neural populations are needed and this can be achieved onl
 y by developing computational tools that can enhance the quality of the da
 ta acquired and can scan 3-D volumes quickly. In this talk we introduce li
 ght-field microscopy and present a method to localize neurons in 3-D. The 
 method is based on the use of proper sparsity priors\, novel optimization 
 strategies and machine learning.\n\n\nThis is joint work with A. Foust\, P
 . Song\, C. Howe\, H. Verinaz\, J. Huang and Y.Su from Imperial College Lo
 ndon\, and C. Higgitt and N. Daly from The National Gallery in London\n
LOCATION:https://stable.researchseminars.org/talk/MPML/61/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Suman Ravuri (DeepMind)
DTSTART:20211125T170000Z
DTEND:20211125T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/62
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 62/">Skilful precipitation nowcasting using deep generative models of rada
 r</a>\nby Suman Ravuri (DeepMind) as part of Mathematics\, Physics and Mac
 hine Learning (IST\, Lisbon)\n\n\nAbstract\nPrecipitation nowcasting\, the
  high-resolution forecasting of precipitation up to two hours ahead\, supp
 orts the real-world socioeconomic needs of many sectors reliant on weather
 -dependent decision-making. State-of-the-art operational nowcasting method
 s typically advect precipitation fields with radar-based wind estimates\, 
 and struggle to capture important non-linear events such as convective ini
 tiations. Recently introduced deep learning methods use radar to directly 
 predict future rain rates\, free of physical constraints. While they accur
 ately predict low-intensity rainfall\, their operational utility is limite
 d because their lack of constraints produces blurry nowcasts at longer lea
 d times\, yielding poor performance on rarer medium-to-heavy rain events. 
 Here we present a deep generative model for the probabilistic nowcasting o
 f precipitation from radar that addresses these challenges. Using statisti
 cal\, economic and cognitive measures\, we show that our method provides i
 mproved forecast quality\, forecast consistency and forecast value. Our mo
 del produces realistic and spatiotemporally consistent predictions over re
 gions up to 1\,536 km × 1\,280 km and with lead times from 5–90
  min ahead. Using a systematic evaluation by more than 50 expert meteoro
 logists\, we show that our generative model ranked first for its accuracy 
 and usefulness in 89% of cases against two competitive methods. When verif
 ied quantitatively\, these nowcasts are skillful without resorting to blur
 ring. We show that generative nowcasting can provide probabilistic predict
 ions that improve forecast value and support operational utility\, and at 
 resolutions and lead times where alternative methods struggle.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/62/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Dan Roberts (MIT\, Center for Theoretical Physics)
DTSTART:20220113T170000Z
DTEND:20220113T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/63
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 63/">The Principles of Deep Learning Theory</a>\nby Dan Roberts (MIT\, Cen
 ter for Theoretical Physics) as part of Mathematics\, Physics and Machine 
 Learning (IST\, Lisbon)\n\n\nAbstract\nDeep learning is an exciting approa
 ch to modern artificial intelligence based on artificial neural networks. 
 The goal of this talk is to provide a blueprint — using tools from physi
 cs — for theoretically analyzing deep neural networks of practical relev
 ance. This task will encompass both understanding the statistics of initia
 lized deep networks and determining the training dynamics of such an ensem
 ble when learning from data.\n\nThis talk is based on a book\, <a href="ht
 tps://arxiv.org/pdf/2106.10165.pdf">"The Principles of Deep Learning Theor
 y\,"</a> co-authored with Sho Yaida and based on research also in collabor
 ation with Boris Hanin. It will be published next year by Cambridge Univer
 sity Press.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/63/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Anders Hansen (Faculty of Mathematics and Department of Applied Ma
 thematics and Theoretical Physics\, University of Cambridge)
DTSTART:20220120T170000Z
DTEND:20220120T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/64
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 64/">Why things don’t work — On the extended Smale's 9th and 18th prob
 lems (the limits of AI) and methodological barriers</a>\nby Anders Hansen 
 (Faculty of Mathematics and Department of Applied Mathematics and Theoreti
 cal Physics\, University of Cambridge) as part of Mathematics\, Physics an
 d Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe alchemists wanted to 
 create gold\, Hilbert wanted an algorithm to solve Diophantine equations\,
  researchers want to make deep learning robust in AI\, MATLAB wants (but f
 ails) to detect when it provides wrong solutions to linear programs etc. W
 hy does one not succeed in so many of these fundamental cases? The reason 
 is typically methodological barriers. The history of science is full of me
 thodological barriers — reasons for why we never succeed in reaching cer
 tain goals. In many cases\, this is due to the foundations of mathematics.
  We will present a new program on methodological barriers and foundations 
 of mathematics\, where — in this talk — we will focus on two basic pro
 blems: (1) The instability problem in deep learning: Why do researchers fa
 il to produce stable neural networks in basic classification and computer 
 vision problems that can easily be handled by humans — when one can prov
 e that there exist stable and accurate neural networks? Moreover\, AI algo
 rithms can typically not detect when they are wrong\, which becomes a seri
 ous issue when striving to create trustworthy AI. The problem is more gene
 ral\, as for example MATLAB's linprog routine is incapable of certifying c
 orrect solutions of basic linear programs. Thus\, we’ll address the foll
 owing question: (2) Why are algorithms (in AI and computations in general)
  incapable of determining when they are wrong? These questions are deeply 
 connected to the extended Smale’s 9th and 18th problems on the list of m
 athematical problems for the 21st century.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/64/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Joosep Pata (National Institute of Chemical Physics and Biophysics
 \, Estonia)
DTSTART:20220203T170000Z
DTEND:20220203T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/65
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 65/">Machine learning for data reconstruction at the LHC</a>\nby Joosep Pa
 ta (National Institute of Chemical Physics and Biophysics\, Estonia) as pa
 rt of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstr
 act\nPhysics analyses at the CERN experiments rely on detector hits being 
 interpreted or reconstructed as particle candidates. The data reconstructi
 on systems are built on decades of physics and detector knowledge and must
  operate reliably on petabytes of data in diverse computing centers spread
  around the world. In the recent years\, machine learning (ML) is playing 
 an increasingly important role at the LHC experiments for reconstructing a
 nd interpreting the data\, from calibrating the detector readouts to the f
 inal interpretation for complex signal processes. We will discuss the vari
 ous aspects of ML at the LHC experiments\, focusing on data reconstruction
  and particle identification approaches using modern machine learning meth
 ods such as graph neural networks. We will bring a concrete detailed examp
 le from machine learned particle flow (MLPF)\, an R&D effort to develop a 
 fully optimizable particle flow reconstruction across detector subsystems 
 in CMS.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/65/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Jan Kieseler (European Organization for Nuclear Research (CERN))
DTSTART:20220303T170000Z
DTEND:20220303T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/66
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 66/">The MODE project</a>\nby Jan Kieseler (European Organization for Nucl
 ear Research (CERN)) as part of Mathematics\, Physics and Machine Learning
  (IST\, Lisbon)\n\n\nAbstract\nThe effective design of instruments that re
 ly on the interaction of radiation with matter for their operation is a co
 mplex task. Furthermore\, the underlying physics processes are intrinsical
 ly stochastic in nature and open a vast space of possible choices for the 
 physical characteristics of the instrument. While even large scale detecto
 rs such as e.g. at the LHC are built using surrogates for the ultimate phy
 sics objective\, the MODE Collaboration (an acronym for Machine-learning O
 ptimized Design of Experiments) aims at developing tools also based on dee
 p learning techniques to achieve end-to-end optimization of the design of 
 instruments via a fully differentiable pipeline capable of exploring the P
 areto-optimal frontier of the utility function for future particle collide
 r experiments and related detectors. The construction of such a differenti
 able model requires inclusion of information-extraction procedures\, inclu
 ding data collection\, detector response\, pattern recognition\, and other
  existing constraints such as cost. This talk will give an introduction to
  the goals of the newly founded MODE collaboration and highlight some of t
 he already existing ingredients.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/66/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Fernando E. Rosas (Faculty of Medicine\, Department of Brain Scien
 ces\, Imperial College)
DTSTART:20220324T170000Z
DTEND:20220324T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/67
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 67/">Towards a deeper understanding of high-order interdependencies in com
 plex systems</a>\nby Fernando E. Rosas (Faculty of Medicine\, Department o
 f Brain Sciences\, Imperial College) as part of Mathematics\, Physics and 
 Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWe live in an increasingly 
 interconnected world and\, unfortunately\, our understanding of interdepen
 dency is still limited. As a matter of fact\, while bivariated relationshi
 ps are at the core of most of our data analysis methods\, there is still n
 o principled theory to account for the different types of interactions tha
 t can occur between three or more variables. This talk explores the vast a
 nd largely unexplored territory of multivariate complexity\, and discusses
  information-theoretic approaches that have been introduced to fill this i
 mportant knowledge gap.\n\nThe first part of the talk is devoted to synerg
 istic phenomena\, which correspond to statistical regularities that affect
  the whole but not the parts. We explain how synergy can be effectively ca
 ptured by information-theoretic measures inspired in the nature of high br
 ain functions\, and how these measures allow us to map complex interdepend
 encies into hypergraphs. The second part of the talk focuses on a new theo
 ry of what constitutes causal emergence\, and how it can be measured from 
 time series data. This theory enables a formal\, quantitative account of d
 ownward causation\, and introduces “causal decoupling” as a complement
 ary modality of emergence. Importantly\, this not only establishes concept
 ual tools to frame conjectures about emergence rigorously\, but also provi
 des practical procedures to test them on data. We illustrate the considere
 d analysis tools on different case studies\, including cellular automata\,
  baroque music\, flocking models\, and neuroimaging datasets.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/67/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Josef Urban (Czech Institute of of Informatics\, Robotics and Cybe
 rnetics (CIIRC))
DTSTART:20220331T160000Z
DTEND:20220331T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/68
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 68/">Machine Learning and Theorem Proving</a>\nby Josef Urban (Czech Insti
 tute of of Informatics\, Robotics and Cybernetics (CIIRC)) as part of Math
 ematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe t
 alk will describe several ways in which machine learning is combined with 
 theorem proving today over large corpora of formal proof. If time permits\
 , I will also show some demos of the systems and mention related topics su
 ch as ML-guided conjecturing and autoformalization.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/68/
END:VEVENT
BEGIN:VEVENT
SUMMARY:André F. T. Martins (Instituto Superior Técnico)
DTSTART:20220224T163000Z
DTEND:20220224T173000Z
DTSTAMP:20260404T094311Z
UID:MPML/69
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 69/">From Sparse Modeling to Sparse Communication</a>\nby André F. T. Mar
 tins (Instituto Superior Técnico) as part of Mathematics\, Physics and Ma
 chine Learning (IST\, Lisbon)\n\n\nAbstract\nNeural networks and other mac
 hine learning models compute continuous representations\, while humans com
 municate mostly through discrete symbols. Reconciling these two forms of c
 ommunication is desirable for generating human-readable interpretations or
  learning discrete latent variable models\, while maintaining end-to-end d
 ifferentiability.\n\nIn the first part of the talk\, I will describe how s
 parse modeling techniques can be extended and adapted for facilitating spa
 rse communication in neural models. The building block is a family of spar
 se transformations called alpha-entmax\, a drop-in replacement for softmax
 \, which contains sparsemax as a particular case. Entmax transformations a
 re differentiable and (unlike softmax) they can return sparse probability 
 distributions\, useful to build interpretable attention mechanisms. Varian
 ts of these sparse transformations have been applied with success to machi
 ne translation\, natural language inference\, visual question answering\, 
 and other tasks.\n\nIn the second part\, I will introduce mixed random var
 iables\, which are in-between the discrete and continuous worlds. We build
  rigorous theoretical foundations for these hybrids\, via a new “direct 
 sum” base measure defined on the face lattice of the probability simplex
 . From this measure\, we introduce new entropy and Kullback-Leibler diverg
 ence functions that subsume the discrete and differential cases and have i
 nterpretations in terms of code optimality. Our framework suggests two str
 ategies for representing and sampling mixed random variables\, an extrinsi
 c (“sample-and-project”) and an intrinsic one (based on face stratific
 ation).\n\nIn the third part\, I will show how sparse transformations can 
 also be used to design new loss functions\, replacing the cross-entropy lo
 ss. To this end\, I will introduce the family of Fenchel-Young losses\, re
 vealing connections between generalized entropy regularizers and separatio
 n margin. I will illustrate with applications in natural language generati
 on\, morphology\, and machine translation.\n\nThis work was funded by the 
 DeepSPIN ERC project - https://deep-spin.github.io\n
LOCATION:https://stable.researchseminars.org/talk/MPML/69/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Dmitry Krotov (Watson AI Lab and IBM Research in Cambridge)
DTSTART:20220414T160000Z
DTEND:20220414T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/70
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 70/">Modern Hopfield Networks in AI and Neurobiology</a>\nby Dmitry Krotov
  (Watson AI Lab and IBM Research in Cambridge) as part of Mathematics\, Ph
 ysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p>Modern Hopfiel
 d Networks or Dense Associative Memories are recurrent neural networks wit
 h fixed point attractor states that are described by an energy function. I
 n contrast to conventional Hopfield Networks\, their modern versions have 
 a very large memory storage capacity\, which makes them appealing tools fo
 r many problems in machine learning and cognitive and neuro-sciences. In t
 his talk I will introduce an intuition and a mathematical formulation of t
 his class of models\, and will give examples of problems in AI that can be
  tackled using these new ideas. I will also explain how different individu
 al models of this class (e.g. hierarchical memories\, attention mechanism 
 in transformers\, etc.) arise from their general mathematical formulation 
 with the Lagrangian functions.</p>\n\n<p><strong>References:</strong></p>\
 n\n<ol>\n	<li><a href="https://arxiv.org/abs/1606.01164">D.Krotov\, J.Hopf
 ield\, "Dense associative memory for pattern recognition"</a></li>\n	<li><
 a href="https://arxiv.org/abs/2008.06996">D.Krotov\, J.Hopfield\, "Large A
 ssociative Memory Problem in Neurobiology and Machine Learning</a>"</li>\n
 	<li><a href="https://arxiv.org/abs/1702.01929">M.Demircigil\, et al.\, "O
 n a model of associative memory with huge storage capacity"</a></li>\n	<li
 ><a href="https://arxiv.org/abs/2008.02217">H.Ramsauer\, et al.\, "Hopfiel
 d Networks is All You Need"</a></li>\n	<li><a href="https://arxiv.org/abs/
 2107.06446">D.Krotov\, "Hierarchical Associative Memory"</a></li>\n</ol>\n
 \n<p>&nbsp\;</p>\n
LOCATION:https://stable.researchseminars.org/talk/MPML/70/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Emtiyaz Khan (RIKEN-AIP\, Tokyo and OIST\, Okinawa\, Japan)
DTSTART:20220428T090000Z
DTEND:20220428T100000Z
DTSTAMP:20260404T094311Z
UID:MPML/71
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 71/">The Bayesian Learning Rule for Adaptive AI</a>\nby Emtiyaz Khan (RIKE
 N-AIP\, Tokyo and OIST\, Okinawa\, Japan) as part of Mathematics\, Physics
  and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nHumans and animals hav
 e a natural ability to autonomously learn and quickly adapt to their surro
 undings. How can we design AI systems that do the same? In this talk\, I w
 ill present Bayesian principles to bridge such gaps between humans and AI.
  I will show that a wide variety of machine-learning algorithms are instan
 ces of a single learning-rule called the Bayesian learning rule. The rule 
 unravels a dual perspective yielding new adaptive mechanisms for machine-l
 earning based AI systems. My hope is to convince the audience that Bayesia
 n principles are indispensable for an AI that learns as efficiently as we 
 do.\n\n<p><strong>Reference: </strong>M.E. Khan\, H. Rue\, The Bayesian Le
 arning Rule [<a href="https://arxiv.org/abs/2107.04562" rel="noreferrer" t
 arget="_blank">arXiv</a>] [<a href="https://twitter.com/EmtiyazKhan/status
 /1414498922584711171?s=20" rel="noreferrer" target="_blank">Tweet</a>]</p>
 \n
LOCATION:https://stable.researchseminars.org/talk/MPML/71/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rianne van den Berg (Microsoft Research Amsterdam)
DTSTART:20220421T160000Z
DTEND:20220421T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/72
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 72/">Generative models for discrete random variables</a>\nby Rianne van de
 n Berg (Microsoft Research Amsterdam) as part of Mathematics\, Physics and
  Machine Learning (IST\, Lisbon)\n\n\nAbstract\nn this talk I will discuss
  how different classes of generative models can be adapted to handle discr
 ete random variables\, and how this can be used to connect generative mode
 ls to downstream tasks such as lossless compression. I will start by discu
 ssing normalizing flow models\, and the challenges that arise when convert
 ing these models that are typically designed for real-valued random variab
 les to discrete random variables. Next\, I will demonstrate how denoising 
 diffusion models with discrete state spaces have a rich design space in te
 rms of the noising process\, and how this influences the performance of th
 e learned denoising model. Finally\, I will show how denoising diffusion m
 odels can be connected to autoregressive models\, and introduce an autoreg
 ressive model with a random generation order.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/72/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Andrea L. Bertozzi (University of California Los Angeles)
DTSTART:20220505T160000Z
DTEND:20220505T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/73
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 73/">Graph based models in semi-supervised and unsupervised learning</a>\n
 by Andrea L. Bertozzi (University of California Los Angeles) as part of Ma
 thematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nSim
 ilarity graphs provide a structure for analyzing high dimensional
  data. These undirected weighted graphs provide structure for identifying
  inherent clusters in datasets and many methods exist to sort through such
  data building on the graph laplacian matrix.  One way to think about suc
 h problems is in terms of penalized cut problems.  These can be expressed
  in terms of the graph total variation which has a well-known analogue in 
 Euclidean space.  We show how to use ideas from geometric methods for PDE
 s to develop efficient and high performing methods for semi-supervised and
  unsupervised learning.  These methods also extend to active learning and
  to modularity optimization for community detection on networks.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/73/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Stanley Osher (Department of Mathematics\, University of Californi
 a\, Los Angeles)
DTSTART:20220519T160000Z
DTEND:20220519T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/74
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 74/">Conservation laws and generalized optimal transport</a>\nby Stanley O
 sher (Department of Mathematics\, University of California\, Los Angeles) 
 as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\n
 Abstract\nIn this talk\, we connect Lax’s entropy-entropy flux in conser
 vation laws with optimal transport type metric spaces. Following this conn
 ection\, we further design variational discretizations for conservation la
 ws and mean field control of conservation laws. In particular\, we design 
 unconditionally stable time discretization methods that are easy to implem
 ent.\n\nOn joint work with Siting Liu\, UCLA and Wuchen Li\, University of
  South Carolina.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/74/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Anja Butter (ITP\, University of Heidelberg)
DTSTART:20220602T160000Z
DTEND:20220602T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/75
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 75/">Machine Learning and LHC Event Generation</a>\nby Anja Butter (ITP\, 
 University of Heidelberg) as part of Mathematics\, Physics and Machine Lea
 rning (IST\, Lisbon)\n\n\nAbstract\nFirst-principle simulations are at the
  heart of the high-energy physics research program. They link the vast dat
 a output of multi-purpose detectors with fundamental theory predictions an
 d interpretation. In the coming LHC runs\, these simulations will face unp
 recedented precision requirements to match the experimental accuracy. New 
 ideas and tools based on neural networks have been developed at the interf
 ace of particle physics and machine learning. They can improve the speed a
 nd precision of forward simulations and handle the complexity of collision
  data. Such networks can be employed within established simulation tools o
 r as part of a new framework. Since neural networks can be inverted\, they
  open new avenues in LHC analyses.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/75/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Paulo Tabuada (UCLA)
DTSTART:20220609T160000Z
DTEND:20220609T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/77
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 77/">Deep neural networks\, universal approximation\, and geometric contro
 l</a>\nby Paulo Tabuada (UCLA) as part of Mathematics\, Physics and Machin
 e Learning (IST\, Lisbon)\n\n\nAbstract\nDeep neural networks have drastic
 ally changed the landscape of several engineering areas such as computer v
 ision and natural language processing. Notwithstanding the widespread succ
 ess of deep networks in these\, and many other areas\, it is still not wel
 l understood why deep neural networks work so well. In particular\, the qu
 estion of which functions can be learned by deep neural networks has remai
 ned unanswered.\nIn this talk we give an answer to this question for deep 
 residual neural networks\, a class of deep networks that can be interprete
 d as the time discretization of nonlinear control systems. We will show th
 at the ability of these networks to memorize training data can be expresse
 d through the control theoretic notion of controllability which can be pro
 ved using geometric control techniques. We then add an additional ingredie
 nt\, monotonicity\, to conclude that deep residual networks can approximat
 e\, to arbitrary accuracy with respect to the uniform norm\, any continuou
 s function on a compact subset of n-dimensional Euclidean space by using a
 t most n+1 neurons per layer. We will conclude the talk by showing how the
 se results pave the way for the use of deep networks in the perception pip
 eline of autonomous systems while providing formal (and probability free) 
 guarantees of stability and robustness.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/77/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Petar Veličković (DeepMind and University of Cambridge)
DTSTART:20220929T160000Z
DTEND:20220929T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/78
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 78/">Geometric Deep Learning: Grids\, Graphs\, Groups\, Geodesics and Gaug
 es</a>\nby Petar Veličković (DeepMind and University of Cambridge) as pa
 rt of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstr
 act\nThe last decade has witnessed an experimental revolution in data scie
 nce and machine learning\, epitomised by deep learning methods. Indeed\, m
 any high-dimensional learning tasks previously thought to be beyond reach 
 –such as computer vision\, playing Go\, or protein folding – are in fa
 ct feasible with appropriate computational scale. Remarkably\, the essence
  of deep learning is built from two simple algorithmic principles: first\,
  the notion of representation or feature learning\, whereby adapted\, ofte
 n hierarchical\, features capture the appropriate notion of regularity for
  each task\, and second\, learning by local gradient-descent type methods\
 , typically implemented as backpropagation.\n\nWhile learning generic func
 tions in high dimensions is a cursed estimation problem\, most tasks of in
 terest are not generic\, and come with essential pre-defined regularities 
 arising from the underlying low-dimensionality and structure of the physic
 al world. This talk is concerned with exposing these regularities through 
 unified geometric principles that can be applied throughout a wide spectru
 m of applications.\n\nSuch a 'geometric unification' endeavour in the spir
 it of Felix Klein's Erlangen Program serves a dual purpose: on one hand\, 
 it provides a common mathematical framework to study the most successful n
 eural network architectures\, such as CNNs\, RNNs\, GNNs\, and Transformer
 s. On the other hand\, it gives a constructive procedure to incorporate pr
 ior physical knowledge into neural architectures and provide principled wa
 y to build future architectures yet to be invented.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/78/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Yongji Wang (Department of Geosciences\, Princeton University)
DTSTART:20220526T160000Z
DTEND:20220526T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/79
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 79/">Physics-informed neural networks for solving 3-D Euler equation</a>\n
 by Yongji Wang (Department of Geosciences\, Princeton University) as part 
 of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract
 \nOne of the most challenging open questions in mathematical fluid dynamic
 s is whether an inviscid incompressible fluid\, described by the 3-dimensi
 onal Euler equations\, with initially smooth velocity and finite energy ca
 n develop singularities in finite time. This long-standing open problem is
  closely related to one of the seven Millennium Prize Problems which consi
 ders the problem the viscous analogue to the Euler equations (the Navier-S
 tokes equations). In this talk\, I will describe how we leverage the power
  of deep learning\, using deep neural networks with equation constraints\,
  namely physics-informed neural networks (PINNs)\, to find a smooth self-s
 imilar blow-up solution for the 3-dimensional Euler equations in the prese
 nce of a cylindrical boundary. To the best of our knowledge\, the solution
  represents the first example of a truly 2-D or higher dimensional backwar
 ds self-similar solution. This new numerical framework based on PINNs is s
 hown to be robust and readily adaptable to other fluid equations\, which s
 heds new light to the century-old mystery of capital importance in the fie
 ld of mathematical fluid dynamics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/79/
END:VEVENT
BEGIN:VEVENT
SUMMARY:John Baez (U.C. Riverside)
DTSTART:20220616T170000Z
DTEND:20220616T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/80
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 80/">Shannon Entropy from Category Theory</a>\nby John Baez (U.C. Riversid
 e) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\
 n\nAbstract\nShannon entropy is a powerful concept. But what properties si
 ngle out Shannon entropy as special? Instead of focusing on the entropy of
  a probability measure on a finite set\, it can help to focus on the "info
 rmation loss"\, or change in entropy\, associated with a measure-preservin
 g function. Shannon entropy then gives the only concept of information los
 s that is functorial\, convex-linear and continuous.\n\nThis is joint work
  with Tom Leinster and Tobias Fritz.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/80/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Dario Izzo (European Space Agency)
DTSTART:20220630T160000Z
DTEND:20220630T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/81
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 81/">Geodesy of irregular small bodies via neural density fields: geodesyN
 ets</a>\nby Dario Izzo (European Space Agency) as part of Mathematics\, Ph
 ysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe problem of de
 termining the density distribution of celestial bodies from the induced gr
 avitational pull is of great importance in astrophysics as well as space e
 ngineering (thinking of situations where spacecraft need to perform orbita
 l and surface proximity operations). Knowledge of a body density distribut
 ion provides also great insights on the body's origin and composition. In 
 practice\, the state-of-the-art approaches for modelling the gravity field
  of extended bodies are spherical harmonics models\, mascon models and pol
 yhedral gravity models. All of these\, however\, while being widely studie
 d and developed since the early works from Laplace\, introduce requirement
 s such as knowledge of a shape model\, assumption of a homogeneous interna
 l density\, being outside the\nBrillouin sphere\, etc...\n\n\nIn this talk
 \, we introduce and explain Neural Density Fields\, a new approach to repr
 esent the density of extended bodies and learn its accurate form inverting
  data from gravitational accelerations\, orbits or the gravity potential. 
 The resulting deep learning model\, called  geodesyNets is able to compete
  with classical approaches while solving most of their limitations. We als
 o introduce eclipseNets\, a deep learning model based on related ideas and
  able to learn the eclipse shadow cones of irregular bodies\, thus allowin
 g highly precise propagation and stability studies.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/81/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Audrey Durand (IID\, Université Laval\, Canada)
DTSTART:20220707T160000Z
DTEND:20220707T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/82
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 82/">Interactive learning for Neurosciences - Between Simulation and Reali
 ty</a>\nby Audrey Durand (IID\, Université Laval\, Canada) as part of Mat
 hematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nLear
 ning a behaviour to conduct a given task can be achieved by interacting wi
 th the the environment. This is the crux of reinforcement learning (RL)\, 
 where an (automated) agent learns to solve a problem through an iterative 
 trial-and-error process. More specifically\, an RL agent can interact with
  the environment and learn from these interactions by observing a feedback
  on the goal task. Therefore\, these methods typically require to be able 
 to intervene on the environment and make (possibly a very large number of)
  mistakes. Although this can be a limiting factor in some applications\, s
 imple RL settings\, such as bandit settings\, can still host a variety of 
 problems for interactively learning behaviours. In other situations\, simu
 lation might be the key.\n\nIn this talk\, we will show that RL can be use
 d to formulate and tackle data acquisition (imaging) problems in neuroscie
 nces. We will see how bandit methods can be used to optimize super-resolut
 ion imaging by learning on real devices through an actual empirical proces
 s. We will also see how simulation can be leveraged to learn more sequenti
 al decision making strategies. These applications highlight the potential 
 of RL to support expert users on difficult task and enable new discoveries
 .\n
LOCATION:https://stable.researchseminars.org/talk/MPML/82/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Joseph Bakarji (University of Washington)
DTSTART:20220714T160000Z
DTEND:20220714T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/83
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 83/">Dimensionally Consistent Learning with Buckingham Pi</a>\nby Joseph B
 akarji (University of Washington) as part of Mathematics\, Physics and Mac
 hine Learning (IST\, Lisbon)\n\n\nAbstract\nDimensional analysis is a robu
 st technique for extracting insights and finding symmetries in physical sy
 stems\, especially when the governing equations are not known. The Bucking
 ham Pi theorem provides a procedure for finding a set of dimensionless gro
 ups from given measurements\, although this set is not unique. We propose 
 an automated approach using the symmetric and self-similar structure of av
 ailable measurement data to discover the dimensionless groups that best co
 llapse this data to a lower dimensional space according to an optimal fit.
  We develop three data-driven techniques that use the Buckingham Pi theore
 m as a constraint: (i) a constrained optimization problem with a nonparame
 tric function\, (ii) a deep learning algorithm (BuckiNet) that projects th
 e input parameter space to a lower dimension in the first layer\, and (iii
 ) a sparse identification of nonlinear dynamics (SINDy) to discover dimens
 ionless equations whose coefficients parameterize the dynamics. I discuss 
 the accuracy and robustness of these methods when applied to known nonline
 ar systems.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/83/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Inês Hipólito (Humboldt-Universität)
DTSTART:20220908T160000Z
DTEND:20220908T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/84
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 84/">The Free Energy Principle in the Edge of Chaos</a>\nby Inês Hipólit
 o (Humboldt-Universität) as part of Mathematics\, Physics and Machine Lea
 rning (IST\, Lisbon)\n\n\nAbstract\nLiving beings do an extraordinary thin
 g. By being alive they are resisting the second law of thermodynamics. Thi
 s law stipulates that open\, living systems tend to dissipation by the inc
 rease of entropy or chaos. From minimal cognitive organisms like plants to
  more complex organisms equipped with nervous systems\, all living systems
  adjust and adapt to their environments\, thereby resisting the second law
 . Impressively\, while all animals cognitively enact and survive their loc
 al environments\, more complex systems do so also by actively constructing
  their local environments\, thereby not only defying the second law\, but 
 also (evolution) selective properties. Because all living beings defy the 
 second law by adjusting and engaging with the environment\, a prominent qu
 estion is how do living organisms persist while engaging in adaptive excha
 nges with their complex environments? In this talk I will offer an overvie
 w of how the Free Energy Principle (FEP) offers a principled solution to t
 his problem. The FEP prescribes that living systems maintain themselves by
  remaining in non-equilibrium steady states by restricting themselves to a
  limited number of states\; it has been widely applied to explain neurocog
 nitive function and embodied action\, develop artificial intelligence and 
 inspire psychopathology models.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/84/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Robert Nowak (University of Wisconsin-Madison)
DTSTART:20221027T160000Z
DTEND:20221027T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/85
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 85/">The Neural Balance Theorem and its Consequences</a>\nby Robert Nowak 
 (University of Wisconsin-Madison) as part of Mathematics\, Physics and Mac
 hine Learning (IST\, Lisbon)\n\nAbstract: TBA\n
LOCATION:https://stable.researchseminars.org/talk/MPML/85/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Frederico Fiuza (SLAC)
DTSTART:20221103T170000Z
DTEND:20221103T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/86
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 86/">Accelerating the understanding of nonlinear dynamical systems using m
 achine learning</a>\nby Frederico Fiuza (SLAC) as part of Mathematics\, Ph
 ysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe description o
 f nonlinear\, multi-scale dynamics is a common challenge in a wide range o
 f physical systems and research fields — from weather forecast to contro
 lled nuclear fusion. The development of reduced models that balance betwee
 n accuracy and complexity is critical to advancing theoretical comprehensi
 on and enabling holistic computational descriptions of these problems. I w
 ill discuss how techniques from statistical and machine learning are offer
 ing new ways of inferring reduced physics models from the increasingly abu
 ndant data of nonlinear dynamics produced by experiments\, observations\, 
 and simulations. In particular\, I will focus on how sparse regression tec
 hniques can be used to infer interpretable plasma physics models (in the f
 orm of nonlinear partial differential equations) directly from the data of
  first-principles fully-kinetic simulations. The potential of this approac
 h is demonstrated by recovering the fundamental hierarchy of plasma physic
 s models based solely on particle-based simulation data of complex plasma 
 dynamics. I will discuss how this data-driven methodology provides a promi
 sing tool to accelerate the development of reduced theoretical models of n
 onlinear dynamical systems and to design computationally efficient algorit
 hms for multi-scale simulations.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/86/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Markus Reichstein (MPI for Biogeochemistry)
DTSTART:20221124T170000Z
DTEND:20221124T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/87
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 87/">Integrating Machine Learning with System Modelling and Observations f
 or a better understanding of the Earth System</a>\nby Markus Reichstein (M
 PI for Biogeochemistry) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nThe Earth is a complex dynamic networked
  system. Machine learning\, i.e. derivation of computational models from d
 ata\, has already made important contributions to predict and understand c
 omponents of the Earth system\, specifically in climate\, remote sensing a
 nd environmental sciences. For instance\, classifications of land cover ty
 pes\, prediction of land-atmosphere and ocean-atmosphere exchange\, or det
 ection of extreme events have greatly benefited from these approaches. Suc
 h data-driven information has already changed how Earth system models are 
 evaluated and further developed. However\, many studies have not yet suffi
 ciently addressed and exploited dynamic aspects of systems\, such as memor
 y effects for prediction and effects of spatial context\, e.g. for classif
 ication and change detection. In particular new developments in deep learn
 ing offer great potential to overcome these limitations. Yet\, a key chall
 enge and opportunity is to integrate (physical-biological) system modeling
  approaches with machine learning into hybrid modeling approaches\, which 
 combines physical consistency and machine learning versatility. A couple o
 f examples are given with focus on the terrestrial biosphere\, where the c
 ombination of system-based and machine-learning-based modelling helps our 
 understanding of aspects of the Earth system.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/87/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Bruno Loureiro (École Polytechnique Fédérale de Lausanne (EPFL)
 )
DTSTART:20221215T170000Z
DTEND:20221215T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/88
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 88/">Phase diagram of Stochastic Gradient Descent in high-dimensional two-
 layer neural networks</a>\nby Bruno Loureiro (École Polytechnique Fédér
 ale de Lausanne (EPFL)) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nDespite the non-convex optimization land
 scape\, over-parametrized shallow networks are able to achieve global conv
 ergence under gradient descent. The picture can be radically different for
  narrow networks\, which tend to get stuck in badly-generalizing local min
 ima. Here we investigate the cross-over between these two regimes in the h
 igh-dimensional setting\, and in particular investigate the connection bet
 ween the so-called mean-field/hydrodynamic regime and the seminal approach
  of Saad & Solla. Focusing on the case of Gaussian data\, we study the int
 erplay between the learning rate\, the time scale\, and the number of hidd
 en units in the high-dimensional dynamics of stochastic gradient descent (
 SGD). Our work builds on a deterministic description of SGD in high-dimens
 ions from statistical physics\, which we extend and for which we provide r
 igorous convergence rates.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/88/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Diogo Gomes (KAUST)
DTSTART:20221014T083000Z
DTEND:20221014T110000Z
DTSTAMP:20260404T094311Z
UID:MPML/89
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 89/">From Calculus of Variations to Reinforcement Learning (Lectures 1 & 2
 )</a>\nby Diogo Gomes (KAUST) as part of Mathematics\, Physics and Machine
  Learning (IST\, Lisbon)\n\n\nAbstract\nThis course begins with a brief in
 troduction to classical calculus of variations and its applications to cla
 ssical problems such as geodesic trajectories and the brachistochrone prob
 lem. Then\, we examine Hamilton-Jacobi equations\, the role of convexity a
 nd the classical verification theorem. Next\, we illustrate the lack of cl
 assical solutions and motivate the definition of viscosity solutions. The 
 course ends with a brief description of the reinforcement learning problem
  and its connection with Hamilton-Jacobi equations.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/89/
END:VEVENT
BEGIN:VEVENT
SUMMARY:José Miguel Urbano (KAUST)
DTSTART:20221014T133000Z
DTEND:20221014T160000Z
DTSTAMP:20260404T094311Z
UID:MPML/90
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 90/">Semi-Supervised Learning and the infinite-Laplacian (Lectures 1 & 2)<
 /a>\nby José Miguel Urbano (KAUST) as part of Mathematics\, Physics and M
 achine Learning (IST\, Lisbon)\n\n\nAbstract\nMotivated by a recent applic
 ation in Semi-Supervised Learning (SSL)\, the minicourse is a brief introd
 uction to the analysis of infinity-harmonic functions. We will discuss the
  Lipschitz extension problem\, its solution via MacShane-Whitney extension
 s and its several drawbacks\, leading to the notion of AMLE (Absolutely Mi
 nimising Lipschitz Extension). We then explore the equivalence between bei
 ng absolutely minimising Lipschitz\, enjoying comparison with cones and so
 lving the infinity-Laplace equation in the viscosity sense.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/90/
END:VEVENT
BEGIN:VEVENT
SUMMARY:João Sacramento (ETH Zürich)
DTSTART:20221110T170000Z
DTEND:20221110T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/91
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 91/">The least-control principle for learning at equilibrium</a>\nby João
  Sacramento (ETH Zürich) as part of Mathematics\, Physics and Machine Lea
 rning (IST\, Lisbon)\n\n\nAbstract\nA large number of models of interest i
 n both neuroscience and machine learning can be expressed as dynamical sys
 tems at equilibrium. This class of systems includes deep neural networks\,
  equilibrium recurrent neural networks\, and meta-learning. In this talk I
  will present a new principle for learning equilibria with a temporally - 
 and spatially - local rule. Our principle casts learning as a least-contro
 l problem\, where we first introduce an optimal controller to lead the sys
 tem towards a solution state\, and then define learning as reducing the am
 ount of control needed to reach such a state. We show that incorporating l
 earning signals within a dynamics as an optimal control enables transmitti
 ng activity-dependent credit assignment information\, avoids storing inter
 mediate states in memory\, and does not rely on infinitesimal learning sig
 nals. In practice\, our principle leads to strong performance matching tha
 t of leading gradient-based learning methods when applied to an array of b
 enchmarking experiments. Our results shed light on how the brain might lea
 rn and offer new ways of approaching a broad class of machine learning pro
 blems.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/91/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Tom Goldstein (University of Maryland)
DTSTART:20221117T170000Z
DTEND:20221117T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/92
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 92/">Building (and breaking) neural networks that think fast and slow</a>\
 nby Tom Goldstein (University of Maryland) as part of Mathematics\, Physic
 s and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMost neural networks 
 are built to solve simple patternmatching tasks\, a process that is often 
 known as “fast” thinking. In this talk\, I’ll use adversarial method
 s to explore the robustness of neural networks. I’ll also discuss whethe
 r vulnerabilities of AI systems that have been observed in academic labs c
 an pose real security threats to industrial systems. Then\, I’ll present
  methods for constructing neural networks that exhibit “slow” thinking
  abilities akin to human logical reasoning. Rather than learning simple pa
 ttern matching rules\, these networks have the ability to synthesize algor
 ithmic reasoning processes and solve difficult discrete search and plannin
 g problems that cannot be solved by conventional AI systems. Interestingly
 \, these reasoning systems naturally exhibit error correction and robustne
 ss properties that make them more difficult to break than their fast think
 ing counterparts.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/92/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Yang-Hui He (London Institute for Mathematical Sciences & Merton C
 ollege\, Oxford University)
DTSTART:20230202T170000Z
DTEND:20230202T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/93
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 93/">COLLOQUIUM: Universes as Bigdata: Physics\, Geometry and Machine-Lear
 ning</a>\nby Yang-Hui He (London Institute for Mathematical Sciences & Mer
 ton College\, Oxford University) as part of Mathematics\, Physics and Mach
 ine Learning (IST\, Lisbon)\n\n\nAbstract\nThe search for the Theory of Ev
 erything has led to superstring theory\, which then led physics\, first to
  algebraic/differential geometry/topology\, and then to computational geom
 etry\, and now to data science. With a concrete playground of the geometri
 c landscape\, accumulated by the collaboration of physicists\, mathematici
 ans and computer scientists over the last 4 decades\, we show how the late
 st techniques in machine-learning can help explore problems of interest to
  theoretical physics and to pure mathematics. At the core of our programme
  is the question: how can AI help us with mathematics?\n
LOCATION:https://stable.researchseminars.org/talk/MPML/93/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sebastian Engelke (University of Geneva)
DTSTART:20230112T170000Z
DTEND:20230112T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/94
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 94/">Machine learning beyond the data range: extreme quantile regression</
 a>\nby Sebastian Engelke (University of Geneva) as part of Mathematics\, P
 hysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMachine learning
  methods perform well in prediction tasks within the range of the training
  data. When interest is in quantiles of the response that go beyond the ob
 served records\, these methods typically break down. Extreme value theory 
 provides the mathematical foundation for estimation of such extreme quanti
 les. A common approach is to approximate the exceedances over a high thres
 hold by the generalized Pareto distribution. For conditional extreme quant
 iles\, one may model the parameters of this distribution as functions of t
 he predictors. Up to now\, the existing methods are either not flexible en
 ough or do not generalize well in higher dimensions. We develop new approa
 ches for extreme quantile regression that estimate the parameters of the g
 eneralized Pareto distribution with tree-based methods and recurrent neura
 l networks. Our estimators outperform classical machine learning methods a
 nd methods from extreme value theory in simulations studies. We illustrate
  how the recurrent neural network model can be used for effective forecast
 ing of flood risk.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/94/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Alhussein Fawzi (DeepMind)
DTSTART:20230119T170000Z
DTEND:20230119T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/95
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 95/">Discovering faster matrix multiplication algorithms with deep reinfor
 cement learning</a>\nby Alhussein Fawzi (DeepMind) as part of Mathematics\
 , Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nImproving the
  efficiency of algorithms for fundamental computational tasks such as matr
 ix multiplication can have widespread impact\, as it affects the overall s
 peed of a large amount of computations. The automatic discovery of algorit
 hms using machine learning offers the prospect of reaching beyond human in
 tuition and outperforming the current best human-designed algorithms. In t
 his talk I'll present AlphaTensor\, our reinforcement learning agent based
  on AlphaZero for discovering efficient and provably correct algorithms fo
 r the multiplication of arbitrary matrices. AlphaTensor discovered algorit
 hms that outperform the state-of-the-art complexity for many matrix sizes.
  Particularly relevant is the case of 4 × 4 matrices in a finite field\, 
 where AlphaTensor's algorithm improves on Strassen's two-level algorithm f
 or the first time since its discovery 50 years ago. I'll present our probl
 em formulation as a single-player game\, the key ingredients that enable t
 ackling such difficult mathematical problems using reinforcement learning\
 , and the flexibility of the AlphaTensor framework.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/95/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sara A. Solla (Northwestern University)
DTSTART:20230302T170000Z
DTEND:20230302T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/96
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 96/">Low Dimensional Manifolds for Neural Dynamics</a>\nby Sara A. Solla (
 Northwestern University) as part of Mathematics\, Physics and Machine Lear
 ning (IST\, Lisbon)\n\n\nAbstract\nThe ability to simultaneously record th
 e activity from tens to hundreds to thousands of neurons has allowed us to
  analyze the computational role of population activity as opposed to singl
 e neuron activity. Recent work on a variety of cortical areas suggests tha
 t neural function may be built on the activation of population-wide activi
 ty patterns\, the neural modes\, rather than on the independent modulation
  of individual neural activity. These neural modes\, the dominant covariat
 ion patterns within the neural population\, define a low dimensional neura
 l manifold that captures most of the variance in the recorded neural activ
 ity. We refer to the time-dependent activation of the neural modes as thei
 r latent dynamics and argue that latent cortical dynamics within the manif
 old are the fundamental and stable building blocks of neural population ac
 tivity.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/96/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Andreas Döpp (Ludwig-Maximilians-Universität München)
DTSTART:20230601T160000Z
DTEND:20230601T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/97
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 97/">Machine-learning strategies in laser-plasma physics</a>\nby Andreas D
 öpp (Ludwig-Maximilians-Universität München) as part of Mathematics\, P
 hysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p>The field of 
 laser-plasma physics has experienced significant advancements in the past 
 few decades\, owing to the increasing power and accessibility of high-powe
 r lasers. Initially\, research in this area was limited to single-shot exp
 eriments with minimal exploration of parameters. However\, recent technolo
 gical advancements have enabled the collection of a wealth of data through
  both experimental and simulation-based approaches.</p>\n\n<p>In this semi
 nar talk\, I will present a range of machine learning techniques that we h
 ave developed for applications in laser-plasma physics [1]. The first part
  of my talk will focus on Bayesian optimization\, where I will showcase ou
 r latest findings on multi-objective and multi-fidelity optimization of la
 ser-plasma accelerators and neural networks [2-4].</p>\n\n<p>In the second
  part of the talk\, I will discuss machine learning solutions for tackling
  complex inverse problems\, such as image deblurring or extracting 3D info
 rmation from 2D sensors [5-6]. Specifically\, I will discuss various adapt
 ations of established convolutional network architectures\, such as the U-
 Net\, as well as novel physics-informed retrieval methods like deep algori
 thm unrolling. These techniques have shown promising results in overcoming
  the challenges posed by these intricate inverse problems.</p>\n\n<p><stro
 ng>References:</strong></p>\n\n<p>[1] Data-driven Science and Machine Lear
 ning Methods in Laser-Plasma Physics<br />\n<a href="https://arxiv.org/abs
 /2212.00026">https://arxiv.org/abs/2212.00026</a></p>\n\n<p>[2] Expected h
 ypervolume improvement for simultaneous multi-objective and multi-fidelity
  optimization<br />\n<a href="https://arxiv.org/abs/2112.13901">https://ar
 xiv.org/abs/2112.13901</a></p>\n\n<p>[3] Multi-objective and multi-fidelit
 y Bayesian optimization of laser-plasma acceleration<br />\n<a href="https
 ://arxiv.org/abs/2210.03484">https://arxiv.org/abs/2210.03484</a></p>\n\n<
 p>[4] Pareto Optimization of a Laser Wakefield Accelerator<br />\n<a href=
 "https://arxiv.org/abs/2303.15825">https://arxiv.org/abs/2303.15825</a></p
 >\n\n<p>[5] Measuring spatio-temporal couplings using modal spatio-spectra
 l wavefront retrieval<br />\n<a href="https://arxiv.org/abs/2303.01360">ht
 tps://arxiv.org/abs/2303.01360</a></p>\n\n<p>[6] Hyperspectral Compressive
  Wavefront Sensing<br />\n<a href="https://arxiv.org/abs/2303.03555">https
 ://arxiv.org/abs/2303.03555</a></p>\n\n<p>&nbsp\;</p>\n
LOCATION:https://stable.researchseminars.org/talk/MPML/97/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ben Edelman (Harvard University)
DTSTART:20230209T170000Z
DTEND:20230209T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/98
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 98/">Studies in feature learning through the lens of sparse boolean functi
 ons</a>\nby Ben Edelman (Harvard University) as part of Mathematics\, Phys
 ics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nHow do deep neural 
 networks learn to construct useful features? Why do self-attention-based n
 etworks such as transformers perform so well on combinatorial tasks such a
 s language learning? Why do some capabilities of networks emerge "disconti
 nuously" as the computational resources used for training are scaled up? W
 e will present perspectives on these questions through the lens of a parti
 cular class of simple synthetic tasks: learning sparse boolean functions. 
 In part one\, we will show that the hypothesis class of one-layer transfor
 mers can learn these functions in a statistically efficient manner. This l
 eads to a view of each layer of a transformer as creating new "variables" 
 out of sparse combinations of the previous layer's outputs. In part two\, 
 we will focus on the classic task of learning sparse parities\, which is s
 tatistically easy but computationally difficult. We will demonstrate that 
 SGD on various neural networks (transformers\, MLPs\, etc.) successfully l
 earns sparse parities\, with computational efficiency that is close to kno
 wn lower bounds. Moreover\, the training curves display no apparent progre
 ss for a long time\, and then quickly drop late in training. We show that 
 despite this apparent delayed breakthrough in performance\, hidden progres
 s is actually being made throughout the course of training.\n\nBased on jo
 int work with Surbhi Goel\, Sham Kakade\, Cyril Zhang\, Boaz Barak\, and E
 ran Malach:\n\nhttps://arxiv.org/abs/2110.10090\n\nhttps://arxiv.org/abs/2
 207.08799\n
LOCATION:https://stable.researchseminars.org/talk/MPML/98/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Valentin De Bortoli (Center for Sciences of Data\, ENS Ulm\, Paris
 )
DTSTART:20230316T170000Z
DTEND:20230316T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/99
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 99/">Diffusion models\, theory and methodology</a>\nby Valentin De Bortoli
  (Center for Sciences of Data\, ENS Ulm\, Paris) as part of Mathematics\, 
 Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nGenerative mode
 ling is the task of drawing new samples from an underlying distribution kn
 own only via an empirical measure. There exists a myriad of models to tack
 le this problem with applications in image and speech processing\, medical
  imaging\, forecasting and protein modeling to cite a few. Among these met
 hods diffusion models are a new powerful class of generative models that e
 xhibit remarkable empirical performance. They consist of a ``noising'' sta
 ge\, whereby a diffusion is used to gradually add Gaussian noise to data\,
  and a generative model\, which entails a ``denoising'' process defined by
  approximating the time-reversal of the diffusion. In this talk we discuss
  three aspects of diffusion models. First\, we will dive into the methodol
 ogy behind diffusion models. Second\, we will present some of their theore
 tical guarantees with an emphasis on their behavior under the so-called ma
 nifold hypothesis. Such theoretical guarantees are non-vacuous and provide
  insight on the empirical behavior of these models. Finally\, I will prese
 nt an extension of diffusion models to the Optimal Transport setting and i
 ntroduce Diffusion Schrodinger Bridges.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/99/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Memming Park (Champalimaud Foundation)
DTSTART:20230323T170000Z
DTEND:20230323T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/100
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 100/">On learning signals in recurrent networks</a>\nby Memming Park (Cham
 palimaud Foundation) as part of Mathematics\, Physics and Machine Learning
  (IST\, Lisbon)\n\n\nAbstract\nNeural dynamical systems with stable attrac
 tor structures such as point attractors and continuous attractors are wide
 ly hypothesized to underlie meaningful temporal behavior that requires wor
 king memory. However\, perhaps counterintuitively\, having good working me
 mory is not sufficient for supporting useful learning signals that are nec
 essary to adapt to changes in the temporal structure of the environment. W
 e show that in addition to the well-known continuous attractors\, the peri
 odic and quasi-periodic attractors are also fundamentally capable of suppo
 rting learning arbitrarily long temporal relationships. Due to the fine tu
 ning problem of the continuous attractors and the lack of\ntemporal fluctu
 ations\, we believe the less explored quasi-periodic attractors are unique
 ly qualified for learning to produce temporally structured behavior. Our t
 heory has wide implications for the design of artificial learning systems\
 , and makes predictions on the observable signatures of biological neural 
 dynamics that can support temporal dependence learning. Based on our theor
 y\, we developed a new initialization scheme for artificial recurrent neur
 al networks which outperforms standard methods for tasks that require lear
 ning temporal dynamics. Finally\, we speculate on their biological impleme
 ntations and make predictions on neuronal dynamics.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/100/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rongjie Lai (Rensselaer Polytechnic Institute)
DTSTART:20230420T160000Z
DTEND:20230420T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/101
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 101/">Learning Manifold-Structured Data using Deep Neural Networks: Theory
  and Applications</a>\nby Rongjie Lai (Rensselaer Polytechnic Institute) a
 s part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nA
 bstract\nDeep artificial neural networks have made great success in many p
 roblems in science and engineering. In this talk\, I will discuss our rece
 nt efforts to develop DNNs capable of learning non-trivial geometry inform
 ation hidden in data. In the first part\, I will discuss our work on advoc
 ating the use of a multi-chart latent space for better data representation
 . Inspired by differential geometry\, we propose a Chart Auto-Encoder (CAE
 ) and prove a universal approximation theorem on its representation capabi
 lity. CAE admits desirable manifold properties that conventional auto-enco
 ders with a flat latent space fail to obey. We further establish statistic
 al guarantees on the generalization error for trained CAE models and show 
 their robustness to noise. Our numerical experiments also demonstrate sati
 sfactory performance on data with complicated geometry and topology. If ti
 me permits\, I will discuss our work on defining convolution on manifolds 
 via parallel transport. This geometric way of defining parallel transport 
 convolution (PTC) provides a natural combination of modeling and learning 
 on manifolds. PTC allows for the construction of compactly supported filte
 rs and is also robust to manifold deformations. I will demonstrate its app
 lications to shape analysis and point clouds processing using PTC-nets. Th
 is talk is based on a series of joint work with my students and collaborat
 ors.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/101/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gonçalo Correia (IST and Priberam Labs)
DTSTART:20230309T170000Z
DTEND:20230309T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/102
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 102/">Learnable Sparsity and Weak Supervision for Data-Efficient\, Transpa
 rent\, and Compact Neural Models</a>\nby Gonçalo Correia (IST and Pribera
 m Labs) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbo
 n)\n\n\nAbstract\nNeural network models have become ubiquitous in Machine 
 Learning literature. These models are compositions of differentiable build
 ing blocks that result in dense representations of the underlying data. To
  obtain good representations\, conventional neural models require many tra
 ining data points. Moreover\, those representations\, albeit capable of ob
 taining a high performance on many tasks\, are largely uninterpretable. Th
 ese models are often overparameterized and give out representations that d
 o not compactly represent the data. To address these issues\, we find solu
 tions in sparsity and various forms of weak supervision. For data-efficien
 cy\, we leverage transfer learning as a form of weak supervision. The prop
 osed model can perform similarly to models trained on millions of data poi
 nts on a sequence-to-sequence generation task\, even though we only train 
 it on a few thousand. For transparency\, we propose a probability normaliz
 ing function that can learn its sparsity. The model learns the sparsity it
  needs differentiably and thus adapts it to the data according to the neur
 al component's role in the overall structure. We show that the proposed mo
 del improves the interpretability of a popular neural machine translation 
 architecture when compared to conventional probability normalizing functio
 ns. Finally\, for compactness\, we uncover a way to obtain exact gradients
  of discrete and structured latent variable models efficiently. The discre
 te nodes in these models can compactly represent implicit clusters and str
 uctures in the data\, but training them was often complex and prone to fai
 lure since it required approximations that rely on sampling or relaxations
 . We propose to train these models with exact gradients by parameterizing 
 discrete distributions with sparse functions\, both unstructured and struc
 tured. We obtain good performance on three latent variable model applicati
 ons while still achieving the practicality of the approximations mentioned
  above. Through these novel contributions\, we challenge the conventional 
 wisdom that neural models cannot exhibit data-efficiency\, transparency\, 
 or compactness.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/102/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Diogo Gomes (KAUST)
DTSTART:20230504T160000Z
DTEND:20230504T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/103
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 103/">Mathematics for data science and AI - curriculum design\, experience
 s\, and lessons learned</a>\nby Diogo Gomes (KAUST) as part of Mathematics
 \, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIn this talk
 \, we will explore the importance of mathematical foundations for AI and d
 ata science and the design of an academic curriculum for graduate students
 . While traditional mathematics for AI and data science has focused on cor
 e techniques like linear algebra\, basic probability\, and optimization me
 thods (e.g.\, gradient and stochastic gradient descent)\, several advanced
  mathematical techniques are now essential to understanding modern data sc
 ience. These include ideas from the calculus of variations in spaces of ra
 ndom variables\, functional analytic methods\, ergodic theory\, control th
 eory methods in reinforcement learning\, and metrics in spaces of probabil
 ity measures. We will discuss the author's experience designing an applied
  mathematics curriculum on data science and draw on the author's experienc
 e and lessons learned in teaching an advanced course on the mathematical f
 oundations of data science. This talk aims to promote discussion and excha
 nge of ideas on how mathematicians can play an important role in AI and da
 ta science and better equip our students to excel in this field.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/103/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Harry Desmond (University of Portsmouth)
DTSTART:20230511T160000Z
DTEND:20230511T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/104
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 104/">Exhaustive Symbolic Regression (or how to find the best function for
  your data)</a>\nby Harry Desmond (University of Portsmouth) as part of Ma
 thematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nSym
 bolic regression aims to find optimal functional representation of dataset
 s\, with broad applications across science. This is traditionally done usi
 ng a "genetic algorithm" which stochastically searches function space usin
 g an evolution-inspired method for generating new trial functions. Motivat
 ed by the uncertainties inherent in this approach -- and its failure on se
 emingly simple test cases -- I will describe a new method which exhaustive
 ly searches and evaluates function space. Coupled to a model selection pri
 nciple based on minimum description length\, Exhaustive Symbolic Regressio
 n is guaranteed to find the simple equations that optimally balance simpli
 city with accuracy on any dataset. I will describe how the method works an
 d showcase it on Hubble rate measurements and dynamical galaxy data.\n\nBa
 sed on work with Deaglan Bartlett and Pedro G. Ferreira: <br>\nhttps://arx
 iv.org/abs/2211.11461 <br>\nhttps://arxiv.org/abs/2301.04368\n
LOCATION:https://stable.researchseminars.org/talk/MPML/104/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Paulo Rosa (Deimos)
DTSTART:20230427T160000Z
DTEND:20230427T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/105
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 105/">Deep Reinforcement Learning based Integrated Guidance and Control fo
 r a Launcher Landing Problem</a>\nby Paulo Rosa (Deimos) as part of Mathem
 atics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nDeep Re
 inforcement Learning (Deep-RL) has received considerable attention in rece
 nt years due to its ability to make an agent learn how to take optimal con
 trol actions\, given rich observation data via the maximization of a rewar
 d function. Future space missions will need new on-board autonomy capabili
 ties with increasingly complex requirements at the limits of the vehicle p
 erformance. This justifies the use of machine learning based techniques\, 
 in particular reinforcement learning in order to allow exploring the edge 
 of the performance trade-off space. The guidance and control systems devel
 opment for Reusable Launch Vehicles (RLV) can take advantage of reinforcem
 ent learning techniques for optimal adaption in the face of multi-objectiv
 e requirements and uncertain scenarios.\n\nIn AI4GNC - a project funded by
  the European Space Agency (ESA)\, led by DEIMOS and participated by INESC
 -ID\, the University of Lund\, and TASC - a Deep-RL algorithm was used to 
 train an actor-critic agent to simultaneously control the engine thrust ma
 gnitude and the two TVC gimbal angles to land a RLV in 6-DoF simulation. T
 he design followed an incremental approach\, progressively augmenting the 
 number of degrees of freedom and introducing more complexity factors such 
 as nonlinearity in models. Ultimately\, the full 6-DoF problem was address
 ed using a high fidelity simulator that includes a nonlinear actuator mode
 l and a realistic vehicle aerodynamic model. Starting from an initial vehi
 cle state along a reentry trajectory\, the problem consists of precisely l
 and the RLV while ensuring system requirements satisfaction\, such as satu
 ration and rate limits in the actuation\, and aiming at fuel consumption o
 ptimality. The Deep Deterministic Policy Gradient (DDPG) algorithm was ado
 pted as candidate strategy to allow the design of an integrated guidance a
 nd control algorithm in continuous action and observation spaces.\n\nThe r
 esults obtained are very satisfactory in terms of landing accuracy and fue
 l consumption. These results were also compared to a more classical and in
 dustrially used solution\, due to its capability to yield satisfactory lan
 ding accuracy and fuel consumption\, composed of a successive convexificat
 ion guidance and a PID controller tuned independently for the non-disturbe
 d nominal scenario. A reachability analysis was also performed to assess t
 he stability and robustness of the closed-loop system composed by the inte
 grated guidance and control NN\, trained for the 1-DoF scenario\, and the 
 RLV dynamics.\n\nTaking into account the fidelity of the benchmark adopted
  and the results obtained\, this approach is deemed to have a significant 
 potential for further developments and ultimately space industry applicati
 ons\, such as In-Orbit Servicing (IOS) and Active Debris Removal (ADR)\, t
 hat also require a high level of autonomy.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/105/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Artemy Kolchinsky (Universal Biology Institute\, University of Tok
 yo)
DTSTART:20230622T160000Z
DTEND:20230622T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/106
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 106/">Information geometry for nonequilibrium processes</a>\nby Artemy Kol
 chinsky (Universal Biology Institute\, University of Tokyo) as part of Mat
 hematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nRece
 ntly\, there has been dramatic progress in nonequilibrium thermodynamics\,
  with diverse applications in biological and chemical systems. The central
  quantity of interest in the field is “entropy production” (EP)\, whic
 h reflects the increase of the entropy of a system and its environment. Ma
 jor questions of interest include (1) quantitative tradeoffs between EP an
 d performance measures like speed and precision\, (2) inference of EP from
  data\, and (3) decomposition of EP into contributions from different sour
 ces of dissipation. In this work\, we study the thermodynamics of nonequil
 ibrium processes by considering the information geometry of fluxes. Our ap
 proach can be seen as a dynamical generalization of existing work on the i
 nformation geometry of probability distributions considered at a given ins
 tant in time. It is applicable to a broad range of nonequilibrium processe
 s\, including nonlinear ones that exhibit oscillations and/or chaos\, and 
 it has implications for thermodynamic tradeoffs\, thermodynamic inference\
 , and decompositions of EP. As one application\, we derive a universal dec
 omposition of EP into “excess” and “housekeeping” contributions\, 
 representing contributions from nonstationarity and cyclic fluxes respecti
 vely.\n\n(joint work with Andreas Dechant\, Kohei Yoshimura\, Sosuke Ito. 
 arXiv:2206.14599)\n
LOCATION:https://stable.researchseminars.org/talk/MPML/106/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rui Castro (Mathematics Department\, TU Eindhoven)
DTSTART:20230518T160000Z
DTEND:20230518T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/107
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 107/">Anomaly detection for a large number of streams: a permutation/rank-
 based higher criticism approach</a>\nby Rui Castro (Mathematics Department
 \, TU Eindhoven) as part of Mathematics\, Physics and Machine Learning (IS
 T\, Lisbon)\n\n\nAbstract\nAnomaly detection when observing a large number
  of data streams is essential in a variety of applications\, ranging from 
 epidemiological studies to monitoring of complex systems. High-dimensional
  scenarios are usually tackled with scan-statistics and related methods\, 
 requiring stringent distributional assumptions for proper test calibration
 . In this talk we take a non-parametric stance\, and introduce two variant
 s of the higher criticism test that do not require knowledge of the null d
 istribution for proper calibration. In the first variant we calibrate the 
 test by permutation\, while in the second variant we use a rank-based appr
 oach. Both methodologies result in exact tests in finite samples. Our perm
 utation methodology is applicable when observations within null streams ar
 e independent and identically distributed\, and we show this methodology i
 s asymptotically optimal in the wide class of exponential models. Our rank
 -based methodology is more flexible\, and only requires observations withi
 n null streams to be independent. We provide an asymptotic characterizatio
 n of the power of the test in terms of the probability of mis-ranking null
  observations\, showing that the asymptotic power loss (relative to an ora
 cle test) is minimal for many common models. As the proposed statistics do
  not rely on asymptotic approximations\, they typically perform better tha
 n popular variants of higher criticism relying on such approximations. Fin
 ally\, we demonstrate the use of these methodologies when monitoring the c
 ontent uniformity of an active ingredient for a batch-produced drug produc
 t\, and monitoring the daily number of COVID-19 cases in the Netherlands.\
 n\nBased on joint work with Ivo Stoepker\, Ery Arias-Castro and Edwin van 
 de den Heuvel:\nhttps://arxiv.org/abs/2009.03117\n
LOCATION:https://stable.researchseminars.org/talk/MPML/107/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sara Magliacane (University of Amsterdam and MIT-IBM Watson AI Lab
 )
DTSTART:20230608T160000Z
DTEND:20230608T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/108
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 108/">Causal vs causality-inspired representation learning</a>\nby Sara Ma
 gliacane (University of Amsterdam and MIT-IBM Watson AI Lab) as part of Ma
 thematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p>
 Causal representation learning (CRL) aims at learning causal factors and t
 heir causal relations from high-dimensional observations\, e.g. images. In
  general\, this is an ill-posed problem\, but under certain assumptions or
  with the help of additional information or interventions\, we are able to
  guarantee that the representations we learn are corresponding to some tru
 e underlying causal factors up to some equivalence class.<br />\nIn this t
 alk I will first present CITRIS (<a href="https://proceedings.mlr.press/v1
 62/lippe22a/lippe22a.pdf" rel="noreferrer" target="_blank">https://proceed
 ings.mlr.press/v162/lippe22a/lippe22a.pdf</a>)\, a variational autoencoder
  framework for causal representation learning from temporal sequences of i
 mages\, in systems in which we can perform interventions. CITRIS exploits 
 temporality and observing intervention targets to identify scalar and mult
 idimensional causal factors\, such as 3D rotation angles. In experiments o
 n 3D rendered image sequences\, CITRIS outperforms previous methods on rec
 overing the underlying causal variables. Moreover\, using pretrained autoe
 ncoders\, CITRIS can even generalize to unseen instantiations of causal fa
 ctors.<br />\n<br />\nWhile CRL is an exciting and promising new field of 
 research\, the assumptions required by CITRIS and other current CRL method
 s can be difficult to satisfy in many settings. Moreover\, in many practic
 al cases learning representations that are not guaranteed to be fully caus
 al\, but exploit some ideas from causality\, can still be extremely useful
 . As examples\, I will describe some of our work on exploiting these "caus
 ality-inspired" representations for adapting policies across domains in RL
  (<a href="https://openreview.net/forum?id=8H5bpVwvt5" rel="noreferrer" ta
 rget="_blank">https://openreview.net/forum?id=8H5bpVwvt5</a>) and to nonst
 ationary environments (<a href="https://openreview.net/forum?id=VQ9fogN1q6
 e" rel="noreferrer" target="_blank">https://openreview.net/forum?id=VQ9fog
 N1q6e</a>)\, and how learning a factored graphical representations (even i
 f not necessarily causal) can be beneficial in these (and possibly other) 
 settings.</p>\n
LOCATION:https://stable.researchseminars.org/talk/MPML/108/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mário Figueiredo (Instituto Superior Técnico and IT)
DTSTART:20230615T160000Z
DTEND:20230615T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/109
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 109/">Causal Discovery from Observations: Introduction and Some Recent Adv
 ances</a>\nby Mário Figueiredo (Instituto Superior Técnico and IT) as pa
 rt of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstr
 act\nCausal discovery is an active research field that aims to uncover the
  underlying causal mechanisms that drive the relationship between a collec
 tion of variables and which has applications in many areas\, including med
 icine\, biology\, economics\, and social sciences. In principle\, identify
 ing causal relationships requires interventions. However\, intervening is 
 often impossible\, impractical\, or unethical\, which has stimulated much 
 research on causal discovery from purely observational data or mixed obser
 vational-interventional data. In this talk\, after overviewing the causal 
 discovery field\, I will discuss some recent advances\, namely on causal d
 iscovery from data with latent interventions and on what is the quintessen
 tial causal discovery problem: distinguishing the cause from the effect on
  a pair of dependent variables.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/109/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Francisco Förster Burón (Universidad de Chile)
DTSTART:20240111T170000Z
DTEND:20240111T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/110
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 110/">The ALeRCE astronomical alert broker</a>\nby Francisco Förster Bur
 ón (Universidad de Chile) as part of Mathematics\, Physics and Machine Le
 arning (IST\, Lisbon)\n\n\nAbstract\nA new generation of large aperture an
 d large field of view telescopes is allowing the exploration of large volu
 mes of the Universe in an unprecedented fashion. In order to take advantag
 e of these new telescopes\, notably the Vera C. Rubin Observatory\, a new 
 time domain ecosystem is developing. Among the tools required are fast mac
 hine learning aided discovery and classification algorithms\, interoperabl
 e tools to allow for an effective communication with the community and fol
 low-up telescopes\, and new models and tools to extract the most physical 
 knowledge from these observations. In this talk I will review the challeng
 es and progress of building one of these systems: the Automatic Learning f
 or the Rapid Classification of Events (ALeRCE) astronomical alert broker. 
 ALeRCE (http://alerce.science/) is an alert annotation and classification 
 system led by an interdisciplinary and interinstitutional group of scienti
 sts from Chile since 2019. ALeRCE is focused around three scientific cases
 : transients\, variable stars and active galactic nuclei. Thanks to its st
 ate-of-the-art machine learning models\, ALeRCE has become the 3rd group t
 o report most transient candidates to the Transient Name Server\, and it i
 s enabling new science with different astrophysical objects\, e.g. AGN sci
 ence. I will discuss some of the challenges associated with the problem of
  alert classification\, including the ingestion of multiple alert streams\
 , annotation\, database management\, training set building\, feature compu
 tation and distributed processing\, machine learning classification and vi
 sualization\, or the challenges of working in large interdisciplinary team
 s. I will also show some results based on the real‐time ingestion and cl
 assification using the Zwicky Transient Facility (ZTF) alert stream as inp
 ut\, as well as some of the tools available.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/110/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Olga Mula (TU Eindhoven)
DTSTART:20230922T130000Z
DTEND:20230922T140000Z
DTSTAMP:20260404T094311Z
UID:MPML/111
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 111/">Optimal State and Parameter Estimation Algorithms and Applications t
 o Biomedical Problems</a>\nby Olga Mula (TU Eindhoven) as part of Mathemat
 ics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIn this t
 alk\, I will present an overview of recent works aiming at solving inverse
  problems (state and parameter estimation) by combining optimally measurem
 ent observations and parametrized PDE models. After defining a notion of o
 ptimal performance in terms of the smallest possible reconstruction error 
 that any reconstruction algorithm can achieve\, I will present practical n
 umerical algorithms based on nonlinear reduced models for which we can pro
 ve that they can deliver a performance close to optimal. The proposed conc
 epts may be viewed as exploring alternatives to Bayesian inversion in favo
 r of more deterministic notions of accuracy quantification. I will illustr
 ate the performance of the approach on simple benchmark examples and we wi
 ll also discuss applications of the methodology to biomedical problems whi
 ch are challenging due to shape variability.\n\nhttps://arxiv.org/pdf/2203
 .07769.pdf\nhttps://arxiv.org/pdf/2009.02687.pdf\n
LOCATION:https://stable.researchseminars.org/talk/MPML/111/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Pedro Domingos (University of Washington)
DTSTART:20240215T170000Z
DTEND:20240215T180000Z
DTSTAMP:20260404T094311Z
UID:MPML/112
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 112/">Deep Networks Are Kernel Machines</a>\nby Pedro Domingos (University
  of Washington) as part of Mathematics\, Physics and Machine Learning (IST
 \, Lisbon)\n\n\nAbstract\nDeep learning's successes are often attributed t
 o its ability to automatically discover new representations of the data\, 
 rather than relying on handcrafted features like other learning methods. I
 n this talk\, however\, I will show that deep networks learned by the stan
 dard gradient descent algorithm are in fact mathematically approximately e
 quivalent to kernel machines\, a learning method that simply memorizes the
  data and uses it directly for prediction via a similarity function (the k
 ernel). This greatly enhances the interpretability of deep network weights
 \, by elucidating that they are effectively a superposition of the trainin
 g examples. The network architecture incorporates knowledge of the target 
 function into the kernel. The talk will include a discussion of both the m
 ain ideas behind this result and some of its more startling consequences f
 or deep learning\, kernel machines\, and machine learning at large.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/112/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Kathryn Hess (EPFL)
DTSTART:20240606T160000Z
DTEND:20240606T170000Z
DTSTAMP:20260404T094311Z
UID:MPML/113
DESCRIPTION:Title: <a href="https://stable.researchseminars.org/talk/MPML/
 113/">Of mice and men</a>\nby Kathryn Hess (EPFL) as part of Mathematics\,
  Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMotivated by t
 he desire to automate classification of neuron morphologies\, we designed 
 a topological signature\, the Topological Morphology Descriptor (TMD)\, th
 at assigns a "barcode" to any any finite binary tree embedded in ${\\mathb
 b R}^3$. Using the TMD we performed an objective\, stable classification o
 f pyramidal cells in the rat neocortex\, based only on the shape of their 
 dendrites.\n\nIn this talk\, I will introduce the TMD\, then focus on a ve
 ry recent application to comparing mouse and human cortical neurons and ch
 aracterizing the differences between them. I'll also briefly discuss the r
 ole of machine learning in our work.\n\nThis talk is based on collaboratio
 ns led by Lida Kanari of the Blue Brain Project.\n
LOCATION:https://stable.researchseminars.org/talk/MPML/113/
END:VEVENT
END:VCALENDAR
