## Lectures on Differential Equations and their Numerical Analysis

The Aziz Lectures were established by a generous gift from Prof. A. Kadir Aziz. The purpose of the lectures is to bring distinguished mathematicians to the University of Maryland, College Park, to give survey lectures on differential equations, their numerical analysis, and related areas.

Kadir Aziz, who died on March 25, 2016 at the age of 92, received his Ph.D. from the University of Maryland, College Park in 1957. He was on the faculty of Georgetown University from 1956 to 1967, and was on the faculty at the University of Maryland, Baltimore County (UMBC) since 1967. He was Professor Emeritus of Mathematics and Statistics at UMBC. Throughout his career Kadir Aziz was an active and highly respected member of the Numerical Analysis community at Maryland.

The Aziz lecture is given at **3pm in the Math Colloquium Room** (MTH 3206).

Usually the speaker gives a related talk in the **Applied Math Colloquium** on the previous day (Thursday at 3:30pm).

## Aziz Lectures 2016

### May 4, 2016

**Quantum Dots and Dislocations: Dynamics of Materials****Prof. Irene Fonseca****Department of Mathematical Sciences****Carnegie Mellon University**

The formation and assembly patterns of quantum dots have a significant impact on the optoelectronic properties of semiconductors. We will address short time existence for a surface diffusion evolution equation with curvature regularization in the context of epitaxially strained three-dimensional films. Further, the nucleation of misfit dislocations will be analyzed.

This is joint work with Nicola Fusco, Giovanni Leoni and Massimiliano Morini.

## Aziz Lectures 2015

### November 18, 2015

**Tensor Sparsity - a Regularity Notion for High Dimensional PDEs****Prof. Wolfgang Dahmen****Institute für Geometrie und Praktische Mathematik****RWTH Aachen University (Germany)**

The numerical solution of PDEs in a spatially high-dimensional regime (such as the electronic Schrödinger or Fokker-Planck equations) is severely hampered by the “curse of dimensionality”: the computational cost required for achieving a desired target accuracy increases exponentially with respect to the spatial dimension.

We explore a possible remedy by exploiting a typically hidden sparsity of the solution to such problems with respect to a problem dependent basis or dictionary. Here sparsity means that relatively few terms from such a dictionary suffice to realize a given target accuracy. Specifically, sparsity with respect to dictionaries comprised of separable functions – rank-one tensors – would significantly mitigate the curse of dimensionality. The main result establishes such tensor-sparsity for elliptic problems over product domains when the data are tensor-sparse, which can be viewed as a structural regularity theorem.

### April 15, 2015

**Waves in random media: the story of the phase****Prof. Lenya Ryzhik****Department of Mathematics ****Stanford University**

The macroscopic description of wave propagation in random media typically focuses on the scattering of the wave intensity, while the phase is discarded as a uselessly random object. At the same time, most of the beauty in wave scattering come from the phase correlations. I will describe some of the miracles, as well as some limit theorems for the wave phase.

### May 6, 2015

**Mathematical challenges in kinetic models of dilute polymers: analysis, approximation and computation****Prof Endre Süli****Mathematical Institute ****University of Oxford****United Kingdom**

We survey recent analytical and computational results for coupled macroscopic-microscopic bead-spring chain models that arise from the kinetic theory of dilute solutions of polymeric fluids with noninteracting polymer chains, involving the unsteady Navier–Stokes system in a bounded domain and a high-dimensional Fokker–Planck equation. The Fokker–Planck equation emerges from a system of (Ito) stochastic differential equations, which models the evolution of a vectorial stochastic process comprised by the centre-of-mass position vector and the orientation (or configuration) vectors of the polymer chain. We discuss the existence of large-data global-in-time weak solutions to the coupled Navier–Stokes–Fokker–Planck system. The Fokker–Planck equation involved in the model is a high-dimensional partial differential equation, whose numerical approximation is a formidable computational challenge, complicated by the fact that for practically relevant spring potentials, such as finitely extensible nonlinear elastic potentials, the drift coefficient in the Fokker–Planck equation is unbounded.

## Aziz Lectures 2014

### November 12, 2014

*The interplay between geometric modeling and simulation of partial differential equations***Prof. Annalisa Buffa**

**Istituto di Matematica Applicata e Tecnologie Informatiche "E. Magenes"**

Pavia, Italy

Pavia, Italy

Computer-based simulation of partial differential equations involves approximation of the unknown fields and a description of geometrical entities such as the computational domain and the properties of the media. There are a variety of situations: in some cases this description is very complex, in some other the governing equations are very difficult to discretize. Starting with an historical perspective, I will describe the recent efforts to improve the interplay between the mathematical approaches characterizing these two aspects of the problem.

## Aziz Lectures 2013

## November 8, 2013

*Universality and chaos in clustering and branching processes***Prof. Robert Pego**

**Carnegie-Mellon University**

Scaling limits of Smoluchowski's coagulation equation are related to probability theory in numerous remarkable ways. E.g., such an equation governs the merging of ancestral trees in critical branching processes, as observed by Bertoin and Le Gall. A simple explanation of this relies on how Bernstein functions relate to a weak topology for Levy triples. From the same theory, we find the existence of 'universal' branching mechanisms which generate complicated dynamics that contain arbitrary renormalized limits. I also plan to describe a remarkable application of Bernstein function theory to a coagulation-fragmentation model introduced in fisheries science to explain animal group size.

### April 2-3, 2013

*Maximum Norm Stability and Error Estimates for Stokes and Navier-Stokes Problems***Prof. Vivette Girault**

**Université Pierre et Marie Curie, Paris, France**

Energy norm stability estimates for the finite element discretization of the Stokes problem follow easily from the variational formulation provided the discrete pressure and velocity satisfy a uniform inf-sup condition. But deriving uniform stability estimates in L^{∞} is much more complex because variational formulations do not lend themselves to maximum norms. I shall present here the main ideas of a proof that relies on weighted L^{2} estimates for regularized Green's functions associated with the Stokes problem and on a weighted inf-sup condition. The domain is a convex polygon or polyhedron. The triangulation is shape-regular and quasi-uniform. The finite element spaces satisfy a super-approximation property, which is shown to be valid for most commonly used stable finite-element spaces. Extending this result to error estimates and to the solution of the steady incompressible Navier-Stokes problem is straightforward.

## Aziz Lectures 2012

## February 22, 2012

*Semismooth Newton Methods: Theory, Numerics and Applications***Prof. Michael Hintermüller**

**Department of Mathematics**

**Humboldt-Universität, Berlin, Germany**

Many mathematical models of processes or problems in engineering sciences, mathematical imaging, biomedical sciences or mathematical finance rely on non-smooth structures, either directly through non-differentiable associated energy models, due to (quasi)variational inequality formulations or the presence of inequality constraints in pertinent energy minimization tasks. Based on non-smooth operator equation based (re)formulations of the above problem classes, in this talk a generalized Newton framework in function space is discussed. For this purpose the concept of semismoothness in function space is addressed. Relying on the latter concept, locally superlinear convergence of the associated semismooth Newton iteration is established and its mesh independent convergence behavior upon discretization is shown. In a second part of the talk, the efficiency and wide applicability of the above semismooth Newton framework is highlighted by considering constrained optimal control problems for fluid flow, contact problems with or without adhesion forces, phase separation phenomena relying on non-smooth homogeneous free energy densities and restoration tasks in mathematical image processing.

## Aziz Lectures 2011

## December 2, 2011

*Optimal and Practical Algebraic Solvers for Discretized PDEs***Prof. Jinchao Xu**

**Center for Computational Mathematics and Applications**

**Penn State University**

An overview of fast solution techniques (such as multi-grid, two-grid, one-grid and nil-grid methods) will be given in this talk on solving large scale systems of equations that arise from the discretization of partial differential equations (such as Poisson, elasticity, Stokes, Navier-Stokes, Maxwell, MHD, and black-oil models). Mathematical optimality, practical applicability and parallel (CPU/GPU) scalability will be addressed for these algorithms and applications.

### February 11, 2011

*Complex Fluids***Prof. Peter Constantin**

**Department of Mathematics**

**University of Chicago**

The talk will be about some of the models used to describe fluids with particulate matter suspended in them. Some of these models are very complicated. After a bit of history and a review of known results, I will try to point out some open problems, isolate some of the mathematical difficulties, and illustrate some of the phenomena on simpler didactic models.

## Aziz Lectures 2010

## November 12, 2010

*Discontinuous Galerkin Finite Element Methods for High Order Nonlinear Partial Differential Equations***Prof. Chi-Wang Shu**

**Brown University**

Discontinuous Galerkin (DG) finite element methods were first designed to solve hyperbolic conservation laws utilizing successful high resolution finite difference and finite volume schemes such as approximate Riemann solvers and nonlinear limiters. More recently the DG methods have been generalized to solve convection dominated convection-diffusion equations (e.g. high Reynolds number Navier-Stokes equations), convection-dispersion (e.g. KdV equations) and other high order nonlinear wave equations or diffusion equations. In this talk we will first give an introduction to the DG method, emphasizing several key ingredients which made the method popular, and then we will move on to introduce a class of DG methods for solving high order PDEs, termed local DG (LDG) methods. We will highlight the important ingredient of the design of LDG schemes, namely the adequate choice of numerical fluxes, and emphasize the stability of the fully nonlinear DG approximations. Numerical examples will be shown to demonstrate the performance of the DG methods.

### March 5, 2010

*A Taste of Compressed Sensing***Prof. Ronald DeVore**

**Texas A&M University**

Compressed Sensing is a new paradigm in signal and image processing. It seeks to faithfully capture a signal/image with the fewest number of measurements. Rather than model a signal as a bandlimited function or an image as a pixel array, it models both of these as a sparse vector in some representation system. This model fits well real world signals and images. For example, images are well approximated by a sparse wavelet decomposition. Given this model, how should we design a sensor to capture the signal with the fewest number of measurements. We shall introduce ways of measuring the effectiveness of compressed sensing algorithms and then show which of these are optimal.

## Aziz Lectures 2009

## October 12, 2009

*Isogeometric Analysis***Prof. Thomas J. R. Hughes**

**Institute for Computational Engineering and Sciences**

**University of Texas at Austin**

Geometry is the foundation of analysis yet modern methods of computational geometry have until recently had very little impact on computational mechanics. The reason may be that the Finite Element Analysis (FEA), as we know it today, was developed in the 1950's and 1960's, before the advent and widespread use of Computer Aided Design (CAD) programs, which occurred in the 1970's and 1980's. Many difficulties encountered with FEA emanate from its approximate, polynomial based geometry, such as, for example, mesh generation, mesh refinement, sliding contact, flows about aerodynamic shapes, buckling of thin shells, etc., and it s disconnect with CAD. It would seem that it is time to look at more powerful descriptions of geometry to provide a new basis for computational mechanics.

The purpose of this talk is to describe the new generation of computational mechanics procedures based on modern developments in computational geometry. The emphasis will be on Isogeometric Analysis in which basis functions generated from NURBS (Non-Uniform Rational B-Splines) and T-Splines are employed to construct an exact geometric model. For purposes of analysis, the basis is refined and/or its order elevated without changing the geometry or its parameterization. Analogues of finite element h- and p-refinement schemes are presented and a new, more efficient, higher-order concept, k-refinement, is described. Refinements are easily implemented and exact geometry is maintained at all levels without the necessity of subsequent communication with a CAD description.

In the context of structural mechanics, it is established that the basis functions are complete with respect to affine transformations, meaning that all rigid body motions and constant strain states are exactly represented. Standard patch tests are likewise satisfied. Numerical examples exhibit optimal rates of convergence for linear elasticity problems and convergence to thin elastic shell solutions. Extraordinary accuracy is noted for k-refinement in structural vibrations and wave propagation calculations. Surprising robustness is also noted in fluid and non-linear solid mechanics problems. It is argued that Isogeometric Analysis is a viable alternative to standard, polynomial-based, finite element analysis and possesses many advantages. In particular, k-refinement seems to offer a unique combination of attributes, that is, robustness and accuracy, not possessed by classical p-methods, and is applicable to models requiring smoother basis functions, such as, thin bending elements, and strain-gradient and various phase-field theories.

A modelling paradigm for patient-specific simulation of cardiovascular fluid-structure interaction is reviewed, and a précis of the status of current mathematical understanding is presented.

### May 1, 2009

*The Fast Multipole Method and its Applications***Leslie Greengard**

**Courant Institute of Mathematical Sciences, New York University**

In this lecture, we will describe the analytic and computational foundations of fast multipole methods (FMMs), as well as some of their applications. They are most easily understood, perhaps, in the case of particle simulations, where they reduce the cost of computing all pairwise interactions in a system of N particles from O(N2) to O(N) or O(N log N) operations. FMMs are equally useful, however, in solving partial differential equations by first recasting them as integral equations. We will present examples from electromagnetics, elasticity, and fluid mechanics.

## Aziz Lectures 2008

## November 14, 2008

*Topology optimization of structures*

Prof. Gregoire Allaire**Ecole Polytechnique**

The typical problem of structural optimization is to find the "best" structure which is, at the same time, of minimal weight and of maximum strength or which performs a desired deformation. In this context I will present the combination of the classical shape derivative and of the level-set methods for front propagation. This approach has been implemented in two and three space dimensions for models of linear or non-linear elasticity and for various objective functions and constraints on the perimeter. It has also been coupled with the bubble or topological gradient method which is designed for introducing new holes in the optimization process. Since the level set method is known to easily handle boundary propagation with topological changes, the resulting numerical algorithm is very efficient for topology optimization. It can escape from local minima in a given topological class of shapes and the resulting optimal design is largely independent of the initial guess.

### March 28, 2008

*New materials from mathematics: real and imagined***Prof. Richard D. James**

**University of Minnesota**

In this talk I will give two examples where mathematics played an important role for the discovery of new materials, and a third example where mathematics suggests a systematic way of searching for broad classes of yet undiscovered materials.

## Aziz Lectures 2007

### Nov. 16, 2007

*Adaptive Approximation by Greedy Algorithms***Prof. Albert Cohen**

**Laboratoire Jacques-Louis Lions, Université Pierre et Marie Curie,**

**Paris, France**

This talk will discuss computational algorithms that deal with the following general task: given a function f and a dictionary of functions D in a Hilbert space, extract a linear combination of N functions of D which approximates f at best. We shall review the convergence properties of existing algorithms. This work is motivated by applications as various as data compression, adaptive numerical simulation of PDE's, statistical learning theory.

### May 4, 2007

**Compressive Sampling****Prof. Emmanuel J. CandesCalifornia Institute of Technology**

One of the central tenets of signal processing is the Shannon/Nyquist sampling theory: the number of samples needed to reconstruct a signal without error is dictated by its bandwidth-the length of the shortest interval which contains the support of the spectrum of the signal under study. Very recently, an alternative sampling or sensing theory has emerged which goes against this conventional wisdom. This theory allows the faithful recovery of signals and images from what appear to be highly incomplete sets of data, i.e. from far fewer data bits than traditional methods use. Underlying this metholdology is a concrete protocol for sensing and compressing data simultaneously.

This talk will present the key mathematical ideas underlying this new sampling or sensing theory, and will survey some of the most important results. We will argue that this is a robust mathematical theory; not only is it possible to recover signals accurately from just an incomplete set of measurements, but it is also possible to do so when the measurements are unreliable and corrupted by noise. We will see that the reconstruction algorithms are very concrete, stable (in the sense that they degrade smoothly as the noise level increases) and practical; in fact, they only involve solving very simple convex optimization programs.

An interesting aspect of this theory is that it has bearings on some fields in the applied sciences and engineering such as statistics, information theory, coding theory, theoretical computer science, and others as well. If time allows, we will try to explain these connections via a few selected examples.

## Aziz Lectures 2006

### December 1, 2006

*Imaging in random media***Prof. George C. Papanicolaou**

**Mathematics Department**

**Stanford University**

I will present an overview of some recently developed methods for imaging with array and distributed sensors when the environment between the objects to be imaged and the sensors is complex and only partially known to the imager. This brings in modeling and analysis in random media, and the need for statistical algorithms that increase the computational complexity of imaging, which is done by backpropagating local correlations rather than traces (interferometry). I will illustrate the theory with applications from non-destructive testing and from other areas.

### April 21, 2006

*String integration of some MHD equations***Prof. Yann Brenier**

**Laboratoire Alexandre Dieudonné**

**Université de Nice-Sophia-Antipolis, France**

We first review the link between strings and some Magnetohydrodynamics equations. Typical examples are the Born-Infeld system, the Chaplygin gas equations and the shallow water MHD model. They arise in Physics at very different (from subatomic to cosmologic) scales. These models can be exactly integrated in one space dimension by solving the 1D wave equation and using the d'Alembert formula. We show how an elementary "string integrator" can be used to solve these MHD equations through dimensional splitting. A good control of the energy conservation is needed due to the repeated use of Lagrangian to Eulerian grid projections. Numerical simulations in 1 and 2 dimensions will be shown.

### February 3, 2006

**Multiscale Analysis in Micromagnetics**

**Prof. Felix Otto**

**Institute for Applied Mathematics**

**University of Bonn, Germany**

From the point of view of mathematics, micromagnetics is an ideal playground for a pattern forming system in m aterials science: There are abundant experiments on a wealth of visually attractive phenomena and there is a well-accepted continuum model.

In this talk, I will focus on two specific experimental pattern for thin film ferromagnetic elements. One pattern is a ground state, the other pattern is a metastable state. Starting point for our analysis is the micromagnetic model which has three length scales and thus many parameter regimes. For both pattern, we identify the appropriate paramater regime and rigorously derive a reduced model via G-convergence. We numerically simulate the reduced model and compare to experimental data.

This is joint work with A. DeSimone, R. V. Kohn, and S. Müller for the first part and with R. Cantero-Alvarez and J. Steiner for the second part.

## Aziz Lectures 2005

### December 9, 2005

*Multiscale Modeling in Biosciences: **Ion Transport through Membranes*Prof. Willi Jäger

**Institute for Applied Mathematics**

University of Heidelberg, Germany

University of Heidelberg, Germany

## Aziz Lectures 2004

### November 19, 2004

**Electromagnetic imaging for small inhomogeneities****Prof. Michael VogeliusDepartment of Mathematics, Rutgers University**

### May 7, 2004

**Mathematical models for cell motion****Prof. Benoît PerthameÉcole Normale Supérieure, Paris**

## Aziz Lectures 2003

### November 14, 2003

*Multiscale Modeling and Computation of Flow in Heterogeneous Media***Prof. Tom Hou**

**Caltech**

### March 7, 2003

**Mathematical and Numerical Modeling of the Cardiovascular System**

**Prof. Alfio Quarteroni**

**Politecnico di Milano, Milan, Italy, and**

**EPFL, Lausanne, Switzerland**

## Aziz Lectures 2002

### Dec. 6, 2002

**The regularity of minimizers in elasticity****Prof. John BallDepartment of Mathematics, Oxford**

### May 3, 2002

**Multigrid: From Fourier to Gauss****Prof. Randolph E. BankDepartment of Mathematics, University of California at San Diego**

## Aziz Lectures 2001

### Nov. 16, 2001

**Mathematical Problems in Meteorology and Oceanography****Prof. Roger TemamInstitute for Scientific Computing and Applied Mathematics, Indiana University**

### April 23, 2001

**Recent Approaches in the Treatment of Subgrid Scales****Prof. Franco BrezziIstituto di Analisi Numerica del CNR and Dipartimento di Matematica, Universita di Pavia, Italy**

## Aziz Lectures 2000

### March 15, 2000

**Time Stepping in Parabolic Problems - Approximation of Analytic Semigroups****Prof. Vidar ThoméeDept. of Mathematics, Chalmers University of Technology and Göteborg University **

## Aziz Lectures 1999

### December 10, 1999

**Colliding Black Holes and Gravity Waves: A New Computational Challenge****Prof. Douglas N. ArnoldDept. of Mathematics, Pennsylvania State University**

### September 24, 1999

*A Priori and A Posteriori Error Estimates in Finite Element Approximation***Prof. Lars B. Wahlbin**

**Dept. of Mathematics, Cornell University**

### February 19, 1999

*Mathematical Problems Related to the Reliability of Finite Element Analysis in Practice: When Can We Trust the Computational Results for Engineering Decisions***Prof. Ivo Babuska**

**University of Texas, Austin, Emeritus Professor at University of Maryland**