Weekly Colloquia

Guest speakers from around the world discuss trends and research in the field of Applied Mathematics and Statistics.

Colloquium takes place at 3 p.m. on Fridays in Chauvenet Hall room 143. Attendance is in person or via Zoom.  Please contact Samy Wu Fung at swufung@mines.edu or Daniel McKenzie at dmckenzie@mines.edu for further information and the Zoom link and password.

Spring 2025

April 25Jess Ellis Hagman, Colorado State University

Title: Centering students' identities to create critical transformations: An example centering multilingual students in college algebra.

Abstract: One way to critically transform undergraduate mathematics is to center aspects of students' identities to inform the structure of the mathematical spaces and systems that they are in. In this talk, I first draw on multiple large scale NSF projects seeking to improve introductory college math programs to propose a process for achieving critical transformations. I then use recent efforts at CSU to improve instruction for multilingual students as an example of this process, sharing how we used research on multilingual students' experiences to inform the instructional practices I used to teach a large (n=80), College Algebra class with a subsection restricted for multilingual students (n=14). I end by drawing on surveys and interviews from a subset of students in this class to discuss what did and did not land for students in this class. I hope this presentation can serve to (1) share practices that can support multilingual students and also many other students, and (2) exemplify how centering students' social identities can be used to design mathematics spaces to support more students to feel good mathematically.
May 2Daniel Sanz-Alonzo
Past Spring 2025
January 31Dr. Dane Taylor, University of Wyoming

Title: Consensus processes over networks: Past, present, and future

Abstract: Models for consensus---that is, the reaching of agreement---have been developed to study, e.g., group decisions over social networks, the collectively movements of animals/agents, and the training of decentralized machine-learning (ML) algorithms on dispersed data. This talk will explore the important role of network structure in shaping consensus by considering 3 extensions for a linear consensus model. First, I will show that the presence of network community structure may or may not impose a bottleneck for consensus ML, which we analyze using random matrix theory [arxiv.org/abs/2011.01334]. Next, I will model and study collective decisions in human-AI teams by formulating consensus over interconnected networks and use spectral perturbation theory to predict the effects of asymmetric coupling between humans and AI agents [doi.org/10.1109/TNSE.2023.3325278]. Time permitting, I will formulate a consensus model with higher-order, multiway interactions using simplicial complexes (i.e., as opposed to graphs) and use algebraic topology to study how homology influences dynamics. [doi.org/10.1063/5.0080370]
February 7Dr. Christian Parkinson, Michigan State University

Title: Compartmental Models for Epidemiology with Noncompliant Behavior

Abstract: We formulate and analyze ODE and PDE models for epidemiology which incorporate human behavioral concerns. Specifically, we assume that as a disease spreads and a governing body implements non-pharmaceutical intervention methods, there is a portion of the population that does not comply with these mandates and that this noncompliance has a nontrivial effect on the spread of the disease. Borrowing from social contagion theory, we then allow this noncompliance to spread parallel to the disease. We derive reproductive ratios and large-time asymptotics for our models, and then analyze them through the lens of optimal control to account for policy maker decisions. We demonstrate the behavior of all of our models with simulations.
March 28Monique Chyba, University of Hawaii at Manoa

Title: A Tour of Controlled Dynamical Systems

Abstract: In this talk, we explore the interplay between geometry, dynamical systems, and control theory, which has driven significant advances in these fields. Control theory seeks to influence the behavior of dynamical systems to achieve desired objectives, often by optimizing a prescribed cost. Geometric optimal control is deeply connected to sub-Riemannian geometry, which, for instance, plays an important role in studying optimal strokes for micro-swimmers.

Control systems can be either continuous or discrete, depending on the application. However, modeling with dynamical systems is often more effective when incorporating both continuous and discrete states. This approach allows for the representation of local dynamics coupled into a global system, where interactions may involve discrete transitions. We will highlight key features of such systems and discuss various applications.
April 4Dr. Wolfgang Bangerth, Colorado State University

Title: On the notion of "information" in inverse problems

Abstract: Inverse problems are ones where one would like to reconstruct a spatially variable function from measurements of a system in which this function appears as a coefficient or right hand side. Examples are biomedical imaging and seismic imaging of the earth.
In many inverse problems, practitioners have an intuitive notion of how much one "knows" about the coefficient in different parts of the domain -- that is, that there is a spatially variable "amount of information". For example, in seismic imaging, we can only know about those parts of the earth that are traversed by seismic waves on their way from earthquake to receiving seismometer, whereas we cannot know about places not on these raypaths.

Despite the fact that this concept of "information" is intuitive to understand, there are no accepted quantitative measures of information in inverse problems. I will present an approach to define such a spatially variable "information density" and illustrate both how it can be computed practically, as well as used in applications. The approach is based on techniques borrowed from Bayesian inverse problems, and an approximation of the covariance matrix using the Cramer-Rao bound.
April 18Dr. Deena R Schmidt, University of Nevada, Reno

Title: Stochastic network models in biology

Abstract: Many biological systems in nature can be represented as a dynamic model on a network. Examples include gene regulatory systems, neuronal networks, social networks, epidemics spreading within a population described by a contact network, and many others. A fundamental question when studying a biological process represented as a dynamic model on a network is to what extent the network structure is contributing to the observed dynamics. I will give a brief introduction to network modeling in biology along with an overview of my work that addresses this question. I will then focus on two recent projects. The first project investigates the spread of norovirus (stomach flu) within a local population using a stochastic adaptive network model. The second project looks at mammalian sleep-wake regulation at different stages of development using an integrate-and-fire neuronal network model. If time allows, I will also discuss a few related projects with current students.
Fall 2024
November 22Ziyu Li - Grad Student Colloquium
November 15Chenlu Shi, New Jersey Institute of Technology

Title: Space-Filling Designs for Computer Experiments

Abstract: Computer simulations are essential tools for studying complex systems across various fields in natural and social sciences. However, the complexity of the models behind these simulations often leads to high computational costs. A useful approach to address this is to build a statistical surrogate model based on a set of data generated by running a computer model - computer experiment. Space-filling designs are widely recognized as effective designs for such experiments. This talk will provide an overview of space-filling designs, with a focus on a specific type known as strong orthogonal arrays. These designs are particularly attractive due to their space-filling properties in lower-dimensional projections of the input space. We will introduce this class of designs and share our recent advancements in this area.

Join in-person in room CH143 or through zoom: https://mines.zoom.us/j/99293132717
November 8

*Postponed until Spring of 2025*
Professor Albert Berahas, University of Michigan

Title: Next Generation Algorithms for Stochastic Optimization with Constraints

Abstract: Constrained optimization problems arise in a plethora of applications in science and engineering. More often than not, real-world optimization problems are contaminated with stochasticity. Stochastic gradient and related methods for solving stochastic unconstrained optimization problems have been studied extensively in recent years. It has been shown that such algorithms and much of their convergence and complexity guarantees extend in straightforward ways when one considers problems involving simple constraints, such as when one can perform projections onto the feasible region of the problem. However, settings with general nonlinear constraints have received less attention. Many of the approaches proposed for solving such problems resort to using penalty or (augmented) Lagrangian methods, which are often not the most effective strategies. In this work, we propose and analyze stochastic optimization algorithms for deterministically constrained problems based on the sequential quadratic optimization (commonly known as SQP) methodology. We discuss the rationale behind our proposed techniques, convergence in expectation, and complexity guarantees for our algorithms. Additionally, we present numerical experiments that we have performed. This is joint work with Raghu Bollapragada, Frank E. Curtis, Michael O'Neill, Daniel P. Robinson, Jiahao Shi, and Baoyu Zhou.

Bio: Albert S. Berahas is an Assistant Professor in the Industrial and Operations Engineering department at the University of Michigan. Before joining the University of Michigan, he was a Postdoctoral Research Fellow in the Industrial and Systems Engineering department at Lehigh University working with Professors Katya Scheinberg, Frank Curtis and Martin Takáč. Prior to that appointment, he was a Postdoctoral Research Fellow in the Industrial Engineering and Management Sciences department at Northwestern University working with Professor Jorge Nocedal. Berahas completed his PhD studies in the Engineering Sciences and Applied Mathematics (ESAM) department at Northwestern University in 2018, advised by Professor Jorge Nocedal. He received his undergraduate degree in Operations Research and Industrial Engineering (ORIE) from Cornell University in 2009, and in 2012 obtained an MS degree in Applied Mathematics from Northwestern University. Berahas’ research broadly focuses on designing, developing and analyzing algorithms for solving large scale nonlinear optimization problems. Specifically, he is interested in and has explored several sub-fields of nonlinear optimization such as: (i) general nonlinear optimization algorithms, (ii) optimization algorithms for machine learning, (iii) constrained optimization, (iv) stochastic optimization, (v) derivative-free optimization, and (vi) distributed optimization. Berahas served as the vice-chair of the Nonlinear Optimization cluster for the INFORMS Optimization Society (2020-2022), the chair of the Nonlinear Optimization cluster for the INFORMS Optimization Society Conference (2021-2022), the co-chair of the Nonlinear Optimization cluster for the ICCOPT 2022 conference (2021-2022), and the president of the INFORMS Junior Faculty Interest Group (JFIG) (2023-2024).
November 1Juliette Mukangango - Grad Student Colloquium
October 25Professor Jocelyn Chi, University of Colorado Boulder

Title: Randomized Kaczmarz Method for Linear Discriminant Analysis

Abstract: We present randomized Kaczmarz method for linear discriminant analysis (rkLDA), an iterative randomized approach to binary-class Gaussian model linear discriminant analysis (LDA) for very large data. We harness a least squares formulation and mobilize the stochastic gradient descent framework to obtain a randomized classifier with performance that can achieve comparable accuracy to that of full data LDA. We present analysis for the expected change in the LDA predictions if one employs the rkLDA solution in lieu of the full-data least squares solution that accounts for both the Gaussian modeling assumptions on the data and algorithmic randomness. Our analysis shows how the expected change depends on quantities inherent in the data such as the scaled condition number and Frobenius norm of the input data, how well the linear model fits the data, and choices from the randomized algorithm. Our experiments demonstrate that rkLDA can offer a viable alternative to full data LDA on a range of step-sizes and numbers of iterations.
October 18Dr. Stephen Kissler, University of Colorado Boulder

Title: Modeling infectious disease dynamics across scales

Abstract: Infectious disease outbreaks are inherently multi-scale processes: what happens inside an individual body impacts what happens in community, which in turn impacts what happens in a nation and the world. Linking these scales has proven challenging, both due to a lack of theory and an absence of adequate data. This data/theory landscape is rapidly changing, however, due in part to the renewed interest in cross-scale disease dynamics spurred by the COVID-19 pandemic. I will discuss our recent and ongoing work to characterize the within-host dynamics of acute viral infections and to link these with interpersonal contact patterns and population-level estimates of disease burden. I will touch upon the power and pitfalls of biological data collection, the beauty of statistical modeling, and the remarkable story of a hundred-year-old mathematical model that has only recently become appreciated for its full complexity.

Bio: Stephen Kissler is an assistant professor of Computer Science and an affiliate of the BioFrontiers institute at the University of Colorado Boulder. A Colorado native, he earned his bachelor’s and master’s degrees in Applied Mathematics before moving to the University of Cambridge to complete his PhD in the same subject. He worked for four years as a postdoctoral fellow at the Harvard T.H. Chan School of Public Health before returning to Colorado. He has been interested in infectious diseases since “before they were cool”, and his work remains dedicated to pandemic preparedness and resilience against respiratory viral infections using mathematical modeling.
October 11Isabella Chittumuri - Grad Student Colloquium
September 27Professor Yawen Guan, Colorado State University

Title: Spatial Confounding: The Myths and Remedies

Abstract: A key task in environmental and epidemiological studies is estimating the effect of an exposure variable on a response variable using spatially correlated observational data. However, adding a spatial random effect to account for spatial dependence can sometimes distort effect estimates, leading to differing inference results from models with and without spatial random effects. Spatial confounding, the underlying issue, has recently gained attention, yet there is no consensus on its definition. Two primary definitions have emerged: confounding due to "location" and confounding caused by an unmeasured variable. How to handle spatial confounding, and whether it should be adjusted at all, remains an ongoing debate.

In this talk, I will provide an overview of the challenges posed by spatial confounding, review current deconfounding methods, and introduce a new framework for adjusting for spatial confounding. Our approach represents spatial confounding through a spectral decomposition, which intuitively breaks down confounding into different spatial scales. Within this framework, we define the necessary assumptions to estimate the exposure effect and propose a series of confounder adjustment methods, ranging from parametric adjustments using the Matern coherence function to more flexible semi-parametric methods based on smoothing splines. These methods are applied to both areal and geostatistical data, demonstrated through simulations and real-world datasets.
September 20Professor Denis Silantyev, University of Colorado at Colorado Springs

Title: Generalized Constantin-Lax-Majda Equation: Collapse vs. Blow Up and Global Existence

Abstract: We investigate the behavior of the generalized Constantin-Lax-Majda (CLM) equation which is a 1D model for the advection and stretching of vorticity in a 3D incompressible Euler fluid. Similar to Euler equations the vortex stretching term is quadratic in vorticity, and therefore is destabilizing and has the potential to generate singular behavior, while the advection term does not cause any growth of vorticity and provides a stabilizing effect.
We study the influence of a parameter a which controls the strength of advection, distinguishing a finite time singularity formation (collapse or blow-up) vs. global existence of solutions. We find a new critical value a_c=0.6890665337007457... Such that for a<a_c there are self-similar collapsing solutions with the spatial extent of blow-up shrinking to zero, and for a_c<a1 we find that the solution exists globally. We identify the leading order complex singularity for general values of a which controls the leading order behavior of the collapsing solution. We also rederive a known exact collapsing solution for a=0 and we find a new exact analytical collapsing solution at a=1/2.
September 6Debkumar De

Title: From Data to Decisions: The Use of Statistics in Finance

Abstract: I will discuss the application of statistics and data science in the finance industry. My presentation is geared towards students and will focus on descriptive explanations rather than technical details. Additionally, I plan to discuss risk models used to manage investments in the equity market.
August 30Professor Daniel McKenzie, Colorado School of Mines

Title: Decision-focused learning: How to differentiate the solution of an optimization problem.

Abstract: Many real-world problems can be distilled into an optimization problem, for which many good algorithms exist. However, it is often the case that certain key parameters in the optimization problem are not observed directly. Instead, one can observe large amounts of data that is correlated with these parameters, but in ways that is not easy to describe. This raises the possibility of combining machine learning ---to predict the unknown parameters --- with optimization --- to solve the problem of interest. This combination is sometimes called decision-focused learning. In this talk I'll give an introduction to this field as well as describe some recent work done by myself and Dr. Samy Wu Fung.
August 23Yifan Wu, a postdoctoral researcher from Mines EE

Title: Gridless Harmonics Estimation with Multi-frequency Measurements based on Convex Optimization

Abstract: Harmonics estimation plays a crucial role in various applications, including array processing, wireless sensing, source localization, and remote sensing. In array processing applications, harmonics usually refer to the direction-of-arrival (DOA) of the source, a parameter that appears in the exponent of the complex sinusoid. Traditional harmonics estimation methods typically rely on single-frequency measurements, which are limited to narrowband signals with frequencies concentrated around a specific point. In this presentation, I will introduce our recent advancements in multi-frequency gridless DOA estimation. By leveraging a multi-frequency measurement model, our approach effectively addresses wideband signals with dispersed frequencies. Unlike conventional grid-based methods that suffer from discretization errors due to grid mismatch, our technique avoids such errors by operating in a gridless framework. The problem is initially formulated as an atomic norm minimization (ANM) problem, which can be equivalently expressed as a semidefinite program (SDP). We provide conditions for the optimality of the SDP, allowing for its verification through a computable metric. Additionally, we present a fast version of the SDP to enhance computational efficiency and extend the method to non-uniform setups. Importantly, the multi-frequency setup enables us to resolve more DOAs (harmonics) than the number of physical sensors. Numerical experiments demonstrate the superiority of the proposed method.
Spring 2024
January 12Ryan Peterson - Graduate Student Colloquium

Title: Spatial Statistical Data Fusion with LatticeKrig
January 19Krishna Balasubramanian

Title: High-dimensional scaling limits of least-square online SGD and its fluctuations.

Abstract: Stochastic Gradient Descent (SGD) is widely used in modern data science. Existing analyses of SGD have predominantly focused on the fixed-dimensional setting. In order to perform high-dimensional statistical inference with such algorithms, it is important to study the dynamics of SGD when both the dimension and the iterations go to infinity at a proportional rate. In this talk, I will discuss such high-dimensional limit theorems for the online least-squares SGD iterates for solving over-parameterized linear regression. Specifically, focusing on the double-asymptotic setting (i.e., when both the dimensionality and iterations tend to infinity), I will present the mean-field limit (in the form of an infinite-dimensional ODE) and fluctuations (in the form of an infinite-dimensional SDEs) for the online least-squares SGD iterates, highlighting certain phase-transitions. A direct consequence of the result is obtaining explicit expressions for the mean-squared estimation/prediction errors and its fluctuations, under high-dimensional scalings.
February 2Mahadevan Ganesh

Title: Time- and Frequency-Domain Wave Propagation Models: Reformulations, Algorithms, Analyses, and Simulations

Abstract: Efficient simulation of wave propagation induced by multiple structures is fundamental for numerous applications. Robust mathematical modeling of the underlying
time-dependent physical process is crucial for designing high-order computational methods for the multiple scattering simulations. Development of related algorithms and analyses are based on celebrated continuous mathematical equations either in the time- or frequency-domain, with the latter involving mathematical manipulations.
Consequently, the meaning of the term "multiple scattering" varies depending on the context in which it is used. Physics literature suggests that the continuous frequency-domain (FD) multiple scattering model is a purely mathematical construct, and that in the time-domain (TD), multiple scattering becomes a definite physical phenomenon. In recent years there has been substantial development of computational multiple scattering algorithms in the FD. In the context of computational multiple scattering, it is important to ensure that the simulated solutions represent the definite physical multiple scattering process. In this talk, we describe our recent contributions to the development of high-order wave propagation computational models in both time- and frequency-domains, and we argue that spectrally accurate FD scattering algorithms are crucial for efficient and practical simulation of physically appropriate TD multiple scattering phenomena in unbounded regions with multiple structures.
February 9Matt Hofkes - Graduate Student Colloquium
February 16Brandon Knutson - Graduate Student Colloquium
February 23Julia Arciero


Title: Modeling oxygen transport and flow regulation in the human retina

Abstract: Impairments in retinal blood flow and oxygenation have been shown to contribute to the progression of glaucoma. In this study, a theoretical model of the retina is used to predict retinal blood flow and oxyhemoglobin saturation at differing levels of capillary density and autoregulation capacity as intraluminal pressure, oxygen demand, or intraocular pressure are varied. The model includes a heterogeneous representation of retinal arterioles and a compartmental representation of capillaries and venules. A Green’s function method is used to model oxygen transport in the arterioles, and a Krogh cylinder model is used in the capillaries and venules. Model results predict that increased intraocular pressure and impaired blood flow regulation can each cause decreased tissue oxygenation. Under baseline conditions of a capillary density of 500 mm-2, an autoregulation plateau is predicted for incoming intraluminal pressures in the range of 32 - 40 mmHg. Decreasing capillary density or increasing intraocular pressure leads to a loss in the autoregulation plateau in that pressure range. If the patient has no ability to regulate flow, the autoregulation plateau disappears entirely. Ultimately, the model framework presented in this study will allow for future comparisons to sectorial-specific clinical data to help assess the potential role of impaired blood flow regulation in ocular disease.
March 1Brennan Sprinkle & Dorit Hammerling

Title: Why applied math and statistics work so well together: Detecting, localizing and quantifying methane emissions on oil and gas facilities.

Abstract: Methane, the main component of natural gas, is the second-largest contributor to climate change after carbon dioxide. Methane has a higher heat-trapping potential but shorter lifetime than carbon dioxide, and therefore, rapid reduction of methane emissions can have quick and large climate change mitigation impacts. Reducing emissions from the oil and gas supply chain, which account for approximately 14% of total methane emissions, turns out to be a particularly promising avenue in part due to rapid development in continuous emission monitoring technology. We will present a fast method for the modeling and simulation of methane emission dispersion, and how we use these simulations as a critical building block within a statistical framework for quick emission detection and localization using continuous methane concentration data. In particular, we will highlight the importance of combining approaches from applied math and scientific computing with modern statistics and data science to furnish a practical method for rapid emission detection on oil and gas production facilities. We'll end by discussing some open questions and ongoing challenges with this work and opportunities to get involved.
March 8Jeff Anderson

Title: Ensemble Kalman Filters for Data Assimilation: An Overview and Future Directions

Abstract: The development of numerical weather prediction was a great computational and scientific
achievement of the last century. Models of the PDEs that govern fluid flow and a vast observing
network are required for these predictions. A third vital component is data assimilation (DA)
that combines observations with predictions from previous times to produce initial conditions
for subsequent predictions.
Ensemble Kalman filters are DA algorithms that use a set of predictions to estimate the PDF of
the model state given observations. They are used for weather, but also for many other
geophysical systems, and for applications like disease transmission. They can be extended to
estimate model parameters, guide model improvement, evaluate observation quality and
design future observing systems.
Basic Kalman and ensemble Kalman filter algorithms are reviewed, followed by a discussion of
some heuristic extensions like localization that are required for application to large models.
Recent work to extend ensemble filters to strongly non-Gaussian nonlinear problems will be
discussed. These extensions are particularly beneficial when applying filters for quantities like
rainfall or tracer concentration where the appropriate priors can be best represented by mixed
PDFs; PDFs that are a sum of continuous and discrete functions.
March 15No Colloquium
March 22No Colloquium Due To Spring Break
March 29John Schreck

Title: Evidential Deep Learning: Enhancing Predictive Uncertainty Estimation for Earth System Science Applications

Abstract: Uncertainty quantification is crucial for reliable weather and climate modeling but challenging to achieve. In this seminar, I will demonstrate evidential deep learning, combining probabilistic modeling with deep neural networks, as an effective technique for predictive modeling and calibrated uncertainty estimation. Through atmospheric science classification and regression tasks, we show evidential models attaining predictive accuracy comparable to standard methods while robustly quantifying uncertainty. Uncertainty decomposition provides insights into aleatoric (data variability) and epistemic (model limitations) components. Gaining insights into these distinct uncertainty sources is paramount for enhancing model reliability, utility, and efficiency. We compare the uncertainty metrics derived from evidential neural networks to those obtained from calibrated ensembles, with evidential networks resulting in significant computational savings. Analyses reveal links between uncertainties and underlying meteorological processes, facilitating interpretation. This study establishes deep evidential networks as an adaptable tool for augmenting neural network predictions across geoscience disciplines, overcoming limitations of prevailing approaches. With the ability to produce trustworthy uncertainties alongside predictions, evidential learning has the potential to transform weather and climate modeling, aiding critical analysis and decision-making under uncertainty.
April 5Jennifer Mueller

Title: Electrical impedance tomography for pulmonary imaging: from inverse problems to the clinic

Abstract: Electrical impedance tomography (EIT) is a non-invasive, non-ionizing imaging technique that produces real-time functional images of ventilation and pulmonary perfusion at the bedside. The inverse problem of EIT is to determine the spatially varying conductivity, which arises as a coefficient in the generalized Laplace equation, from measurements of the voltages that arise on electrodes on the surface of the body from applied currents on those same electrodes. The mathematical problem is also known as the Calderon problem of reconstructing the conductivity coefficient from the Neumann-to-Dirichlet map. My lab at CSU is focused on collaborations with engineers and physicians to develop EIT technology for clinical use. In this talk, I will discuss the mathematics of the inverse problem of forming real-time images as well as clinical applications in collaboration with Children's Hospital Colorado and Anschutz Hospital in Aurora. Results from patient data with chronic and critical lung disease will be shown and discussed.
April 12No Colloquium Due To E-Days
April 19Brian Reich

Title: Bayesian computational methods for spatial models with intractable likelihoods

Abstract: Extreme value analysis is critical for understanding the effects of climate change. Exploiting the spatiotemporal structure of climate data can improve estimates by borrowing strength across nearby locations and provide estimates of the probability of simultaneous extreme events. A fundamental probability model for spatially-dependent extremes is the max-stable processes. While this model is theoretically justified, it leads to an intractable likelihood function. We propose to use deep learning to overcome this computational challenge. The approximation is based on simulating millions of draws from the prior and then the data-generating process, and then using deep learning density regression to approximate the posterior distribution. We verify through extensive simulation experiments that this approach leads to reliable Bayesian inference, and discuss extensions to other spatial processes with intractable likelihoods including the autologistic model for binary data and SIR model for the spread of an infectious disease.
April 26Doug Nychka

Title: Hybrid L1 and L2 Smoothing

Abstract: Spline smoothing, and more generally Gaussian process smoothing, have become a successful methodology for estimating a smooth trend or surface from noisy data. Similarly, the LASSO and related L1 penalties have become important tools for variable selection and also admit of a Bayesian version based on the Laplace distribution. This project combines these two approaches as a method to detect discontinuous behavior in an otherwise smooth signal. Every day the Foothills Facility of Denver Water filters more than 250 million gallons of water for the metropolitan area. This process runs continuously and is monitored across an array of filters, each the size of a small swimming pool, at 5-minute intervals. It is important to be able detect anomalous behavior in a filter in a prompt manner or to process past measurements to determine trends. The anomalies take the form of discontinuities or appear as step changes in the smooth filtering cycle. This application is the motivation for a mixed smoothing approach where normal operation is captured by a smoothing spline and the anomalies by basis function coefficients determined by an L1 penalty. As part of this research a frequentist penalty method is compared against its equivalent Bayesian hierarchical model (BHM) based on Gaussian processes and a Laplace prior for the anomaly coefficients. This talk will discuss some of the challenges in implementing both models. Specifically, we study how to choose penalty parameters for the frequentist model and how to formulate then BHM in a way that the MCMC sampling algorithm mixes efficiently. Both approaches appear to classify anomalies in the filter cycles well with the spline model being much faster but the BHM providing measures of uncertainty in the detected anomalies. The similarities between these frequentist and Bayesian models relies on the correspondence between splines and Gaussian processes. This was first described by Grace Wahba, a long-time faculty member of the UW statistics department, and George Kimeldorf. Some background for this connection will be given as part of developing the Bayesian model.
May 3Raul Perez Pelaez
May 10No Colloquium
Fall 2023
September 1Laura Albright - Graduate Student Colloquium
September 5 (Tuesday) at 3:30 pmSophie Marbach

Title: The Countoscope: Counting particles in boxes to probe individual and collective dynamics

Abstract: Any imaging technique is limited by its field of view. As objects or particles move in and out of the observation field, tracking their motion, especially over long periods, becomes challenging. Here, we shift this paradigm by introducing a technique that deliberately harvests the limited field of view. We divide images into observation boxes and count the particles in each box. By analyzing the statistical properties of the number of particles, with varying observation box sizes, we show that we can infer the kinetic properties of the particles, such as their diffusion coefficient, without relying on particle tracking. We use a combination of experiments on colloidal suspensions, simulations with fluctuating hydrodynamics and analytical theory to support our findings. By investigating suspensions with increasing packing fraction, we show how box counting can probe, beyond the self-diffusion coefficient, hydrodynamic and steric effects, and collective motion. We extend our technique to various suspensions, such as ions or active particles. The ​"Countoscope" offers the unique possibility to systematically link individual and collective behavior, opening up broad soft matter and statistical physics perspectives.
September 15Rob Webber

Title: Rocket-propelled Cholesky: Addressing the challenges of large-scale kernel computations.

Abstract: Kernel methods are used for prediction and clustering in many data science and scientific computing applications, but applying kernel methods to a large number of data points N is expensive due to the high cost of manipulating the N x N kernel matrix. A basic approach for speeding up kernel computations is low-rank approximation, in which we replace the kernel matrix A with a factorized approximation that can be stored and manipulated more cheaply. When the kernel matrix A has rapidly decaying eigenvalues, mathematical existence proofs guarantee that A can be accurately approximated using a constant number of columns (without ever looking at the full matrix). Nevertheless, for a long time designing a practical and provably justified algorithm to select the appropriate columns proved challenging.

Recently, we introduced RPCholesky ("randomly pivoted" or "rocket-propelled" Cholesky), a natural algorithm for approximating an N x N positive semidefinite matrix using k adaptively sampled columns. RPCholesky can be implemented with just a few lines of code; it requires only (k + 1) N entry evaluations and O(k^2 N) additional arithmetic operations. In experiments, RPCholesky matches or improves on the performance of alternative algorithms for low-rank psd approximation. Moreover, RPCholesky provably achieves near-optimal approximation guarantees. The simplicity, effectiveness, and robustness of this algorithm strongly support its use for large-scale kernel computations.
September 22Fatemeh Pourahmadian

Title: Recent progress in inverse elastic scattering

Abstract: The first part of this talk highlights recent laboratory applications of sampling approaches to inverse scattering with a particular focus on the linear sampling method and its generalized form for reconstruction from noisy measurements. For this purpose, I leverage two types of experiments; the first setup is designed to mimic fracking with the aim of seismic sensing of evolving fractures in rock, while the other setup pertains to laser ultrasonic testing for characterization of additively manufactured components. I will also include some preliminary results on our recent efforts to potentially augment and automate the inversion process via deep learning. This would set the stage for the second part of this talk which is mostly theoretical and dedicated to differential evolution indicators for imaging evolving processes in unknown environments.
September 29Ernest Ryu

Title: Toward a Grand Unified Theory of Accelerations in Optimization and Machine Learning

Abstract: Momentum-based acceleration of first-order optimization methods, first introduced by Nesterov, has been foundational to the theory and practice of large-scale optimization and machine learning. However, finding a fundamental understanding of such acceleration remains a long-standing open problem. In the past few years, several new acceleration mechanisms, distinct from Nesterov’s, have been discovered, and the similarities and dissimilarities among these new acceleration phenomena hint at a promising avenue of attack for the open problem. In this talk, we discuss the envisioned goal of developing a mathematical theory unifying the collection of acceleration mechanisms and the challenges that are to be overcome.
October 6Emily King

Title: Interpretable, Explainable, and Adversarial AI: Data Science Buzzwords and You (Mathematicians)

Abstract: Many state-of-the-art methods in machine learning are black boxes which do not allow humans to understand how decisions are made. In a number of applications, like medicine and atmospheric science, researchers do not trust such black boxes. Explainable AI can be thought of as attempts to open the black box of neural networks, while interpretable AI focuses on creating white boxes. Adversarial attacks are small perturbations of data, often images, that cause a neural network to misclassify the data. Such attacks are potentially very dangerous when applied to technology like self-driving cars. After a gentle introduction to these topics and data science in general, a sampling of methods from geometry, linear algebra, and harmonic analysis to attack these issues will be presented.
October 13No Colloquium
October 20Kate Bubar

Title: Fundamental limits to the effectiveness of traveler screening with molecular tests

Abstract: Screening airport travelers during an emerging infectious disease outbreak is a common approach to limit the geographical spread of infection. Previous modeling work has explored the effectiveness of screening travelers for symptoms or exposure risk for a variety of pathogens. We developed a probabilistic modelling framework to build on this past work via three main contributions, (1) estimating the effectiveness of screening with molecular tests (e.g., PCR or rapid tests), (2) integrating important heterogeneities in individuals’ detectability and infectiousness with temporal within-host viral kinetics models, and (3) quantifying the fundamental limits of traveler screening. In this talk, I will describe the relevant biological and epidemiological background, our modelling approach and analysis, and the implications for public health policy makers.
October 26Matt Picklo - Graduate Student Colloquium
October 27Megan Wawro

Title: The Inquiry-Oriented Linear Algebra Project

Abstract: The goal of the Inquiry-Oriented Linear Algebra (IOLA) project is to promote a research-based, student-centered approach to the teaching and learning of introductory linear algebra at the university level. Based on the instructional design theory of Realistic Mathematics Education, the IOLA curricular materials build from a set of experientially real tasks that allow for active student engagement in the guided reinvention of key mathematical ideas through student and instructor inquiry. The online instructional support materials include various resources such as rationales for task design, implementation suggestions, and examples of typical student work. In this talk, I will share some IOLA tasks and associated examples of student reasoning, as well as some guiding principles for inquiry-oriented instruction.
November 3Elizabeth Newman

Title: Training Made Easy: Harnessing Structure and Curvature Information to Train Neural Networks Efficiently

Abstract: Deep neural networks (DNNs) have achieved inarguable success as high-dimensional function approximators in countless applications. However, this success comes at a significant hidden expense: the cost of training. Typically, the training problem is posed as a stochastic optimization problem with respect to the learnable DNN weights. With millions of weights, a non-convex objective function, and many hyperparameters to tune, solving the training problem well is no easy task.

In this talk, we will present new algorithms that make DNN training easier by exploiting common structure, automating hyperparameter tuning, and computing curvature information efficiently. We will first focus on training separable DNNs; that is, architectures for which the weights of the final layer are applied linearly. We will leverage this structure first in a deterministic setting by eliminating the linear weights through variable projection (i.e., partial optimization). Then, we will extend to a stochastic setting using a powerful iterative sampling approach, which notably incorporates automatic regularization parameter selection. Time-permitting, we will discuss up-and-coming work that introduces a memory and computationally efficient Gauss-Newton optimizer for training large-scale DNN models rapidly. Throughout the talk, we will demonstrate the efficacy of these approaches through numerical examples.
November 10Michael Ivanitskiy - Graduate Student Colloquium
November 17Andrew Zammit Mangion

Title: Bayesian Neural Networks for Spatial Process Modelling

Abstract: Statistical models for spatial processes play a central role in statistical analyses of spatial data. Yet, it is the simple, interpretable, and well understood models that are routinely employed even though, as is revealed through prior and posterior predictive checks, these can poorly characterise the spatial heterogeneity in the underlying process of interest. In this talk I will propose a new, flexible class of spatial-process models, which I refer to as spatial Bayesian neural networks (SBNNs). An SBNN leverages the representational capacity of a Bayesian neural network; it is tailored to a spatial setting by incorporating a spatial "embedding layer" into the network and, possibly, spatially-varying network parameters. An SBNN is calibrated by matching its finite-dimensional distribution at locations on a fine gridding of space to that of a target process of interest. That process could be easy to simulate from or we have many realisations from it. I will be formulating several variants of SBNNs, most of which are able to match the finite-dimensional distribution of the target process at the selected grid better than conventional BNNs of similar complexity. I will show that a single SBNN can remarkably be used to represent a variety of spatial processes often used in practice, such as Gaussian processes and lognormal processes. I will briefly discuss the tools that could be used to make inference with SBNNs, and will conclude with a discussion on their advantages and limitations.
November 24No Colloquium
November 30Heather Zinn Brooks

Title: The Mathematics of Opinion Dynamics

Abstract: Given the large audience and the ease of sharing content, the shifts in opinion driven by online interaction have important implications for interpersonal interactions, public opinion, voting, and policy. There is a critical and growing demand to understand the mechanisms behind the spread of content online. While the majority of the research on online content focuses on these phenomena from an empirical or computational perspective, mechanistic mathematical modeling also has an important role to play. Mathematical models can help develop a theory to understand the mechanisms underpinning the spread of content and diffusion of information. These models provide an excellent framework because they are often relatively simple models with surprisingly rich dynamics. In this talk, I’ll introduce you to a variety of mathematical models for opinion dynamics, and I’ll highlight some particular problems that we study in my research group.
December 1Graduate Student Colloquium
December 8No Colloquium
December 15No Colloquium
Spring 2023
January 13Siting Liu

Title: An Inverse Problem in Mean Field Games from Partial Boundary Measurement

Abstract: In this talk, we consider a novel inverse problem in mean-field games (MFG). We aim to recover the MFG model parameters that govern the underlying interactions among the population based on a limited set of noisy partial observations of the population dynamics under the limited aperture. Due to its severe ill-posedness, obtaining a good quality reconstruction is very difficult. Nonetheless, it is vital to recover the model parameters stably and efficiently to uncover the underlying causes of population dynamics for practical needs.

 Our work focuses on the simultaneous recovery of running cost and interaction energy in the MFG equations from a finite number of boundary measurements of population profile and boundary movement. To achieve this goal, we formalize the inverse problem as a constrained optimization problem of a least squares residual functional under suitable norms. We then develop a fast and robust operator splitting algorithm to solve the optimization using techniques including harmonic extensions, three-operator splitting scheme, and primal-dual hybrid gradient method. Numerical experiments illustrate the effectiveness and robustness of the algorithm.

This is based on joint work with Yat Tin Chow, Samy Wu Fung, Levon Nurbekyan, and Stanley J. Osher.
January 20Lyndsay Shand (In-Person)

Title: An autotuning approach to DOE’s earth system model

Abstract: Calibration is the science (and art) of matching the model to observed data. Global Climate Model (GCM) calibration is a multi-step process done by hand and is a tedious and time-consuming process. GCM calibration involves both high-dimensional input and output spaces. Many rigorous calibration methods have been proposed in both statistical and climate literature, but many are not practical to implement. In this talk, I will demonstrate a promising and practical calibration approach on the atmosphere only model of the Department of Energy’s Energy Exascale Earth System Model (E3SM). We generate a set of designed ensemble runs that span our input parameter space and fit a surrogate model using polynomial chaos expansions on a reduced space. We then use surrogate in an optimization scheme to identify input parameter sets that best match our simulated output to observations. Finally, we run E3SM with the optimal parameter set and compare prediction results across 44 spatial fields to the hand-tuned optimal parameter set chosen by experts. This flexible approach is straightforward to implement and seems to do as well or better than the tuning parameters chosen by the expert while considering high-dimensional output and operating in a fraction of the time.
January 27Shelby Stowe - Graduate Student Colloquium
February 3Dave Montgomery - Graduate Student Colloquium
February 10Nathan Lenssen

Title: What will the weather be next year? How we can (and can’t) predict the chaotic evolution of the climate system.

Abstract: The Earth’s atmosphere and ocean are coupled chaotic nonlinear dynamical systems that drive the weather and long-term climate patterns we experience. Variability in the climate, or changes in the distribution of weather, can lead to increased chances of climate extremes such as drought and weather extremes such as hurricanes. The El Niño-Southern Oscillation (ENSO) is the dominant source of subseasonal, seasonal, and multi-year climate variability, driving changes in climate and weather worldwide. ENSO is also the dominant source of theoretical predictability on these timescales, giving us the opportunity to predict the weather in the coming months and years.

This seminar will discuss the history of modeling and predicting ENSO as a dynamical system from the conception in the 1980s up to the state-of-the art. We will show results from ongoing research into predicting ENSO up to 2 years in advance. Open questions in climate prediction will be discussed with an emphasis on possible applications for applied mathematics, statistics, and machine learning.
February 17Matthias Katzfuss

Title: Scalable Gaussian-Process Inference via Sparse Inverse Cholesky Factorization.

Abstract: Gaussian processes (GPs) are popular, flexible, and interpretable probabilistic models for functions in geospatial analysis, computer-model emulation, and machine learning. However, direct application of GPs involves dense covariance matrices and is computationally infeasible for large datasets. We consider a framework for fast GP inference based on the so-called Vecchia approximation, which implies a sparse Cholesky factor of the inverse covariance matrix. The approximation can be written in closed form and computed in parallel, and it includes many popular existing approximations as special cases. We discuss various applications and extensions of the framework, including high-dimensional inference and variational approximations for latent GPs.
February 24Dan Cooley (In-Person)

Title: Transformed-Linear Methods for Multivariate Extremes and Application to Climate

Abstract: Statistical methods for extremes are widely used in climate science. Distributions like the generalized extreme value and generalized Pareto are familiar tools used by climate scientists to describe the extreme behavior of univariate data. Multivariate (and spatial and time series) extremes largely focuses on accurately capturing the tail dependence of several variables. Multivariate extremes models can be complicated and can be difficult to fit in high dimensions. In this talk, we will use methods from classical (non-extreme) statistics as inspiration to create sensible methods for multivariate extremes. Many classical statistical methods (e.g., PCA, spatial, factor analysis, and time series) employ the covariance matrix to learn about dependence, to construct models, or to perform prediction. Most familiar statistics methods are linear. However, extremal dependence is poorly described by the covariance matrix. Linear methods have not been widely employed for extremes, and are difficult to implement for data that are non-negative.

In this talk, we will introduce transformed linear methods for extremes. By using the tail pairwise dependence matrix (TPDM) in place of the covariance matrix, and by employing transformed linear operations, extreme analogues can be developed for familiar linear statistical methods. Here, we will focus on developing transformed linear time series models to capture dependence in the upper tail. These models are extremal analogues to familiar ARMA models. We apply these models to perform attribution for seasonal wildfire conditions. To focus on change in fire risk due to climate, we model the fire weather index (FWI) time series. We use our fitted model to perform an attribution study. According to our fitted model, the 2020 Colorado fire season is many times more likely to occur under recent climate than under the climate of 50 years ago. If time allows, we will also present results from a PCA analysis of CONUS extreme precipitation.
March 3No Colloquium
March 10Steve Pankavich

Title: Kinetic Models of Collisionless Plasmas

Abstract: Collisionless plasmas arise in a variety of settings, ranging from magnetically confined plasmas to study thermonuclear energy to space plasmas in planetary magnetospheres and solar winds. The two fundamental models that describe such phenomena are systems of nonlinear partial differential equations known as the Vlasov-Maxwell (VM) and Vlasov-Poisson (VP) systems. We will derive these kinetic models and discuss the possibility of shocks arising from a continuous distribution of particles. In the process of this investigation, it will be important to delineate the difference between the mathematical formulation of a shock and related phenomena often described by physicists. Finally, we will describe the stability and instability of velocity-dependent steady states in plasmas and discuss some recent computational results regarding their behavior.
March 17Christian Parkinson (Remote)

Title: The Hamilton-Jacobi Formulation of Optimal Path Planning for Autonomous Vehicles

Abstract: We present a partial-differential-equation-based optimal path planning framework for simple self-driving cars. This formulation relies on optimal control theory, dynamic programming, and a Hamilton-Jacobi-Bellman equation, and thus provides an interpretable alternative to black-box machine learning algorithms. We design grid-based numerical methods used to resolve the solution to the Hamilton-Jacobi-Bellman equation and generate optimal trajectories. We then describe how efficient and scalable algorithms for solutions of high dimensional Hamilton-Jacobi equations can be developed to solve similar problems in higher dimensions and in nearly real-time. We demonstrate our methods with several examples.
March 24No Colloquium
March 31Tusharkanti Ghosh (In-Person)

Title: Bayesian Hierarchical Hidden Markov Models for Identifying Differentially-Methylated Cytosines from Bisulfite-Sequencing Data

Abstract: DNA methylation is a crucial epigenetic mechanism for controlling gene expression, silencing, and genomic imprinting in living cells, and aberrant methylation has been associated with a variety of important biological processes and disease, including ageing and cancer. Recent developments in ultra-high throughput sequencing technologies and the massive accumulation of sequencing data bring the hope of understanding the workings of DNA methylation at a molecular level, however, these data pose significant challenges in modelling and analysis. In this talk, I discuss how we developed a Bayesian statistical framework and methodology for the identification of differential patterns of DNA methylation between different groups of cells, focusing on a study of human ageing. Our approach develops and extends a class of Bayesian hierarchical hidden Markov models (HHMMs) that can accommodate various degrees of dependence among the sequence-level measurements, both within and across the sequences, and provides the ability to select between competing alternative models for the most appropriate one for a specific methylation data set. Our proposed methodology to determine differentially-methylated Cytosines (DMCs) is implemented through a fast and efficient hybrid Markov chain Monte Carlo algorithm, and we demonstrate how it significantly improves correct prediction rates, with a reduced false discovery rate, compared to several existing methods for DMC detection.
April 7Olivia Walch (In-Person)

Title: Circadian Interventions in Shift Workers: Translating math to the real world

Abstract: Shift workers experience profound circadian disruption due to the nature of their work, which often has them on-the-clock at times when their internal clock is sending a strong, sleep-promoting signal. This constant disruption of their sleep and circadian rhythms can put them at risk of injury and development of long term chronic disease. Mathematical models can be used to generate recommendations for shift workers that move their internal clock state to better align with their work schedules, promote overall sleep, promote alertness at key times, or achieve other desired outcomes. Yet for these schedules to have a positive effect in the real world, they need to be acceptable to the shift workers themselves. In this talk, I will survey the types of schedules a shift worker may be recommended by an algorithm, and how they can collide with the preferences of the real people being asked to follow them, and how math can be used to arrive at new schedules that take these human factors into account.
April 14Michael Ivanitskiy - Graduate Student Colloquium
April 21Keaton Hamm (In-Person)

Title: Optimal Transport Based Manifold Learning

Abstract: We will discuss the use of optimal transport in the setting of nonlinear dimensionality reduction and applications to image data. We illustrate the idea with an algorithm called Wasserstein Isometric Mapping (Wassmap) which works for data that can be viewed as a set of probability measures in Wasserstein space. The algorithm provides a low-dimensional, approximately isometric embedding. We show that the algorithm is able to exactly recover parameters of some image manifolds including those generated by translations or dilations of a fixed generating measure. We will discuss computational speedups to the algorithm such as use of linearized optimal transport or the Nystr\"{o}m method. Testing of the proposed algorithms on various image data manifolds show that Wassmap yields good embeddings compared with other global and local techniques.
April 28Elizabeth Barnes (In-Person)

Title: Explainable AI for Climate Science: Opening the Black Box to Reveal Planet Earth

Abstract: Earth’s climate is chaotic and noisy. Finding usable signals amidst all of the noise can be challenging: be it predicting if it will rain, knowing which direction a hurricane will go, understanding the implications of melting Arctic ice, or detecting the impacts of human-induced climate warming. Here, I will demonstrate how explainable artificial intelligence (XAI) techniques can sift through vast amounts of climate data and push the bounds of scientific discovery: allowing scientists to ask “why?” but now with the power of machine learning.
May 5Rebecca Morrison (In-Person)