Fast sparse Gaussian elimination over an arbitrary field and applications to quantum stabilizer states
We present a new LDL factorization algorithm that requires O(nt^{omega-1}) field operations, where t is the treewidth of the graph associated with the nonzeros of a full-rank sparse matrix. This algorithm matches the best known bounds for Cholesky and extends to LU. For low-rank matrices or rectangular LU, our algorithm also gives the factorization in implicit form. Obtaining an explicit form of low rank factors in O(nt^{\omega-1}) time remains an open problem. The main idea of the new algorithm is based on iterative application of a block factorization of saddle-point systems that is equivalent to the null space method and to LDL. We also discuss connections of tree-decomposition-based LDL and butterfly/hierarchical low-rank factorization of the inverse of a sparse matrix.
We then discuss the relationship between LDL and the representation of quantum stabilizer states, which enable polynomial-time simulation of Clifford circuits. In particular, we provide a simple explicit formula for the amplitudes of a Hadamard-transformed graph state, based on a modified Gauss-Jordan matrix inversion process that is closely related to LDL factorization. This formula enables us to directly characterize all pairs of graphs states that are equivalent under application of local Clifford unitaries, as well as match the best known bounds for Clifford circuit simulation. Overall, the LDL and stabilizer-state results reverse the construction of Gosset, Grier, Kerzner, and Schaeffer (Quantum 2024), who provide an O(nt^{omega-1})-time LDL algoithm for solving symmetric linear systems over GF(2) by reduction to a tree-decomposition-based Clifford circuit simulation algorithm.
Matrix Factorizations with Uniformly Random Pivoting
What do Jacobi’s eigenvalue algorithm and an iterative algorithm for finding the QR decomposition have in common? Under a randomized pivoting strategy, a unified analysis can show that these algorithms converge with the same linear rate of convergence. We can view these algorithms, as well as iterative algorithms for SVD and Cholesky decomposition, as elements of a more general class of matrix factorizations. I will define this general class of matrix factorizations and discuss how to prove anything in this class converges under a randomized pivoting strategy.
Matrix Decomposition and Function Approximation in Quantum Many-Body Physics Calculations
In this talk, I will discuss how techniques from numerical linear algebra have played a crucial role in many-body Green’s function calculations in condensed matter physics and quantum chemistry. The main focus will be on the problem of hybridization fitting. I will begin by reviewing how progress in this area has advanced imaginary-time Green’s function and Feynman diagram computations, emphasizing the roles of singular value decomposition (SVD), interpolative decomposition (ID), rational approximation, and sum-of-exponentials (SOE) approximation. I will then turn to more recent developments concerning real-time calculations. Finally, I will highlight several open problems that are directly related to physical applications.
Numerical Solution of Double Saddle-Point Systems
Double saddle-point systems are drawing increasing attention in the past few years, due to the importance of multiphysics applications and recent advances in the solution of indefinite linear systems. In this talk I will describe some of the numerical properties of these linear systems. Eigenvalue bounds will be presented, and we will show what role Schur complements and nested Schur complements play in preconditioning. Some numerical experiments on the Stokes-Darcy equations will be described.
When is accurate and efficient expression evaluation and linear algebra possible?
When, and at what cost, can we solve linear algebra problems, or even just evaluate multivariate polynomials, with high relative accuracy? Suppose we only know that we can perform the 4 basic operations (+, -, *, /) with high relative accuracy, i.e. rnd(a op b) = (a op b)*(1+delta) with |delta| <= epsilon << 1. Which 2 of the following 3 problems can we solve with high relative accuracy?
(1) Evaluate the Motzkin polynomial p(x,y,z) = z^6 + x^2*y^2*(x^2 + y^2 - 3*z^2)
(2) Compute the eigenvalues of the Cauchy matrix C(i,j) = 1/(x(i) + x(j)), given 0 < x(1) < x(2) < ... , an example of which is the Hilbert matrix.
(3) Evaluate x+y+z.
We will answer this question, and many related ones, and talk about the open problem of creating a decision procedure that would take a polynomial, and decide whether an accurate algorithm for it exists. Properties of the variety of the polynomial will play a central role. We will also discuss some properties of floating point arithmetic that will let us expand the list of basic operations above, and so expand the set of problems we can solve accurately.
Markov chains, graph Laplacians, and column subset selection
The column subset selection problem (CSSP) is an essential target of research in low-rank approximation. I will describe how our particular motivation in reduced order models of Markov chains has led to new theory and algorithms for this long-standing problem. Our problem setting, relevant to ubiquitous dynamical systems in chemistry, physics, and computer science, is nicely mappable to solving the CSSP for the (suitably regularized) inverse of the respective graph Laplacian. Using contour integral and probabilistic arguments, I will first show how the dynamical error of reversible Markov chain compression can be formulated and simply bounded in terms of Nyström approximation. Then I will describe how nuclear maximization is an especially attractive method for the CSSP in this setting, showing new matrix-free algorithms based on randomized diagonal estimation via approximate Cholesky factorization. I will close by demonstrating our algorithms' empirical performance and theoretical guarantees, detailing a novel combination of ideas from submodularity, probability theory, and determinantal point processes. Time permitting, I will outline our successes in extending our methods and error analysis to general CX, Nyström, and CUR approximation.
Rank-revealing factorizations: applications, algorithms, and theory
Rank-revealing matrix factorizations play a key role in applications ranging from structured low-rank approximation to computing localized basis functions in computational quantum chemistry. We will discuss some motivating applications to illustrate the breadth of desiderata for such factorizations. This will be followed by a brief overview of algorithms (where we will restrict our discussion to pivoted QR factorizations for simplicity) and highlights of historical and recent theoretical developments. We will close by highlighting some open problems in the area.
A randomized linear algebra perspective on geometric optimization methods in scientific machine learning
Subsampled natural gradient (SNG) and Gauss-Newton (SGN) methods have demonstrated impressive performance for parametric optimization problems in scientific machine learning, including neural network wavefunctions and physics-informed neural networks. However, their development has been hindered by a lack of theoretical understanding. To make progress on this problem, we study these algorithms on a simplified parametric optimization problem involving a linear model and a quadratic loss function. In this setting we show that both SNG and SGN can be interpreted as sketch-and-project methods and can thus enjoy rigorous convergence guarantees for arbitrary batch sizes. This interpretation also explains a recently proposed accelerated variant of SNG and clarifies how this accelerated variant can be extended to the setting of SGN. Finally, these explorations inspire a number of new questions about sketch-and-project algorithms, some of which have been resolved and many of which remain open.
Use of Johnson–Lindenstrauss (JL) sketching operators in faster hierarchical matrix construction
We present an adaptive, partially matrix-free, hierarchical matrix construction framework using a broader class of Johnson–Lindenstrauss (JL) sketching operators. On the theoretical side, we extend the earlier concentration bounds to all JL sketching operators and examine this bound for specific classes of such operators including the original Gaussian sketching operators, subsampled randomized Hadamard transform (SRHT) and the sparse Johnson-Lindenstrauss transform (SJLT). We discuss the implementation details of applying SJLT and SRHT efficiently. Then we demonstrate experimentally that using SJLT or SRHT instead of Gaussian sketching operators leads to up to 2.5 times speedups of the serial HSS construction implementation in the STRUMPACK C++ library. Additionally, we discuss the implementation of a parallel distributed HSS construction that leverages Gaussian or SJLT sketching operators. We observe a performance improvement of up to 35 times when using SJLT sketching operators over Gaussian sketching operators. The generalized algorithm allows users to select their own JL sketching operators with theoretical lower bounds on the size of the operators which may lead to faster runtime with similar HSS construction accuracy.
Obstacles to a Satisfying Convergence Theory for Restarted Arnoldi with Exact Shifts
The restarted Arnoldi method iteratively applies polynomial filters to enhance the orientation of the starting vector toward the desired invariant subspace. At each step the roots of these filters are the Ritz values that least resemble the desired eigenvalues. For Hermitian eigenvalue problems, Sorensen proved that this procedure must lead to convergence (under mild and natural conditions). However, one can construct non-Hermitian examples where a filter polynomial root falls precisely on a desired eigenvalue, making convergence impossible in theory. To handle such examples, we require a deeper understanding of how Ritz eigenvalue estimates are distributed in the numerical range. Given a set of points {theta_1, ..., theta_k} in the numerical range of the n-by-n matrix A, does there exist some n-by-k matrix Q with orthonormal columns for which the spectrum of Q'*A*Q is {theta_1, ..., theta_k}?
Revisiting MR^3 for the Bidiagonal SVD
The Multiple Relatively Robust Representations (MRRR or MR^3) algorithm is optimal for the symmetric tridiagonal eigenvalue problem -- that is, it can compute k eigenpairs (with numerically orthogonal eigenvectors) in only O(nk) operations. Accordingly, MR^3 yields the fastest algorithm for the symmetric eigenvalue problem when paired with standard tridiagonal reduction. Given that the bidiagonal SVD and symmetric tridiagonal eigenvalue problems are two sides of the same coin -- and indeed the fastest algorithms for computing an arbitrary SVD begin with bidiagonal reduction -- it seems natural that MR^3 should imply an optimal algorithm for this problem as well. Nevertheless, it has yet to be successfully deployed in the SVD setting. This is in spite of the fact that one of the key insights of MR^3 is that representing a tridiagonal as a product of bidiagonals ensures high accuracy. In this talk we revisit this problem, tracing the origins of MR^3, the pitfalls of using it to compute the bidiagonal SVD, and the many efforts to make it work anyways.
Kronecker Matrix-Vector Complexity is Weird (and exponentially bad)
Matrix-vector complexity has been a fruitful paradigm in the design and analysis of efficient matrix algorithms. For tensor methods, a natural generalization of the matrix-vector model is the "Kronecker Matrix-Vector Model", and many existing tensor methods belong to this class.
In this talk, we will formalize and motivate this computational model, and study the performance of various NLA algorithms in this model. Along the way, we will discover strange facts, including exponential lower bounds, the fact that Rademacher random variables should not be used in this model, and that we cannot use the subgaussian norm to accurately bound the statistical complexity of our algorithms.
Spectral Learning for Dynamical Systems via Infinite-Dimensional Numerical Linear Algebra
Koopman operators recast nonlinear dynamics as infinite-dimensional linear systems, enabling spectral analysis of time-series data—effectively, data-driven infinite-dimensional numerical linear algebra. Over the past decade, they’ve found widespread use in fields ranging from robot control and climate variability to neural networks, epidemiology, and neuroscience. But which spectral features can actually be reliably computed from finite data—and which cannot? We show that the answer depends critically on the choice of observable space (e.g., L2 vs RKHS) and the nature of the underlying dynamics. In particular, we construct adversarial systems where any algorithm provably fails, even with unlimited data and randomized methods. By identifying when convergence is possible, we design algorithms that succeed both in theory and in practice, overcoming the notorious problem of spurious eigenvalues that undermine prediction. Applications to climate systems—including sea ice and ocean flow—demonstrate how spectral analysis reveals coherent structures and enables state-of-the-art forecasting. This leads to a classification theory that quantifies the difficulty of data-driven problems for dynamical systems and proves the optimality of certain algorithms. I’ll conclude with a look ahead at where this field is going, and the exciting opportunities it presents for both numerical linear algebra and complexity theory.