# Projects

##### Here are a few ongoing research projects:

Randomized low-rank matrix approximation is an approach for compressing a large matrix to a low-rank factored form while preserving the data as accurately as possible. Randomized low-rank matrix approximation can accelerate machine learning algorithms for prediction and clustering, making it a vital tool for modern data science.

With Ethan Epperly and Joel Tropp, I am developing faster, more accurate algorithms for randomized low-rank matrix approximation, including "randomized block Krylov iteration" for general matrices and "randomly pivoted Cholesky" for positive semidefinite kernel matrices.

For more details, see:

*Randomly pivoted Cholesky: Practical approximation of a kernel matrix with few entry evaluations*[arXiv]- 2022 seminar talk on YouTube

Rare events can be highly impactful. Yet, estimating the probability *p* of a rare event by direct numerical simulation requires a very large sample size (>100p^{-1}). Since generating such a large sample can be prohibitively expensive, are there more practical methods for calculating rare event probabilities?

To help answer this question, I have developed several "splitting and killing" algorithms for rare event probability estimation. These algorithms "split" selected trajectories to promote progress toward a rare event and randomly "kill" other trajectories to control the computational cost.

With Dorian Abbot, Sam Hadden, and Jonathan Weare, I recently applied splitting and killing to evaluate the probability that Mercury will become unstable and collide with another celestial body over the next 2.2 billion years. We calculated the probability to be ~10^{-4} and obtained a speed-up of nearly 100x over direct numerical simulation.

For more details, see:

- 2023 SIAM News blogpost about our work
*Rare event sampling improves Mercury instability statistics,*[Astrophysical Journal, 2021]*Practical rare event sampling for extreme mesoscale weather,*[Chaos, 2019]- 2022 seminar talk on YouTube

The ground state and the first few excited states determine the fundamental properties of a quantum system at low temperatures. However, as the system size increases, it becomes exponentially more difficult to calculate eigenstates using traditional numerical methods. To address this curse of dimensionality, I helped develop two modern methods that harness the power of Monte Carlo sampling.

- With Michael Lindsey, I introduced the Rayleigh-Gauss-Newton (RGN) method, which uses Monte Carlo sampling to efficiently optimize a neural network model for the ground-state wavefunction.
- With Timothy Berkelbach, Samuel Greene, and Jonathan Weare, I helped develop Fast Randomized Iteration (FRI), which provides stochastic estimates for the dominant eigenvalues and eigenvectors of a large matrix.

My collaborators and I have applied RGN to spin systems with up to 400 spins (hence 2^{400} possible spin configurations) and have applied FRI to molecules as large as oxo-Mn(salen) (which has 28 interacting electrons).

For more details, see:

*Rayleigh-Gauss-Newton optimization with enhanced sampling for variational Monte Carlo*, [Physical Review Research, 2022]*Approximating matrix eigenvalues by subspace iteration with repeated random sparsification*, [SIAM Journal on Scientific Computing, 2022]- 2022 tutorial talk and conference talk on YouTube