Codes and Expansions (CodEx) Seminar


Robert Webber (Caltech):

Rocket-propelled Cholesky: Addressing the challenges of large-scale kernel computations

Kernel methods are used for prediction and clustering in many data science and scientific computing applications, but applying kernel methods to a large number of data points \(N\) is expensive due to the high cost of manipulating the \(N \times N\) kernel matrix. A basic approach for speeding up kernel computations is low-rank approximation, in which we replace the kernel matrix \(A\) with a factorized approximation that can be stored and manipulated more cheaply. When the kernel matrix  \(A\) has rapidly decaying eigenvalues, mathematical existence proofs guarantee that  \(A\) can be accurately approximated using a constant number of columns (without ever looking at the full matrix). Nevertheless, for a long time designing a practical and provably justified algorithm to select the appropriate columns proved challenging.

Recently, we introduced RPCholesky ("randomly pivoted" or "rocket-propelled" Cholesky), a natural algorithm for approximating an \(N \times N\) positive semidefinite matrix using k adaptively sampled columns. RPCholesky can be implemented with just a few lines of code; it requires only \((k + 1) N\) entry evaluations and \(O(k^2 N)\) additional arithmetic operations. In experiments, RPCholesky matches or improves on the performance of alternative algorithms for low-rank psd approximation. Moreover, RPCholesky provably achieves near-optimal approximation guarantees. The simplicity, effectiveness, and robustness of this algorithm strongly support its use for large-scale kernel computations.

Joint work with Yifan Chen, Ethan Epperly, and Joel Tropp. Available at arXiv:2207.06503.