Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Robust, randomized preconditioning for kernel ridge regression (2304.12465v4)

Published 24 Apr 2023 in math.NA, cs.NA, and stat.ML

Abstract: This paper investigates two randomized preconditioning techniques for solving kernel ridge regression (KRR) problems with a medium to large number of data points ($104 \leq N \leq 107$), and it introduces two new methods with state-of-the-art performance. The first method, RPCholesky preconditioning, accurately solves the full-data KRR problem in $O(N2)$ arithmetic operations, assuming sufficiently rapid polynomial decay of the kernel matrix eigenvalues. The second method, KRILL preconditioning, offers an accurate solution to a restricted version of the KRR problem involving $k \ll N$ selected data centers at a cost of $O((N + k2) k \log k)$ operations. The proposed methods solve a broad range of KRR problems, making them ideal for practical applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Mateo Díaz (20 papers)
  2. Ethan N. Epperly (16 papers)
  3. Zachary Frangella (14 papers)
  4. Joel A. Tropp (64 papers)
  5. Robert J. Webber (30 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets