Oskar Allerbo, Chalmers University of Technology & University of Gothenburg: Solving Kernel Ridge Regression with Gradient Descent
Overview
- Date:Starts 11 May 2023, 13:15Ends 11 May 2023, 14:00
- Location:MV:L14, Chalmers tvärgata 3
- Language:English
Abstract: We present an equivalent formulation for the objective function of kernel ridge regression (KRR), that opens up for studying KRR from the perspective of gradient descent. Utilizing gradient descent with infinitesimal step size, allows us to formulate a new regularization for kernel regression through early stopping.
The gradient descent formulation of KRR allows us expand to a time dependent stationary kernel, where we decrease the bandwidth to zero during training. This circumvents the need of hyper parameter selection. Furthermore, we are able to achieve both zero training error and a double descent behavior, phenomena that do not occur for KRR with constant bandwidth, but are known to appear for neural networks.
The new formulation of KRR also enables us to explore other penalties than the ridge penalty. Specifically, we explore the ℓ1 and ℓ∞ penalties and show that these correspond to two flavors of gradient descent, thus alleviating the need of computationally heavy proximal gradient descent algorithms. We show theoretically and empirically how these formulations correspond to signal-driven and robust regression, respectively.
- Senior Lecturer, Applied Mathematics and Statistics, Mathematical Sciences
