Seminar
The event has passed

Seminar Geometry, Algebra and Physics in Deep Neural Networks (GAPinDNNs)

Debarghya Ghoshdastidar, TU Munich: Are self-supervised models doing kernel PCA?

Overview

The event has passed
  • Date:Starts 6 March 2025, 13:30Ends 6 March 2025, 14:30
  • Location:
    MV:L15, Chalmers tvärgata 3
  • Language:English

Abstract: The short answer to the title is NO, but there are some theoretical connections between neural networks with self-supervised pre-training and kernel principal component analysis. At a high level, the equivalence is based on two ideas: (i) optimal solutions for many self-supervised losses correspond to spectral embeddings; and (ii) infinite-width neural networks converge to neural tangent kernel (NTK) models.

I will first give a short overview of this equivalence and discuss why it could be useful for both theory and practice of foundation models. I will then discuss two recent works on the NTK convergence under self-supervised losses (arXiv:2403.08673, arXiv:2411.11176). Specifically, I will show that one cannot directly use NTK results from supervised learning / regression, but a careful analysis is needed to prove that the NTK indeed remains constant during self-supervised training.

Jan Gerken
  • Assistant Professor, Algebra and Geometry, Mathematical Sciences