Seminar
The event has passed

Seminar Geometry, Algebra and Physics in Deep Neural Networks (GAPinDNNs)

Yonatan Kahn, University of Toronto: Finite-width feature learning in deep networks with orthogonal weight initialization

Overview

The event has passed
  • Date:Starts 3 April 2025, 15:30Ends 3 April 2025, 16:30
  • Location:
    MV:L13, Chalmers tvärgata 3
  • Language:English

Abstract: Fully-connected deep neural networks with weights initialized from independent Gaussian distributions can be tuned to criticality, which prevents the exponential growth or decay of signals propagating through the network. However, such networks still exhibit fluctuations that grow linearly with the depth of the network, which may impair the training of networks with width comparable to depth. I will provide some theoretical and experimental evidence that using weights initialized from the ensemble of orthogonal matrices leads to better training and generalization behavior even for deep networks, and argue that these results demonstrate the practical usefulness of finite-width perturbation theory.

Max Guillen
  • Postdoc, Algebra and Geometry, Mathematical Sciences
Seminar Geometry, Algebra and Physics in Deep Neural Networks (GAPinDNNs) | Chalmers