INNER: information theory of deep neural networks

Over the last decade, deep-learning algorithms have dramatically improved the state of the art in many machine-learning problems, including computer vision, speech recognition, natural language processing, and audio recognition. Despite their success, however, there is no satisfactory mathematical theory that explains the functioning of such algorithms. Indeed, a common critique is that deep-learning algorithms are often used as black box, which is unsatisfactory in all applications for which performance guarantees are critical (e.g., traffic-safety applications).

The purpose of this project is to increase our theoretical understanding of deep neural networks. This will be done by relying on tools of information theory and focusing on specific tasks that are relevant to computer vision.

Samarbetande organisationer

  • Chalmers AI-forskningscentrum (CHAIR) (Centrumbildning, Sweden)
Startdatum 2019-01-01
Slutdatum 2021-12-31

Sidansvarig Publicerad: må 25 nov 2019.