Explaining Deep Neural Networks using Information Theory and Geometry

Publicerad

Författare

Typ

Examensarbete för masterexamen
Master's Thesis

Modellbyggare

Tidskriftstitel

ISSN

Volymtitel

Utgivare

Sammanfattning

Abstract Deep neural networks (DNNs) have achieved remarkable success in various machine learning tasks, yet a satisfactory explanation for their generalisation performance remains elusive. In this thesis, we stand on the shoulders of giants and explore two extensions to the information bottleneck (IB) principle, which formalises the idea that DNNs are ultimately compression machines which turn input data into minimal sufficient representations. The first extension is based on geometry and replaces the information-theoretic IB with neural collapse (NC), a phenomenon where the penultimate layer representations of the training data converge to a regular simplex, a configuration of points that maximises pairwise distance. The second extension enforces more realistic constraints on the computation of mutual information used in IB, by taking into account the functional family used to decode said information. This leads to the natural definition of đ’±-information. We build on these ideas by extending them to all layers of a DNN. Motivated by realworld computational constraints on computing NC and đ’±-information, we develop a (to our knowledge) novel algorithm for measuring NC, and (re)discover theoretical motivations behind many current practices in DNN training. Experiments indicate that compression, both geometric and information-theoretic, is not necessary for generalisation, yet DNNs do compress in a predictable manner, with compression increasing as one moves deeper into the network.

Beskrivning

Ämne/nyckelord

Keywords: neural collapse, information theory, decodable information bottleneck, compression, generalisation

Citation

Arkitekt (konstruktör)

Geografisk plats

Byggnad (typ)

ByggÄr

Modelltyp

Skala

Teknik / material

Index

item.page.endorsement

item.page.review

item.page.supplemented

item.page.referenced