Abstract / Description of output
Deep learning models have seen significant successes in numerous applications, but their inner workings remain elusive. The purpose of this work is to quantify the learning process of deep neural networks through the lens of a novel topological invariant called magnitude. Magnitude is an isometry invariant; its properties are an active area of research as it encodes many known invariants of a metric space. We use magnitude to study the internal representations of neural networks and propose a new method for determining their generalisation capabilities. Moreover, we theoretically connect magnitude dimension and the generalisation error, and demonstrate experimentally that the proposed framework can be a good indicator of the latter.
Original language | English |
---|---|
Title of host publication | Proceedings of 2nd Annual Workshop on Topology, Algebra, and Geometry in Machine Learning (TAG-ML) |
Publisher | Journal of Machine Learning Research: Workshop and Conference Proceedings |
Pages | 242-253 |
Number of pages | 11 |
Volume | 221 |
Edition | TAG-ML, ICML |
Publication status | Published - 28 Jul 2023 |
Event | The Fortieth International Conference on Machine Learning - Honolulu, United States Duration: 23 Jul 2023 → 29 Jul 2023 Conference number: 40 https://icml.cc/ |
Publication series
Name | Proceedings of Machine Learning Research |
---|---|
Publisher | MIT-Press |
Volume | 221 |
ISSN (Electronic) | 2640-3498 |
Conference
Conference | The Fortieth International Conference on Machine Learning |
---|---|
Abbreviated title | ICML 2023 |
Country/Territory | United States |
City | Honolulu |
Period | 23/07/23 → 29/07/23 |
Internet address |