Beyond two-point correlations: from higher-order data statistics to neural representations
beyond2 aims to develop a theory on how deep neural networks learn from high-order correlations in non-Gaussian data, enhancing understanding and practical deployment in critical applications.
Projectdetails
Introduction
Deep neural networks (DNNs) have revolutionised how we learn from data. Rather than requiring careful engineering and domain knowledge to extract features from raw data, DNNs learn the relevant features for a task automatically from data. In particular, high-order correlations (HOCs) of the data are crucial for both the performance of DNNs and the type of features they learn.
Problem Statement
However, existing theoretical frameworks cannot capture the impact of HOCs. They either study “lazy” regimes where DNNs do not learn data-specific features, or they rely on the unrealistic assumption of Gaussian inputs devoid of non-trivial HOCs.
Project Goals
beyond2 will develop a theory for how and what neural networks learn from the high-order correlations of their data. We break the problem into two parts:
-
How? We analyse the learning dynamics of neural networks trained by stochastic gradient descent to unveil the mechanism by which they learn from HOCs efficiently. This includes understanding the minimum amount of training data and learning time required to attain satisfactory predictive performance.
-
What? Our preliminary research suggests that neural filters are primarily determined by the “principal components” of HOCs. We investigate how these principal components relate to fundamental data properties, such as symmetries of the inputs.
Methodology
We attack these problems by extending methods from statistical physics and high-dimensional statistics to handle non-Gaussian input distributions. Studying the interplay between data structure and learning dynamics will allow us to understand how specific learning mechanisms, like attention or recursion, are able to unwrap HOCs.
Expected Outcomes
By shifting the focus from unstructured to non-Gaussian data models, beyond2 will yield new insights into the inner workings of neural networks. These insights will bring theory closer to practice and might facilitate the safe deployment of neural networks in high-stakes applications.
Financiële details & Tijdlijn
Financiële details
Subsidiebedrag | € 1.499.999 |
Totale projectbegroting | € 1.499.999 |
Tijdlijn
Startdatum | 1-1-2025 |
Einddatum | 31-12-2029 |
Subsidiejaar | 2025 |
Partners & Locaties
Projectpartners
- SCUOLA INTERNAZIONALE SUPERIORE DI STUDI AVANZATI DI TRIESTEpenvoerder
Land(en)
Vergelijkbare projecten binnen European Research Council
Project | Regeling | Bedrag | Jaar | Actie |
---|---|---|---|---|
Higher-Order Hodge Laplacians for Processing of multi-way SignalsThis project aims to enhance graph signal processing by developing methods for analyzing higher-order relations in complex systems using Hodge-Laplacians and algebraic topology. | ERC Starting... | € 1.500.000 | 2022 | Details |
Dynamics-Aware Theory of Deep LearningThis project aims to create a robust theoretical framework for deep learning, enhancing understanding and practical tools to improve model performance and reduce complexity in various applications. | ERC Starting... | € 1.498.410 | 2022 | Details |
Computational Hardness Of RepresentAtion LearningCHORAL aims to bridge theory and practice in neural networks by quantifying learning costs through a statistical framework, enhancing representation learning for structured data and multi-layer networks. | ERC Starting... | € 1.280.750 | 2022 | Details |
Understanding Deep LearningThe project aims to establish a solid theoretical foundation for deep learning by investigating optimization, statistical complexity, and representation, enhancing understanding and algorithm development. | ERC Starting... | € 1.499.750 | 2022 | Details |
From reconstructions of neuronal circuits to anatomically realistic artificial neural networksThis project aims to enhance artificial neural networks by extracting wiring principles from brain connectomics to improve efficiency and reduce training data needs for deep learning applications. | ERC Proof of... | € 150.000 | 2022 | Details |
Higher-Order Hodge Laplacians for Processing of multi-way Signals
This project aims to enhance graph signal processing by developing methods for analyzing higher-order relations in complex systems using Hodge-Laplacians and algebraic topology.
Dynamics-Aware Theory of Deep Learning
This project aims to create a robust theoretical framework for deep learning, enhancing understanding and practical tools to improve model performance and reduce complexity in various applications.
Computational Hardness Of RepresentAtion Learning
CHORAL aims to bridge theory and practice in neural networks by quantifying learning costs through a statistical framework, enhancing representation learning for structured data and multi-layer networks.
Understanding Deep Learning
The project aims to establish a solid theoretical foundation for deep learning by investigating optimization, statistical complexity, and representation, enhancing understanding and algorithm development.
From reconstructions of neuronal circuits to anatomically realistic artificial neural networks
This project aims to enhance artificial neural networks by extracting wiring principles from brain connectomics to improve efficiency and reduce training data needs for deep learning applications.