Toward a Physics of Deep Learning and Brains

AI in healthcare
Published: arXiv: 2509.22649v1
Authors

Arsham Ghavasieh Meritxell Vila-Minana Akanksha Khurd John Beggs Gerardo Ortiz Santo Fortunato

Abstract

Deep neural networks and brains both learn and share superficial similarities: processing nodes are likened to neurons and adjustable weights are likened to modifiable synapses. But can a unified theoretical framework be found to underlie them both? Here we show that the equations used to describe neuronal avalanches in living brains can also be applied to cascades of activity in deep neural networks. These equations are derived from non-equilibrium statistical physics and show that deep neural networks learn best when poised between absorbing and active phases. Because these networks are strongly driven by inputs, however, they do not operate at a true critical point but within a quasi-critical regime -- one that still approximately satisfies crackling noise scaling relations. By training networks with different initializations, we show that maximal susceptibility is a more reliable predictor of learning than proximity to the critical point itself. This provides a blueprint for engineering improved network performance. Finally, using finite-size scaling we identify distinct universality classes, including Barkhausen noise and directed percolation. This theoretical framework demonstrates that universal features are shared by both biological and artificial neural networks.

Paper Summary

Problem
The main problem addressed in this research paper is to find a unified theoretical framework that underlies both deep learning and brain function. The authors aim to show that the equations used to describe neuronal avalanches in living brains can also be applied to cascades of activity in deep neural networks.
Key Innovation
The key innovation of this work is the application of crackling noise theory, which is typically used to describe brain function, to deep neural networks. The authors demonstrate that deep networks can operate near criticality, which is a state where the system is highly sensitive to small changes. This criticality can predict the performance of the network and is supported by quasi-critical plateaus, rather than exact point criticality.
Practical Impact
The practical impact of this research is significant, as it provides a shared physics between deep learning and brains. This shared physics can offer mechanistic insight and a design playbook for building and steering future generation models. The authors suggest that deep neural networks and brains both use avalanches or cascades of activity to transmit information through stages or layers of processing units, and that operating near the critical point best satisfies the requirement for preserving information.
Analogy / Intuitive Explanation
Imagine a river flowing through a narrow canyon. If the river is flowing too slowly, it will be stuck in a rut and unable to move. If it's flowing too quickly, it will be turbulent and unable to transmit information effectively. But if the river is flowing at just the right speed, it will be in a state of criticality, where it's highly sensitive to small changes in the environment. This is similar to the state of criticality that deep neural networks and brains operate in, where they're highly sensitive to small changes in inputs and are able to transmit information effectively.
Paper Information
Categories:
cond-mat.dis-nn cond-mat.stat-mech cs.AI nlin.AO physics.bio-ph
Published Date:

arXiv ID:

2509.22649v1

Quick Actions