Statistical physics offers a way to understand how learning systems organize information over time. It also can provide a path to interpretable machine learning by treating models as dynamical systems. Viewing training as stochastic dynamics on high-dimensional energy landscapes lets us predict typical behavior—how noise (e.g., SGD stochasticity) aids exploration, when generalization emerges, and why performance can change abruptly like a phase transition. Beyond explaining learning, my research builds classifiers as dynamical systems: classes correspond to stable attractors, trajectories explain how an input flows to a decision, and stability/bifurcations map to robustness and failure modes. In this framework, tools such as order parameters, free energies, and dynamical mean-field theory turn messy optimization into measurable quantities, while control-style interventions reshape the vector field to improve accuracy and transparency. The goal is theory-guided AI that is not only performant, but mechanistically understandable.
Learning in the Wilson-Cowan model for metapopulation. Article link.
The Wilson-Cowan model for metapopulation, a Neural Mass Network Model, treats different subcortical regions of the brain as connected nodes, with connections representing various types of structural, functional, or effective neuronal connectivity between these regions. Each region comprises interacting populations of excitatory and inhibitory cells, consistent with the standard Wilson-Cowan model. In this paper, we show how to incorporate stable attractors into such a metapopulation model’s dynamics. By doing so, we transform the Neural Mass Network Model into a biologically inspired learning algorithm capable of solving different classification tasks. We test it on MNIST and Fashion MNIST in combination with convolutional neural networks, as well as on CIFAR-10 and TF-FLOWERS, and in combination with a transformer architecture (BERT) on IMDB, consistently achieving high classification accuracy.
Daydreaming Hopfield Networks and their surprising effectiveness on correlated data. Article link.
To improve the storage capacity of the Hopfield model, we develop a version of the dreaming algorithm that perpetually reinforces the patterns to be stored (as in the Hebb rule), and erases the spurious memories (as in dreaming algorithms). For this reason, we called it Daydreaming. Daydreaming is not destructive and it converges asymptotically to stationary retrieval maps. When trained on random uncorrelated examples, the model shows optimal performance in terms of the size of the basins of attraction of stored examples and the quality of reconstruction. We also train the Daydreaming algorithm on correlated data obtained via the random-features model and argue that it spontaneously exploits the correlations thus increasing even further the storage capacity and the size of the basins of attraction. Moreover, the Daydreaming algorithm is also able to stabilize the features hidden in the data. Finally, we test Daydreaming on the MNIST dataset and show that it still works surprisingly well, producing attractors that are close to unseen examples and class prototypes.