Arthur Pellegrino
Arthur Pellegrino
I am an AI Fellow at the Data Science Centre at the Ecole Normale Supérieure PSL. I previously was a research fellow at the Gatsby Unit at UCL and before that I completed my PhD at the School of Informatics at the University of Edinburgh. My research interests lie at the interface between mathematics, machine learning and computational neuroscience.
I develop mathematical tools for better data-driven modeling, which I in turn use to test fundamental hypotheses about the geometry and dynamics of computations in biological and artificial neural networks. In particular, I combine concepts from dynamical systems, differential geometry and tensors to probe for low-dimensional neural manifolds and model dynamics on them.
News
04/12/25 - I will be giving an oral presentation about our work on Riemannian geometry & feature learning at the Workshop on Symmetry and Geometry in Neural Representations at NeurIPS (2:50pm).
05/11/25 - Our paper RNNs perform task computations by dynamically warping neural representations was accepted at NeurIPS.
04/11/25 - Our paper Emergent Riemannian geometry over learning discrete computations on continuous manifolds was accepted for publication at PMLR (NeurReps).
02/11/25 - I just started as an AI Fellow at the Data Science Centre and Paris AI Research Institute at the Dept. of Computer Science of the Ecole Normale Supérieure PSL
Projects
The Riemmanian geometry of dynamical systems representations
Characterising how neurons represent variables of the world in their activation has reshaped our understanding of the computations performed by biological and artificial neural networks. In parallel, there is mounting interest in understanding how dynamical systems perform computations on task variables through their internal dynamics. Here, we use concepts of Riemannian geometry to study the internal representations of dynamical systems. In particular, we showed that RNNs perform computations by dynamically warping their representation of task variables, for example compressing irrelevant information and expending relevant ones.
Pellegrino, A., & Chadwick, A. (2025). RNNs perform task computations by dynamically warping neural representations. Advances in Neural Information Processing Systems (NeurIPS) 38.
Dimensionality reduction beyond subspaces
Recent work has argued that large-scale neural recordings are often well described by patterns of coactivation across neurons, which assumes that the data lies on a fixed-low-dimensional subspace. This view may overlook higher-dimensional structure, including stereotyped neural sequences or slowly evolving latent spaces. Here we argue that task-relevant variability in neural data can also cofluctuate over trials or time, defining distinct ‘covariability classes’ that may co-occur within the same dataset. To demix these covariability classes, we develop sliceTCA (slice tensor component analysis), a new unsupervised dimensionality reduction method for neural data tensors.
Low-tensor-rank RNNs
Learning relies on coordinated synaptic changes in neural connectivity. Yet, previous methods to infer a dynamical system from data assumed the existence of a single fixed weight matrix. This discrepency made such methods fundamentally unsuited for data recorded from learning animals. To probe for such changes in connectivity and dynamics in data, we introduced low-tensor-rank RNNs which fits RNNs wiith connectivity connectivity.
Low (tensor) rank gradient dynamics
Recent work has argued that machine learning models trained using gradient descent (or variants thereof) tend to have low-rank weights. But how this low rank structure unfolds over learning is unknown. To address this, we investigate the rank of the 3-tensor formed by stacking the weight matrices throughout learning. We derived several results bounding the rank of the gradient and weight tensor of linear dynamical systems and RNNs, using adjoint dynamics.
Those results were published along ltrRNN in the appendix to:
Dynamical systems principles underlie data manifolds
The manifold hypothesis posists that low-dimensional manifolds can be ubiquitously found in high-dimensional data, and in biology such data are often generated by an underlying dynamical system. Yet, while dynamical systems models describe mechanistically the interactions between biological variables, manifolds are usually empirically extracted from data using dimensionality reduction methods. This has left elusive the link between biological geometry and dynamics. In this work, we argue that the low-dimensional manifolds seen in experimental data throughout the biological sciences are naturally born from dynamical systems principles.
Representational drift
The representation of task variables in the brain continuously changes even in the absence of learning, which is a phenomenon called "representational drift." Yet, neural network models trained under gradient descent in the absence of noise do not exhibit such an effect. Recent work has suggested that adding noise to the gradient dynamics is sufficient to introduce drift in the representation. Yet, in general such noise-driven drift can possibly change the intrinsic geometry of the representation, which is inconsistent with data. In this project, we investigate in artificial neural networks the conditions under which learning dynamics preserve intrinsic geometry of neural representation while changing the extrinsic geometry.
Pellegrino, A., & Palmigiano A. Ongoing project.
Teaching
I've TAed classes on Machine Learning and Pattern Recognition and Computational Neuroscience, at the School of Informatics in Edinburgh as well as Introduction to probability and statistics at the Math Department of the University of Colorado in Boulder. I designed teaching materials in math, machine learning and computational neuroscience. I have also designed summer course materials for CAJAL Machine Learning for Computational Neuroscience course and the Bio-inspired Deep Learning workshop.
I like to provide visual intuition about mathematical concepts
In labs, I use interactive Jupyter notebooks
In class, I teach using a combination of slides and live derivations
Supervision
Notable student thesis projects supervised:
Foundation models of large-scale neural data. Melina Laimon, David O'Neil (Masters' in Machine Learning / Statistics, UCL)
The effect of sparsity on neural representations. Gianluca Carrozo (Masters' in Machine Learning, UCL)
Adaptive low-rank gradient dynamics. Capucine Brillet (Masters' in Mathematics and Applications, ENS / PSL)
Observation models for data-driven manifold inferrence. Julian Brandon (Masters' in Artificial Intelligence, Edinburgh)
Tensor decomposition methods for RNN connectivity. Melina Muller (Masters' in Artificial Intelligence, Edinburgh)
Contact
Feel free to contact me at arthur (dot) pellegrino (at) ens.fr regarding:
Masters's or undergraduate research projects
Applications of our methods to your data
Any interesting ML/Math/Neuro results you'd like to share