Abstract: I will discuss the design trend of ‘event-driven’ bioinspired or neuromorphic audio front-end circuits such as spiking silicon cochleas for edge audio intelligent devices that need to satisfy stringent power constraints of ‘always-on’ operation. These bioinspired circuits bypass the conventional combined stage of the analog-to-digital conversion block and subsequent digital filtering. The microphone input signal is first processed by continuous-time analog bandpass filters followed by a rectification stage and a spike generation stage. The output of the neuromorphic feature extractor is used to train spiking-inspired delta recurrent neural networks for audio wake-up tasks including voice activity detection and keyword spotting. I further show that our keyword ASIC that combines a bio-inspired audio front-end with this deep-neural-network classifier achieves microWatts power consumption and that the combination of the spatial sparsity and temporal sparsity in these networks increases the energy efficiency of the network implementation on an FPGA system by 42X in a speech recognition task.
Abstract: Event-based eye-tracking systems are revolutionizing fields like consumer electronics and neuroscience. Human eye movements can exceed speeds of 300°/s, necessitating high-speed event cameras for accurate tracking. In consumer electronics, particularly AR/VR, these systems reduce power consumption through sparse data streams, enabling lighter, more efficient headsets with extended usage and enhanced comfort. This advancement significantly improves the immersive AR/VR experience and the capabilities of portable technology. In neuroscience, event-based eye tracking aids in understanding visual attention and diagnosing neurological disorders. This talk will explore developing an event-based eye-tracking system for precise rapid eye movement tracking, promising lighter devices and deeper insights into neuroscience and cognitive research.
Abstract: While artificial neural networks represent a highly successful mapping from neuroscience AI, and clearly capture important aspects neuronal information processing, data from experimental neuroscience strongly suggests that the ANN abstraction omits important biological computational principles. This includes such features as the pulsed nature of neural communication, the diversity and function of neuronal morphology, and the inherently time-continuous mode of operation. To investigate these computational principles, we need to be able to train large and complex networks of spiking neurons for specific tasks. In this talk, I will show how effective online and approximate learning rules enable the supervised training of large-scale networks of detailed spiking neuronal models, and how these models can be integrated with brain-derived decision-making circuits to operate continuously. As I will argue, this approach opens up the investigation of both network and neuronal architectures based on functional principles, while at the same time demonstrating the potential power and energy efficiency of AI-solutions based on neuromorphic computing as embodied by spiking neural networks.
Abstract: A major aim in neuromorphic computing is to develop physical systems that learn. Unfortunately, the ubiquitous backpropagation (BP) algorithm is suboptimal given the need to propagate exact gradients throughout the system. In this talk, I present an alternative approach which learns through random perturbations in combination with input decorrelation. This approach is local in nature and ideally suited for implementation on noisy physical devices. We show that input decorrelation allows faster training of networks compared to BP. Furthermore, we show that learning based on random perturbations approaches the performance of BP. In our more recent work we focus on porting this approach to FPGAs and other physical devices with the aim of enabling efficient and effective on-chip learning on neuromorphic devices.
Abstract: Natural gradient descent illuminates how gradient vectors, pointing at directions of steepest descent, can be improved by considering the local curvature of loss landscapes. We extend this perspective and show that to fully solve the problem illuminated by natural gradients in neural networks, one must recognise that correlations in the data at any linear transformation, including node activations at every layer of a neural network, cause a non-orthonormal relationship between the model's parameters. To solve this requires a solution to decorrelate inputs at each individual layer of a neural network. Implementing decorrelation within multi-layer neural networks, we can show that not only is training via backpropagation sped up significantly but also existing approximations of backpropagation such, as node perturbation and feedback alignment, can once again be used to train networks. This has the potential to provide a route forward for approximate gradient descent methods which have previously been discarded, training approaches for analogue and neuromorphic hardware, and potentially insights as to the efficacy and utility of decorrelation processes in the brain.
Abstract: Resistive memory devices offer outstanding potential for near- and in-memory artificial intelligence (AI) accelerators due to their nonvolatility, high density, and multilevel capacity. However, these devices also present significant challenges due to their high variability and unpredictability, essentially rendering them functionally similar to random variables. This feature presents an intriguing parallel with machine learning, where Bayesian methodologies are expressly designed to work with random variables, delivering the added advantage of being able to quantify prediction uncertainty—a hurdle that conventional artificial intelligence techniques often struggle to overcome. In this study, we propose that these Bayesian methods could serve as an effective strategy to harness the potential of resistive memory devices, without the accompanying drawbacks. Our exploration introduces three distinct approaches that increasingly integrate the random nature of resistive memories. Each methodology is substantiated through practical implementation using circuits fabricated in a hybrid CMOS/hafnium-oxide memristor process. We first introduce Bayesian machines, which leverage near-memory computing for energy-efficient Bayesian reasoning, enabling explainable decision-making under uncertain conditions by using all available data and knowledge. These machines digitally encode parameters in resistive memory, minimizing energy costs associated with data movement. Two designs were realized using a 130-nm hybrid CMOS/HfOx memristor process, featuring stochastic computing and integer logarithmic computation for efficient processing, both demonstrating remarkable energy efficiency. Then, we show the realization of Bayesian neural networks (BNNs), which treat synapses as random variables to model uncertainty in predictions, and utilize resistive memories to simulate the random behavior of synapses. This approach, validated through the programming of 50 in-memory neural networks for arrhythmia detection, allows the networks to quantify prediction certainty by leveraging the inherent randomness of memristors. The performance of these BNNs remains stable over time, incorporating resistive memory fluctuations into their uncertainty modeling. Finally, we introduce the Metropolis-Hastings Markov Chain Monte Carlo (MCMC) technique, tailored to exploit the random nature of memristors for learning, enabling synaptic weights to be sampled in a way that mimics natural Gaussian distributions [4]. This method was applied to train experimentally a 16k array of resistive memories for cancerous tissue identification, achieving a 98% accuracy rate, comparable to traditional software-based methods, thereby illustrating the potential of memristors in advanced Bayesian learning applications.
Abstract: Cognitive systems need to elaborate increasingly amount of data while featuring low-power operation and area efficiency. Memristive technology holds great promise for the design of these systems. The potential for energy-efficient and parallel computing, combined with the ability to integrate complex neural and synaptic dynamics within a single device, provides avenues for high-performance hardware implementations. Therefore, memristive technology, if correctly combined with CMOS technology, can extend the functionality of current cognitive systems. We discuss challenges and opportunities to realise memristive neuromorphic computing by developing novel hardware architectures and learning algorithms specifically tailored to best exploit the intrinsic properties of memristive technology.
Abstract: Currently, neural-network processing in machine learning applications relies on layer synchronization, whereby neurons in a layer aggregate incoming currents from all neurons in the preceding layer, before evaluating their activation function. This is practiced even in artificial Spiking Neural Networks (SNNs), which are touted as consistent with neurobiology, in spite of processing in the brain being, in fact asynchronous. A truly asynchronous system however would allow all neurons to evaluate concurrently their threshold and emit spikes upon receiving any presynaptic current. Omitting layer synchronization is potentially beneficial, for latency and energy efficiency, but asynchronous execution of models previously trained with layer synchronization may entail a mismatch in network dynamics and performance. We present a study that documents and quantifies this problem in three datasets on our simulation environment that implements network asynchrony, and we show that models trained with layer synchronization either perform sub-optimally in absence of the synchronization, or they will fail to benefit from any energy and latency reduction, when such a mechanism is in place. We then “make ends meet” and address the problem with unlayered backprop, a novel backpropagation-based training method, for learning models suitable for asynchronous processing. We train with it models that use different neuron exe-cution scheduling strategies, and we show that although their neurons are more reactive, these models consistently exhibit lower overall spike density, reach a correct decision faster without integrating all spikes, and achieve superior accuracy. Our findings suggest that asynchronous event-based (neuromorphic) AI computing is indeed more efficient, but we need to seriously rethink how we train our SNN models, to benefit from it.