An Information Theory Inspired Study of Memristor Devices and their Potential Use in Neuromorphic Circuits

Lead Research Organisation: Imperial College London
Department Name: Electrical and Electronic Engineering

Abstract

This proposed project is aligned with the EPSRC Engineering theme and is best categorised under the Artificial Intelligence Technologies Research Area. It aims to explore, from a communications and information-theoretic perspective, the potential use of memristor devices as synaptic circuit elements in neuromorphic circuits, inspired by Friston's Free Energy Principle to explain the function of the brain [1]. I would like to explore the potential to use devices called memristors (a class of devices whose resistance can be modulated using an applied voltage. They were formalised as a new two-terminal circuit element by Leon Chua in 1971 [2], to complete the set of 4 ideal passive circuit elements) to practically implement the processes described by the theory, through simulation and, if possible, through practical demonstration. Some forms of memristor are good candidates for mimicking the function of binary activation units with a nonlinear thresholding (such as neurons in the human brain) for use in novel neural architectures.

I have previously explored the potential use of memristors as storage devices, modelling them as communication channels using a Generative Adversarial Network (GAN), and using an Autoencoder architecture to compress and transmit data over the noisy memristor channel: a technique known as Deep Joint Source-Channel Coding. Such techniques from deep learning can be extended throughout the course of the project in order to model the devices and their non-idealities, such as imperfect values after programming or resistance drift over time.

The idea is to create a neuromorphic architecture that uses control and optimisation rules to learn. These rules will come from the dynamics of a physical system rather than from programmed rules and will be implemented through negative feedback of an error function in an electronic circuit. This method of optimisation requires no explicit gradient computation, in contrast to the explicit computation of the gradient of the parameters with respect to a loss function, as the overwhelming majority of current neural network architectures in the field of Machine Learning and Deep Learning perform through the algorithm of back propagation.

Current memristive neural networks have attempted to translate the algorithm of back propagation into hardware - using Ohm's law for multiplication and Kirchhoff's current law for addition. They do however demonstrate another advantage: a reduction in power consumption and increase in speed. Many machine learning algorithms run on GPU hardware, using an external memory. For memristive neurons, processing and storage (of weights of the network) are separate from one another. Processing and memory are associated due to a fundamental shift in architecture: memristors have their own storage in the form of the resistance value that they retain after they have been controlled by a voltage or a current. This means that processing and storage are both done in a so called "in-situ" fashion - both in the same location. This is a move from the traditional Von-Neumann (separate storage and processor) architecture of computers towards architectures that function more similarly to the biological, spiking networks found in the human brain. This reduces power and time expended in transferring data between a processing unit and a storage device.
[1] Friston, K. (2010). The free-energy principle: A unified brain theory? Nature Reviews Neuroscience, Vol. 11, pp. 127-138.
[2] L. Chua, "Memristor-The missing circuit element," IEEE Trans. Circuit Theory, vol. 18, no. 5, pp. 507-519, 1971.

Publications

10 25 50
 
Description The potential for semantic information storage on emerging neuromorphic devices has been studied and potential bottlenecks in the storage and recovery of information on the devices have been identified. Models of the devices have been studied and an understanding of the dynamics common to multiple memristive devices has been developed, which are potentially useful for future simulations of the devices in neuromorphic circuits or in other domains.
Exploitation Route Cost and energy-efficient information storage on memristive devices is one potential application. The models developed for use in simulating the devices and a deeper understanding of the most important features of memristors could be very useful for future research into the use of memristive devices for neuromorphic applications, for example.
Sectors Digital/Communication/Information Technologies (including Software)

 
Description Other researchers are interested in the memristor models developed. Through presentation to and discussion with other researchers, the work done during this project has contributed to the nucleation of research into memristive devices as neuromorphic components. For example, the identification of resistive drift in memristive devices as a potential benefit, rather than a disadvantage, for the simulation of the long and short-term memory of biological neurons
Sector Digital/Communication/Information Technologies (including Software)
Impact Types Societal

 
Title Neural Network for Simulating Memristive Device Data 
Description A generative neural network used to simulate memristive device resistance drift for use in applications where a differentiable model is necessary, including deep learning applications. We hope to make the model available publicly soon, once a paper describing it has been submitted and/or published. 
Type Of Technology Software 
Year Produced 2023 
Impact This model is used for further research into information storage on the devices using deep learning.