Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks

Recently, a Cambrian explosion of a novel, non-volatile memory (NVM) devices known as memristive devices have inspired effort in building hardware neural networks that learn like the brain. Early experimental prototypes built simple perceptrons from nanosynapses, and recently, fully-connected multi-...

Full description

Bibliographic Details
Main Authors: Christopher H. Bennett, Vivek Parmar, Laurie E. Calvet, Jacques-Olivier Klein, Manan Suri, Matthew J. Marinella, Damien Querlioz
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8726293/
id doaj-47533fa00ed34d889501fb275d953415
record_format Article
spelling doaj-47533fa00ed34d889501fb275d9534152021-03-29T23:44:53ZengIEEEIEEE Access2169-35362019-01-017739387395310.1109/ACCESS.2019.29200768726293Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural NetworksChristopher H. Bennett0https://orcid.org/0000-0002-6989-292XVivek Parmar1https://orcid.org/0000-0001-7380-0816Laurie E. Calvet2Jacques-Olivier Klein3Manan Suri4https://orcid.org/0000-0003-1417-3570Matthew J. Marinella5Damien Querlioz6Centre de Nanosciences et de Nanotechnologies, University of Paris-Sud, Universite Paris-Saclay, Palaiseau, FranceDepartment of Electrical Engineering, IIT Delhi, New Delhi, IndiaCentre de Nanosciences et de Nanotechnologies, University of Paris-Sud, Universite Paris-Saclay, Palaiseau, FranceCentre de Nanosciences et de Nanotechnologies, University of Paris-Sud, Universite Paris-Saclay, Palaiseau, FranceDepartment of Electrical Engineering, IIT Delhi, New Delhi, IndiaSandia National Laboratories, Albuquerque, NM, USACentre de Nanosciences et de Nanotechnologies, University of Paris-Sud, Universite Paris-Saclay, Palaiseau, FranceRecently, a Cambrian explosion of a novel, non-volatile memory (NVM) devices known as memristive devices have inspired effort in building hardware neural networks that learn like the brain. Early experimental prototypes built simple perceptrons from nanosynapses, and recently, fully-connected multi-layer perceptron (MLP) learning systems have been realized. However, while backpropagating learning systems pair well with high-precision computer memories and achieve state-of-the-art performances, this typically comes with a massive energy budget. For future Internet of Things/peripheral use cases, system energy footprint will be a major constraint, and emerging NVM devices may fill the gap by sacrificing high bit precision for lower energy. In this paper, we contrast the well-known MLP approach with the extreme learning machine (ELM) or NoProp approach, which uses a large layer of random weights to improve the separability of high-dimensional tasks, and is usually considered inferior in a software context. However, we find that when taking the device non-linearity into account, NoProp manages to equal hardware MLP system in terms of accuracy. While also using a sign-based adaptation of the delta rule for energy-savings, we find that NoProp can learn effectively with four to six 'bits' of device analog capacity, while MLP requires eight-bit capacity with the same rule. This may allow the requirements for memristive devices to be relaxed in the context of online learning. By comparing the energy footprint of these systems for several candidate nanosynapses and realistic peripherals, we confirm that memristive NoProp systems save energy compared with MLP systems. Lastly, we show that ELM/NoProp systems can achieve better generalization abilities than nanosynaptic MLP systems when paired with pre-processing layers (which do not require backpropagated error). Collectively, these advantages make such systems worthy of consideration in future accelerators or embedded hardware.https://ieeexplore.ieee.org/document/8726293/Hardware neural networksmemristive devicesonline learningedge computing
collection DOAJ
language English
format Article
sources DOAJ
author Christopher H. Bennett
Vivek Parmar
Laurie E. Calvet
Jacques-Olivier Klein
Manan Suri
Matthew J. Marinella
Damien Querlioz
spellingShingle Christopher H. Bennett
Vivek Parmar
Laurie E. Calvet
Jacques-Olivier Klein
Manan Suri
Matthew J. Marinella
Damien Querlioz
Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks
IEEE Access
Hardware neural networks
memristive devices
online learning
edge computing
author_facet Christopher H. Bennett
Vivek Parmar
Laurie E. Calvet
Jacques-Olivier Klein
Manan Suri
Matthew J. Marinella
Damien Querlioz
author_sort Christopher H. Bennett
title Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks
title_short Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks
title_full Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks
title_fullStr Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks
title_full_unstemmed Contrasting Advantages of Learning With Random Weights and Backpropagation in Non-Volatile Memory Neural Networks
title_sort contrasting advantages of learning with random weights and backpropagation in non-volatile memory neural networks
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description Recently, a Cambrian explosion of a novel, non-volatile memory (NVM) devices known as memristive devices have inspired effort in building hardware neural networks that learn like the brain. Early experimental prototypes built simple perceptrons from nanosynapses, and recently, fully-connected multi-layer perceptron (MLP) learning systems have been realized. However, while backpropagating learning systems pair well with high-precision computer memories and achieve state-of-the-art performances, this typically comes with a massive energy budget. For future Internet of Things/peripheral use cases, system energy footprint will be a major constraint, and emerging NVM devices may fill the gap by sacrificing high bit precision for lower energy. In this paper, we contrast the well-known MLP approach with the extreme learning machine (ELM) or NoProp approach, which uses a large layer of random weights to improve the separability of high-dimensional tasks, and is usually considered inferior in a software context. However, we find that when taking the device non-linearity into account, NoProp manages to equal hardware MLP system in terms of accuracy. While also using a sign-based adaptation of the delta rule for energy-savings, we find that NoProp can learn effectively with four to six 'bits' of device analog capacity, while MLP requires eight-bit capacity with the same rule. This may allow the requirements for memristive devices to be relaxed in the context of online learning. By comparing the energy footprint of these systems for several candidate nanosynapses and realistic peripherals, we confirm that memristive NoProp systems save energy compared with MLP systems. Lastly, we show that ELM/NoProp systems can achieve better generalization abilities than nanosynaptic MLP systems when paired with pre-processing layers (which do not require backpropagated error). Collectively, these advantages make such systems worthy of consideration in future accelerators or embedded hardware.
topic Hardware neural networks
memristive devices
online learning
edge computing
url https://ieeexplore.ieee.org/document/8726293/
work_keys_str_mv AT christopherhbennett contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
AT vivekparmar contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
AT laurieecalvet contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
AT jacquesolivierklein contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
AT manansuri contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
AT matthewjmarinella contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
AT damienquerlioz contrastingadvantagesoflearningwithrandomweightsandbackpropagationinnonvolatilememoryneuralnetworks
_version_ 1724188971934679040