A neural-level model of spatial memory and imagery

We present a model of how neural representations of egocentric spatial experiences in parietal cortex interface with viewpoint-independent representations in medial temporal areas, via retrosplenial cortex, to enable many key aspects of spatial cognition. This account shows how previously reported n...

Full description

Bibliographic Details
Main Authors: Andrej Bicanski, Neil Burgess
Format: Article
Language:English
Published: eLife Sciences Publications Ltd 2018-09-01
Series:eLife
Subjects:
Online Access:https://elifesciences.org/articles/33752
id doaj-3d1db15fbc24434e89e8f29c5119470b
record_format Article
spelling doaj-3d1db15fbc24434e89e8f29c5119470b2021-05-05T16:08:04ZengeLife Sciences Publications LtdeLife2050-084X2018-09-01710.7554/eLife.33752A neural-level model of spatial memory and imageryAndrej Bicanski0https://orcid.org/0000-0003-3356-1034Neil Burgess1https://orcid.org/0000-0003-0646-6584Institute of Cognitive Neuroscience, University College London, London, United KingdomInstitute of Cognitive Neuroscience, University College London, London, United KingdomWe present a model of how neural representations of egocentric spatial experiences in parietal cortex interface with viewpoint-independent representations in medial temporal areas, via retrosplenial cortex, to enable many key aspects of spatial cognition. This account shows how previously reported neural responses (place, head-direction and grid cells, allocentric boundary- and object-vector cells, gain-field neurons) can map onto higher cognitive function in a modular way, and predicts new cell types (egocentric and head-direction-modulated boundary- and object-vector cells). The model predicts how these neural populations should interact across multiple brain regions to support spatial memory, scene construction, novelty-detection, ‘trace cells’, and mental navigation. Simulated behavior and firing rate maps are compared to experimental data, for example showing how object-vector cells allow items to be remembered within a contextual representation based on environmental boundaries, and how grid cells could update the viewpoint in imagery during planning and short-cutting by driving sequential place cell activity.https://elifesciences.org/articles/33752computational modelepisodic memoryspatially selective cellsspatial cognitionscene constructiontrace cells
collection DOAJ
language English
format Article
sources DOAJ
author Andrej Bicanski
Neil Burgess
spellingShingle Andrej Bicanski
Neil Burgess
A neural-level model of spatial memory and imagery
eLife
computational model
episodic memory
spatially selective cells
spatial cognition
scene construction
trace cells
author_facet Andrej Bicanski
Neil Burgess
author_sort Andrej Bicanski
title A neural-level model of spatial memory and imagery
title_short A neural-level model of spatial memory and imagery
title_full A neural-level model of spatial memory and imagery
title_fullStr A neural-level model of spatial memory and imagery
title_full_unstemmed A neural-level model of spatial memory and imagery
title_sort neural-level model of spatial memory and imagery
publisher eLife Sciences Publications Ltd
series eLife
issn 2050-084X
publishDate 2018-09-01
description We present a model of how neural representations of egocentric spatial experiences in parietal cortex interface with viewpoint-independent representations in medial temporal areas, via retrosplenial cortex, to enable many key aspects of spatial cognition. This account shows how previously reported neural responses (place, head-direction and grid cells, allocentric boundary- and object-vector cells, gain-field neurons) can map onto higher cognitive function in a modular way, and predicts new cell types (egocentric and head-direction-modulated boundary- and object-vector cells). The model predicts how these neural populations should interact across multiple brain regions to support spatial memory, scene construction, novelty-detection, ‘trace cells’, and mental navigation. Simulated behavior and firing rate maps are compared to experimental data, for example showing how object-vector cells allow items to be remembered within a contextual representation based on environmental boundaries, and how grid cells could update the viewpoint in imagery during planning and short-cutting by driving sequential place cell activity.
topic computational model
episodic memory
spatially selective cells
spatial cognition
scene construction
trace cells
url https://elifesciences.org/articles/33752
work_keys_str_mv AT andrejbicanski aneurallevelmodelofspatialmemoryandimagery
AT neilburgess aneurallevelmodelofspatialmemoryandimagery
AT andrejbicanski neurallevelmodelofspatialmemoryandimagery
AT neilburgess neurallevelmodelofspatialmemoryandimagery
_version_ 1721459538398281728