A new model of memory developed by Salk Institute scientists explains how neurons retain select memories a few hours after an event.
A new model of memory developed by Salk Institute scientists explains how neurons retain select memories a few hours after an event. This new framework provides a more complete picture of how memory works, which can inform research into disorders liked Parkinson's, Alzheimer's, post-traumatic stress and learning disabilities.
"Previous models of memory were based on fast activity patterns," says Terry Sejnowski, holder of Salk's Francis Crick Chair and a Howard Hughes Medical Institute Investigator. "Our new model of memory makes it possible to integrate experiences over hours rather than moments."
Over the past few decades, neuroscientists have revealed much about how long-term memories are stored. For significant events—for example, being bit by a dog—a number of proteins are quickly made in activated brain cells to create the new memories. Some of these proteins linger for a few hours at specific places on specific neurons before breaking down.
This series of biochemical events allow us to remember important details about that event—such as, in the case of the dog bite, which dog, where it was located, if it required an emergency room visit, and so on.
One problem scientists have had with modeling memory storage is explaining why only selective details and not everything in that 1-2 hour window is strongly remembered. By incorporating data from previous literature, Sejnowski and first author Cian O'Donnell, a Salk postdoctoral researcher, developed a model that bridges findings from both molecular and systems observations of memory to explain how this 1-2 hour memory window works. The work is detailed in the latest issue of Neuron.
Using computational modeling, O'Donnell and Sejnowski show that, despite the proteins being available to a number of neurons in a given circuit, memories are retained when subsequent events activate the same neurons as the original event. The scientists found that the spatial positioning of proteins at both specific neurons and at specific areas around these neurons predicts which memories are recorded. This spatial patterning framework successfully predicts memory retention as a mathematical function of time and location overlap.
Advertisement
The new model also provides a potential framework for understanding how generalizations from memories are processed during dreams.
Advertisement
"During sleep there's a reorganizing of memory—you strengthen some memories and lose ones you don't need anymore," says O'Donnell. "In addition, people learn abstractions as they sleep, but there was no idea how generalization processes happen at a neural level."
By applying their theoretical findings on overlap activity within the 1-2 hour window, they came up with a theoretical model for how the memory abstraction process might work during sleep.
Source-Eurekalert