BERKELEY
— Imagine tapping into the mind of a
coma patient, or watching one’s own dream on YouTube. With a cutting-edge blend
of brain imaging and computer simulation, scientists at the University of
California, Berkeley, are bringing these futuristic scenarios within reach.
Using
functional Magnetic Resonance Imaging (fMRI) and computational models, UC
Berkeley researchers have succeeded in decoding and reconstructing people’s
dynamic visual experiences – in this case, watching Hollywood movie trailers.As
yet, the technology can only reconstruct movie clips people have already
viewed. However, the breakthrough paves the way for reproducing the movies
inside our heads that no one else sees, such as dreams and memories, according
to researchers. The approximate reconstruction (right) of a movie clip
(left) is achieved through brain imaging and computer simulation
“This
is a major leap toward reconstructing internal imagery,” said Professor Jack
Gallant, a UC Berkeley neuroscientist and coauthor of the study published
online in the journal Current Biology. “We are opening a window into the
movies in our minds.”Eventually, practical applications of the technology could
include a better understanding of what goes on in the minds of people who
cannot communicate verbally, such as stroke victims, coma patients and people with
neurodegenerative diseases.It may also lay the groundwork for brain-machine
interface so that people with cerebral palsy or paralysis, for example, can
guide computers with their minds.
However,
researchers point out that the technology is decades from allowing users to
read others’ thoughts and intentions, as portrayed in such sci-fi classics as
“Brainstorm,” in which scientists recorded a person’s sensations so that others
could experience them.
Previously,
Gallant and fellow researchers recorded brain activity in the visual cortex
while a subject viewed black-and-white photographs. They then built a
computational model that enabled them to predict with overwhelming accuracy
which picture the subject was looking at.In their latest experiment, researchers
say they have solved a much more difficult problem by actually decoding brain
signals generated by moving pictures.
“Our
natural visual experience is like watching a movie,” said Shinji Nishimoto,
lead author of the study and a post-doctoral researcher in Gallant’s lab. “In
order for this technology to have wide applicability, we must understand how
the brain processes these dynamic visual experiences.”