Leads to new insights about how the brain processes information out of context.
From: Johns Hopkins University
July 8, 2021 -- When people see a
toothbrush, a car, a tree -- any individual object -- their brain automatically
associates it with other things it naturally occurs with, allowing humans to
build context for their surroundings and set expectations for the world.
By using machine-learning and brain
imaging, researchers measured the extent of the "co-occurrence"
phenomenon and identified the brain region involved. The findings appear
in Nature Communications.
"When we see a refrigerator, we
think we're just looking at a refrigerator, but in our mind, we're also calling
up all the other things in a kitchen that we associate with a
refrigerator," said corresponding author Mick Bonner, a Johns Hopkins
University cognitive scientist. "This is the first time anyone has
quantified this and identified the brain region where it happens."
In a two-part study, Bonner and
co-author, Russell Epstein, a psychology professor at the University of
Pennsylvania, used a database with thousands of scenic photos with every object
labeled. There were pictures of household scenes, city life, nature -- and the
pictures had labels for every mug, car, tree, etc. To quantify object
co-occurrences, or how often certain objects appeared with others, they created
a statistical model and algorithm that demonstrated the likelihood of seeing a
pen if you saw a keyboard, or seeing a boat if you saw a dishwasher.
With these contextual associations
quantified, the researchers next attempted to map the brain region that handles
the links.
While subjects were having their brain
activity monitored with functional magnetic resonance imaging, or fMRI, the
team showed them pictures of individual objects and looked for evidence of a
region whose responses tracked this co-occurrence information. The spot they
identified was a region in the visual cortex commonly associated with the
processing of spatial scenes.
"When you look at a plane, this
region signals sky and clouds and all the other things," Bonner said.
"This region of the brain long thought to process the spatial environment
is also coding information about what things go together in the world."
Researchers have long-known that people
are slower to recognize objects out of context. The team believes this is the
first large-scale experiment to quantify the associations between objects in
the visual environment as well as the first insight into how this visual
context is represented in the brain.
"We show in a fine-grained way that
the brain actually seems to represent this rich statistical information,"
Bonner said.
https://www.sciencedaily.com/releases/2021/07/210708134756.htm
No comments:
Post a Comment