When the eyes are open, visual information flows from the retina through the optic nerve and into the brain, which assembles this raw information into objects and scenes.
Scientists have previously hypothesized that objects are distinguished in the inferior temporal (IT) cortex, which is near the end of this flow of information, also called the ventral stream.
A stop sign at an intersection appears to be larger than a parked car across the street.
Is our brain playing tricks on us?
Researchers at the George Washington University gained important insights into how the human brain processes information and allocates attention.
Their study, “Attention Scales According to Inferred Real-World Object Size,” shows people pay attention to objects based on their real-world size, rather than how they are perceived by the eye.
“Since a person can only pay attention to a limited amount of information at a time, our brain uses object size to determine how much attention to allocate to that object,” Sarah Shomstein, a professor of cognitive neuroscience at the GW Columbian College of Arts and Sciences, said. “
However, the way our eyes perceive an object can be different from its actual size, such as a car appearing large when it is close and small when it is far.
Our study has shown for the first time that the brain adjusts attention based on our knowledge of an object’s size, not how our eyes view it.”
The findings could have significant impacts on many everyday tasks, from a radiologist looking for tumors of various sizes, to an airport security agent tasked with searching for threats hidden in luggage.
This study suggests attention allocation can be trained, which could lead to improvements and efficiency in performing such tasks.
To determine whether the human brain differentially allocates attention depending on an object’s size, Dr. Shomstein and her team showed participants several images of small and large everyday objects such as a domino or billiards table, but presented them at an identical fixed size.
The team measured how long it took participants to identify probe targets that were embedded within object images.
Researchers found that responses were quicker across known smaller real-world sized objects compared to larger ones, even though the space occupied by both objects on the eye was the same.
“If objects are of identical size on your eye, but you know that one of them is smaller—such as a domino nearby versus a pool table far away—you allocate more attention to the smaller item,” Dr. Shomstein said. “Your knowledge of the object’s size overrides its physical size thereby adjusting how much attention your brain allocates to the objects.”
The researchers also showed participants images of everyday objects on a screen and asked them to rate the size of each object on a scale of one to six, one being very small and six being very large.
Findings showed a direct correlation between how participants rated the size of each object and how long it took them to respond to target stimuli within the image.
“Your own personal ratings determine how efficient you are going to be at attending to that object,” co-author Andrew J. Collegio, a GW alumnus and researcher in Dr. Shomstein’s Attention and Cognition Lab, said. “If you think the pool table is really large, then your attention is going to be less focused.”
Dr. Shomstein and Dr. Collegio hope these findings will help future researchers predict how effectively people process particular objects as they pay attention to the world around them.
More information: Andrew J. Collegio et al, Attention scales according to inferred real-world object size, Nature Human Behaviour (2018). DOI: 10.1038/s41562-018-0485-2
Provided by George Washington University