Quantcast

Scans Show It’s Not Only Sight That Helps Us Get Our Bearings

May 26, 2011

Our brain’s understanding of spatial awareness is not triggered by sight alone, scientists have found, in a development that could help design technology for the visually impaired

Our brain’s understanding of spatial awareness is not triggered by sight alone, scientists have found, in a development that could help design technology for the visually impaired.

Researchers at the University of Edinburgh have found that our brain can use other senses ““ such as touch ““ to help us understand spatial awareness.

Scientists took MRI brain scans of both sighted volunteers and others who had been blind since birth while they examined three-dimensional spaces.

Both groups were first asked to feel three dimensional Lego models representing a geometric layout of a room and models of abstract objects containing no enclosed spaces. The sighted volunteers were then also asked to look at photographs of the same rooms and objects.

The scans showed that activity in the part of the brain that computes the spatial layout of a scene ““ known as the parahippocampal place area ““ was doubled for the sighted volunteers when looking at images of a room layout compared with when they looked at images of abstract objects.

The research reinforces previous findings linking the parahippocampal place area to our understanding of spatial awareness.

Crucially, this brain activity was also much stronger for rooms compared with objects when the sighted volunteers touched the models without being able to see them. Given that the non-sighted participants showed the same results, these findings cannot be explained by visual imagery but instead demonstrate that the parahippocampal place area receives spatial information from multiple senses.

On the Net:




comments powered by Disqus