Say what you see in this picture out loud: "The cat ran over the car." A.I. is helping researchers unlock how your brain knows that sentence is different than: "The car ran over the cat.". Credit: University of Rochester Medical Center

Have you ever wondered why you are able to hear a sentence and understand its meaning—given that the same words in a different order would have an entirely different meaning? New research involving neuroimaging and A.I., describes the complex network within the brain that comprehends the meaning of a spoken sentence.

"It has been unclear whether the integration of this meaning is represented in a particular site in the , such as the anterior temporal lobes, or reflects a more level operation that engages multiple ," said Andrew Anderson, Ph.D., research assistant professor in the University of Rochester Del Monte Institute for Neuroscience and lead author on of the study which was published in the Journal of Neuroscience. "The meaning of a is more than the sum of its parts. Take a very simple example—'the car ran over the cat' and 'the cat ran over the car'—each sentence has exactly the same words, but those words have a totally different meaning when reordered."

The study is an example of how the application of artificial neural networks, or A.I., are enabling researchers to unlock the extremely complex signaling in the brain that underlies functions such as processing language. The researchers gather brain activity data from study participants who read sentences while undergoing fMRI. These scans showed activity in the brain spanning across a network of different regions—anterior and posterior temporal lobes, inferior parietal cortex, and inferior frontal cortex. Using the computational model InferSent—an A.I. model developed by Facebook trained to produce unified semantic representations of sentences—the researchers were able to predict patterns of fMRI activity reflecting the encoding of sentence across those brain regions.

"It's the first time that we've applied this model to predict brain activity within these regions, and that provides new evidence that contextualized semantic representations are encoded throughout a distributed language network, rather than at a single site in the brain."

Anderson and his team believe the findings could be helpful in understanding clinical conditions. "We're deploying similar methods to try to understand how language comprehension breaks down in early Alzheimer's disease. We are also interested in moving the models forward to predict brain activity elicited as language is produced. The current study had people read sentences, in the future we're interested in moving forward to predict as people might speak sentences."

More information: Andrew James Anderson et al, Deep artificial neural networks reveal a distributed cortical network encoding propositional sentence-level meaning, The Journal of Neuroscience (2021). DOI: 10.1523/JNEUROSCI.1152-20.2021

Journal information: Journal of Neuroscience