Neural network model shows why people with autism read facial expressions differently

Neural Network Model Shows Why People with Autism Read Facial Expressions Differently
A Tohoku University Hospital pediatric counseling room. Credit: Yuta Takahashi, et al

People with autism spectrum disorder have difficulty interpreting facial expressions.

Using a that reproduces the brain on a computer, a group of researchers based at Tohoku University have unraveled how this comes to be.

The journal Scientific Reports published the results on July 26, 2021.

"Humans recognize different emotions, such as sadness and anger by looking at . Yet little is known about how we come to recognize different emotions based on the visual information of facial expressions," said paper coauthor, Yuta Takahashi.

"It is also not clear what changes occur in this process that leads to people with disorder struggling to read facial expressions."

The research group employed predictive processing theory to help understand more. According to this theory, the brain constantly predicts the next sensory stimulus and adapts when its prediction is wrong. Sensory information, such as facial expressions, helps reduce prediction error.

The artificial neural network incorporated the predictive processing theory and reproduced the developmental process by learning to predict how parts of the face would move in videos of facial expression. After this, the clusters of emotions were self-organized into the neural network model's higher level neuron space—without the model knowing which emotion the facial expression in the video corresponds to.

Neural Network Model Shows Why People with Autism Read Facial Expressions Differently
A Conceptual diagram of the current study. Credit: Yuta Takahashi, et al

The model could generalize unknown facial expressions not given in the training, reproducing facial part movements and minimizing prediction errors.

Following this, the researchers conducted experiments and induced abnormalities in the neurons' activities to investigate the effects on learning development and cognitive characteristics. In the model where heterogeneity of activity in neural population was reduced, the generalization ability also decreased; thus, the formation of clusters in higher-level neurons was inhibited. This led to a tendency to fail in identifying the emotion of unknown facial expressions, a similar symptom of autism spectrum disorder.

According to Takahashi, the study clarified that predictive processing theory can explain emotion recognition from facial expressions using a neural network model.

"We hope to further our understanding of the process by which humans learn to recognize emotions and the cognitive characteristics of people with ," added Takahashi. "The study will help advance developing appropriate intervention methods for people who find it difficult to identify emotions."


Explore further

Autistic people find it harder to identify anger in facial expressions, new study shows

More information: Yuta Takahashi et al, Neural network modeling of altered facial expression recognition in autism spectrum disorders based on predictive processing framework, Scientific Reports (2021). DOI: 10.1038/s41598-021-94067-x
Journal information: Scientific Reports

Provided by Tohoku University
Citation: Neural network model shows why people with autism read facial expressions differently (2021, August 5) retrieved 20 September 2021 from https://medicalxpress.com/news/2021-08-neural-network-people-autism-facial.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
683 shares

Feedback to editors