This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

Research team breaks down musical instincts with AI

Research team breaks down musical instincts with AI
Distinct representation of music in deep neural networks trained for natural sound detection without music. a Example log-Mel spectrograms of the natural sound data in the AudioSet . b Architecture of the deep neural network used to detect the natural sound categories in the input data. The purple box indicates the average pooling layer. c Performance (mean average precision, mAP) of the network trained without music for music-related categories (top, red bars) and other categories (bottom, blue). d Density plot of the t-SNE embedding of feature vectors obtained from the network in C. The lines represent iso-proportion lines at 80%, 60%, 40%, and 20% levels. Credit: Nature Communications (2024). DOI: 10.1038/s41467-023-44516-0

Music, often referred to as the universal language, is known to be a common component in all cultures. Could "musical instinct" be something that is shared to some degree, despite the extensive environmental differences among cultures?

A team of KAIST researchers led by Professor Hawoong Jung from the Department of Physics have used an artificial neural network model to identify the principle by which musical instincts emerge from the human brain without special learning.

The research, conducted by first author Dr. Gwangsu Kim of the KAIST Department of Physics (current affiliation: MIT Department of Brain and Cognitive Sciences) and Dr. Dong-Kyum Kim (current affiliation: IBS) is published in Nature Communications under the title "Spontaneous emergence of rudimentary music detectors in deep neural networks."

Previously, researchers have attempted to identify the similarities and differences among the music that exists in various cultures, and have tried to understand the origin of the universality. A paper published in Science in 2019 revealed that music is produced in all ethnographically distinct cultures, and that similar forms of beats and tunes are used. Neuroscientists have also learned that a specific part of the , the , is responsible for processing musical information.

Professor Jung's team used an artificial neural network model to show that cognitive functions for music forms spontaneously as a result of processing auditory information received from nature, without being taught music. The research team utilized AudioSet, a large-scale collection of sound data provided by Google, and taught the artificial neural network to learn the various sounds.

Interestingly, the research team discovered that certain neurons within the network model would respond selectively to music. In other words, they observed the spontaneous generation of neurons that reacted minimally to various other sounds like those of animals, nature, or machines, but showed high levels of response to various forms of music, both instrumental and vocal.

The neurons in the artificial neural model showed similar reactive behaviors to those in the auditory cortex of a real brain. For example, artificial neurons responded less to the sound of music that was cropped into short intervals and were rearranged. This indicates that the spontaneously-generated music-selective neurons encode the temporal structure of music. This property was not limited to a specific genre of music, but emerged across 25 different genres including classic, pop, rock, jazz, and electronic.

Furthermore, suppressing the activity of the music-selective neurons was found to greatly impede the cognitive accuracy for other natural sounds. That is to say, the neural function that processes musical information helps process other sounds, and that "" may be an instinct formed as a result of an evolutionary adaptation acquired to better process sounds from nature.

Professor Jung, who advised the research, said, "The results of our study imply that evolutionary pressure has contributed to forming the universal basis for processing musical information in various cultures." As for the significance of the research, he explained, "We look forward for this artificially built model with human-like musicality to become an original model for various applications including AI music generation, musical therapy, and for research in musical cognition."

He also commented on its limitations, adding, "This research however does not take into consideration the developmental process that follows the learning of , and it must be noted that this is a study on the foundation of processing musical information in early development."

More information: Gwangsu Kim et al, Spontaneous emergence of rudimentary music detectors in deep neural networks, Nature Communications (2024). DOI: 10.1038/s41467-023-44516-0

Journal information: Nature Communications , Science
Citation: Research team breaks down musical instincts with AI (2024, January 23) retrieved 22 April 2024 from https://medicalxpress.com/news/2024-01-team-musical-instincts-ai.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Study finds brain regions respond to familiar and unfamiliar music, even without musical training

56 shares

Feedback to editors