This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

All in the mind: Decoding brain waves to identify the music we are hearing

All in the mind – decoding brainwaves to identify the music we are listening to
Dr. Ian Daly injecting conductive gel into an EEG cap. Credit: University of Essex

A new technique for monitoring brain waves can identify the music someone is hearing.

Researchers at the University of Essex hope the project could lead to helping people with severe communication disabilities such as locked-in syndrome or stroke sufferers by decoding language signals within their brains through non-invasive techniques.

Dr. Ian Daly from Essex's School of Computer Science and Electronic Engineering, who led the research, said, "This method has many potential applications. We have shown we can decode music, which suggests that we may one day be able to decode language from the ."

Essex scientists wanted to find a less invasive way of decoding acoustic information from signals in the brain to identify and reconstruct a piece of music someone was listening to.

While there have been successful previous studies monitoring and reconstructing acoustic information from , many have used more invasive methods such as electrocortiography (ECoG), which involves placing electrodes inside the skull to monitor the actual surface of the brain.

The research, published in the journal Scientific Reports, used a combination of two non-invasive methods—fMRI, which measures through the entire brain, and electroencephalogram (EEG), which measures what is happening in the brain in real time—to monitor a person's brain activity while they are listening to a piece of music. Using a deep learning neural network model, the data was translated to reconstruct and identify the piece of music.

Music is a complex acoustic signal, sharing many similarities with , so the model could potentially be adapted to translate speech. The eventual goal of this strand of research would be to translate thought, which could offer an important aid in the future for people who struggle to communicate, such as those with locked-in syndrome.

Dr. Daly added, "One application is brain-computer interfacing (BCI), which provides a directly between the brain and a computer. Obviously, this is a long way off but eventually we hope that if we can successfully decode language, we can use this to build communication aids, which is another important step towards the ultimate aim of BCI research and could one day provide a lifeline for people with severe communication disabilities."

The research involved reusing fMRI and EEG data collected originally as part of a project at the University of Reading. That project involved participants listening to a series of 40-second pieces of simple piano music from a set of 36 pieces differing in tempo, pitch, harmony and rhythm. Using these combined data sets, the model was able to accurately identify the piece of with a success rate of 71.8%.

More information: Ian Daly, Neural decoding of music from the EEG, Scientific Reports (2023). DOI: 10.1038/s41598-022-27361-x

Journal information: Scientific Reports
Citation: All in the mind: Decoding brain waves to identify the music we are hearing (2023, January 19) retrieved 23 April 2024 from https://medicalxpress.com/news/2023-01-mind-decoding-brain-music.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

System identifies music selections via brain scanning

50 shares

Feedback to editors