This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

Researchers outline promises, challenges of understanding AI for biological discovery

Researchers outline promises, challenges of understanding AI for biological discovery
An overview of three common pitfalls of IML interpretation in biological contexts and how to avoid these pitfalls. Credit: Nature Methods (2024). DOI: 10.1038/s41592-024-02359-7, https://www.nature.com/articles/s41592-024-02359-7

Machine learning is a powerful tool in computational biology, enabling the analysis of a wide range of biomedical data such as genomic sequences and biological imaging. But when researchers use machine learning in computational biology, understanding model behavior remains crucial for uncovering the underlying biological mechanisms in health and disease.

In a recent article in Nature Methods, researchers at Carnegie Mellon University's School of Computer Science propose guidelines that outline pitfalls and opportunities for using interpretable machine learning methods to tackle problems. The Perspectives article, "Applying Interpretable Machine Learning in Computational Biology—Pitfalls, Recommendations and Opportunities for New Developments," is featured in the journal's August special issue on AI.

"Interpretable machine learning has generated significant excitement as machine learning and artificial intelligence tools are being applied to increasingly important problems," said Ameet Talwalkar, an associate professor in CMU's Machine Learning Department (MLD).

"As these models grow in complexity, there is great promise not only in developing highly but also in creating tools that help end users understand how and why these models make certain predictions. However, it is crucial to acknowledge that interpretable machine learning has yet to deliver turnkey solutions to this interpretability problem."

The paper is a between doctoral students Valerie Chen in MLD and Muyu (Wendy) Yang in the Ray and Stephanie Lane Computational Biology Department. Chen's earlier work critiquing the interpretable machine learning community's lack of grounding in downstream use cases inspired the article, and the idea was developed through discussions with Yang and Jian Ma, the Ray and Stephanie Lane Professor of Computational Biology.

"Our collaboration began with a deep dive into computational biology papers to survey the application of interpretable machine learning methods," Yang said. "We noticed that many applications used these methods in a somewhat ad hoc manner. Our goal with this paper was to provide guidelines for more robust and consistent use of interpretable machine learning methods in computational ."

One major pitfall the paper addresses is the reliance on a single interpretable machine learning method. Instead, the researchers recommend using multiple interpretable machine learning methods with diverse sets of hyperparameters and comparing their results to obtain a more comprehensive understanding of the model behavior and its underlying interpretations.

"While some machine learning models seem to work surprisingly well, we often do not fully understand why," Ma said. "In scientific domains like biomedicine, understanding why models work is crucial for discovering fundamental biological mechanisms."

The paper also warns against cherry-picking results when evaluating interpretable machine learning methods, as this can lead to incomplete or biased interpretations of scientific findings.

Chen emphasized that the guidelines may have broader implications for a wider audience of researchers interested in applying interpretable machine-learning methods to their work.

"We hope that machine learning researchers developing new interpretable machine learning methods and tools—particularly those working on explaining large language models—will carefully consider the human-centric aspects of interpretable machine learning," Chen said. "This includes understanding who their target user is and how the method will be used and evaluated."

While understanding model behavior remains crucially important for and a fundamentally unsolved problem, the authors hope these challenges spur further interdisciplinary collaborations to facilitate the broader use of AI for scientific impact.

More information: Chen, V. et al. Applying Interpretable Machine Learning in Computational Biology — Pitfalls, Recommendations and Opportunities for New Developments, Nature Methods (2024). DOI: 10.1038/s41592-024-02359-7, www.nature.com/articles/s41592-024-02359-7

Journal information: Nature Methods
Citation: Researchers outline promises, challenges of understanding AI for biological discovery (2024, August 9) retrieved 9 August 2024 from https://medicalxpress.com/news/2024-08-outline-ai-biological-discovery.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Scientists could discover physical laws faster using new machine learning technique

36 shares

Feedback to editors