This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


peer-reviewed publication

trusted source


Machine learning combined with multispectral infrared imaging to guide cancer surgery

Machine learning combines with multispectral infrared imaging to guide cancer surgery
Scientists employ multispectral emission profiles instead of the conventional fluorescence intensity profile to train machine learning models for accurately identifying tumor boundaries. Credit: Waterhouse et al, DOI: 10.1117/1.JBO.28.9.094804

Surgical tumor removal remains one of the most common procedures during cancer treatment, with about 45% of cancer patients undergoing this surgery at some point. Thanks to recent progress in imaging and biochemical technologies, surgeons are now better able to tell tumors apart from healthy tissue. Specifically, this is enabled by a technique called "fluorescence-guided surgery" (FGS).

In FGS, the patient's tissue is stained with a dye that emits when irradiated with a special light source. The dye preferentially binds to the surface of tumor cells, so that its light-wave emissions provide information on the location and extent of the tumor. In most FGS-based approaches, the absolute intensity of the infrared emissions is used as the main criterion for discerning the pixels corresponding to tumors. However, it turns out that the intensity is sensitive to , the camera setup, the amount of dye used, and the time elapsed after staining. As a result, the intensity-based classification is prone to erroneous interpretation.

But what if we could instead use an intensity-independent approach to classify healthy and tumor cells? A recent study published in the Journal of Biomedical Optics and led by Dale J. Waterhouse from University College London, U.K., has now proposed such an approach. The research team has developed a new technique that combines machine learning with short-wave infrared (SWIR) fluorescence imaging to detect precise boundaries of tumors.

Their method relies on capturing multispectral SWIR images of the dyed tissue rather than simply measuring the total intensity over one particular wavelength. Put simply, the team sequentially placed six different wavelength frequency (color) filters in front of their SWIR optical system and registered six measurements for each pixel. This allowed the researchers to create the spectral profiles for each type of pixel (background, healthy, or tumor). Next, they trained seven machine learning models to identify these profiles accurately in multispectral SWIR images.

The researchers trained and validated the models in vivo, using SWIR images with a lab model for an aggressive type of neuroblastoma. They also compared different normalization approaches aimed at making the classification of pixels independent of the absolute intensity such that it was governed by the pixel's spectral profile only.

Out of the seven tested models, the best performing model achieved a remarkable per-pixel classification accuracy of 97.5% (the accuracies for , healthy, and background pixels were 97.1%, 93.5%, and 99.2%, respectively). Moreover, thanks to the normalization of the spectral profiles, the results of the model were far more robust against changes in imaging conditions. This is a particularly desirable feature for clinical applications since the ideal conditions under which new imaging technologies are usually tested are not representative of the real-world clinical environment.

Based on their findings, the team has high hopes for the proposed methodology. They anticipate that a on its implementation in could help revolutionize the field of FGS. Additionally, multispectral FGS could be extended beyond the scope of the present study. For example, it could be used to remove surgical or background lights from images, remove unwanted reflections, and provide noninvasive ways for measuring lipid content and oxygen saturation. Moreover, multispectral systems enable the use of multiple fluorescent dyes with different emission characteristics simultaneously, since the signals from each dye can be untangled from the total measurements based on their spectral profile. These multiple dyes can be used to target multiple aspects of disease, providing surgeons with even greater information.

Future studies will surely unlock the full potential of multispectral FGS, opening doors to more effective surgical procedures for treating cancer and other diseases.

More information: Dale J. Waterhouse et al, Enhancing intraoperative tumor delineation with multispectral short-wave infrared fluorescence imaging and machine learning, Journal of Biomedical Optics (2023). DOI: 10.1117/1.JBO.28.9.094804

Journal information: Journal of Biomedical Optics
Provided by SPIE
Citation: Machine learning combined with multispectral infrared imaging to guide cancer surgery (2023, March 29) retrieved 26 May 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Lighting up tumors could help surgeons remove them more precisely


Feedback to editors