Open Access
23 July 2024 Convolutional neural network advances in demosaicing for fluorescent cancer imaging with color–near-infrared sensors
Author Affiliations +
Abstract

Significance

Single-chip imaging devices featuring vertically stacked photodiodes and pixelated spectral filters are advancing multi-dye imaging methods for cancer surgeries, though this innovation comes with a compromise in spatial resolution. To mitigate this drawback, we developed a deep convolutional neural network (CNN) aimed at demosaicing the color and near-infrared (NIR) channels, with its performance validated on both pre-clinical and clinical datasets.

Aim

We introduce an optimized deep CNN designed for demosaicing both color and NIR images obtained using a hexachromatic imaging sensor.

Approach

A residual CNN was fine-tuned and trained on a dataset of color images and subsequently assessed on a series of dual-channel, color, and NIR images to demonstrate its enhanced performance compared with traditional bilinear interpolation.

Results

Our optimized CNN for demosaicing color and NIR images achieves a reduction in the mean square error by 37% for color and 40% for NIR, respectively, and enhances the structural dissimilarity index by 37% across both imaging modalities in pre-clinical data. In clinical datasets, the network improves the mean square error by 35% in color images and 42% in NIR images while enhancing the structural dissimilarity index by 39% in both imaging modalities.

Conclusions

We showcase enhancements in image resolution for both color and NIR modalities through the use of an optimized CNN tailored for a hexachromatic image sensor. With the ongoing advancements in graphics card computational power, our approach delivers significant improvements in resolution that are feasible for real-time execution in surgical environments.

1.

Introduction

The prevalence of cancer, with one in every three individuals globally being affected, highlights the indispensable role of surgical procedures in addressing localized cancers.1,2 The success of these interventions, pivotal for enhancing patient survival rates, largely depends on the thorough removal of primary tumors and the prompt detection of any metastatic occurrences.35 Advancements in near-infrared (NIR) image-guided surgery are at the forefront of improving surgical accuracy and patient prognosis.6 This innovative approach leverages cutting-edge tumor-targeted probes,7,8 state-of-the-art imaging devices,9 and the integration of machine learning algorithms into surgical workflows,10 significantly aiding surgeons in identifying primary tumors with unprecedented precision.11,12 The recent regulatory approval of CYTALUX, a folate-targeted probe for lung and ovarian cancers, and Lumicell, a cathepsin-activated probe for breast cancer, marks a milestone in utilizing tumor-targeted probes within the NIR spectrum.1315 In addition, several promising tumor-targeted agents are on the cusp of completing phase III clinical trials, poised to soon make a significant impact on the market and further revolutionize the field of NIR image-guided surgery.16

The journey of intraoperative imaging instruments for NIR image-guided surgery spans over a decade, paralleling advancements in compact sensor technologies, enhanced quantum efficiency, and sensitivity within the NIR spectrum. This evolution mirrors the historical progression of color imaging technologies.17,18 Initially, color imaging relied on time multiplexing technology, utilizing rotating spectral filters (red, green, and blue) in front of a sensor, assuming a static scene during the acquisition of each color. The challenges of motion artifacts introduced by this technique were later mitigated by incorporating a beam splitter in the optical path, dividing the incoming light into three channels to simultaneously capture color images, thereby enabling real-time, high-resolution video capture.9,19 However, this solution increased the imaging sensor’s size, rendering it impractical for numerous applications. Temperature-dependent coregistration further complicates matters.20 The advent of Bayer filters revolutionized color imaging,21 leading to the single-chip color sensor becoming the predominant technology over the past five decades. The demand for compact color sensors, driven largely by the mobile phone industry, has fortuitously benefited the endoscopic medical field, fostering new endoscopic applications.

NIR imaging instruments have traced a similar trajectory.2225 Initial models employed a single NIR filter over a grayscale imaging device to capture NIR fluorescent images. This approach was expanded through time-multiplexing color and NIR filters, facilitating the simultaneous imaging of visible and NIR fluorescence.26 However, the same motion artifact challenges observed in color imaging were encountered, leading to the adoption of a beam splitter solution.27 This method, dividing the incoming light into two paths for color and NIR fluorescence imaging, has become prevalent among food and drug administration-approved instruments. Yet, as the focus shifts toward minimally invasive techniques28 and imaging multiple NIR fluorescent probes, this approach is showing its limitations.

Drawing inspiration from the 1975 introduction of the Bayer color filter array, our group has developed pixelated color–NIR filters optimized for NIR fluorescent imaging and seamlessly integrated with an imaging device.20,29,30 This innovation addresses coregistration issues and improves power and data efficiency. To enhance the imaging of multiple NIR fluorescent markers, we combined vertically stacked imaging technology with pixelated spectral filters, enabling the simultaneous capture of color images and three distinct observations in the NIR spectrum. This technology, inspired by the visual system of the mantis shrimp, facilitates the concurrent imaging of two NIR fluorescent probes and three-dimensional (3D) reconstruction using NIR structured illumination.

However, similar to the Bayer filter in color imaging,21 pixelated NIR–color cameras experience a reduction in spatial resolution.31,32 Fortunately, this challenge is not new to color imaging, and various demosaicing techniques have been developed to mitigate spatial resolution loss.33 Color demosaicing techniques, essential for converting raw sensor data into full-color images, can be broadly classified into three main approaches: interpolation, dictionary-based, and learning-based. Interpolation methods, known for their computational simplicity, leverage the local attributes of an image and the correlation between color channels to estimate missing pixels.34 However, this approach often results in zipper artifacts and inaccuracies in color representation. To mitigate some of these drawbacks, adaptive and iterative interpolation methods have been developed, offering improved results at the cost of increased computational demands and slower processing times, hindering real-time application.35 Dictionary-based methods, on the other hand, employ interpolation strategies informed by learned image patches,36 leading to enhanced spatial detail and more accurate color reproduction. These approaches draw from a predefined set of patterns to better reconstruct the image, balancing efficiency with improved image quality.

Convolutional neural networks (CNNs) have emerged as a powerful tool in addressing both low- and high-level challenges in computer vision, spanning tasks from image enhancement and object recognition to surgical guidance and intraoperative robotics.37,38 Specifically in the realm of image demosaicing, CNNs have outperformed traditional state-of-the-art techniques by mastering the direct translation from mosaic images to their full-resolution counterparts.39 This is achieved through the network’s exposure to vast datasets of mosaic and corresponding full-resolution images, enabling it to discern and learn vital features within the local vicinity of an image, such as edges, noise patterns, inter-channel correlations, and other nuanced details critical for accurately filling in missing color information. Initially conceived for image super-resolution,40 this approach was successfully adapted for color demosaicing, applying to image sensors equipped with Bayer filters or similar color filter arrays, showcasing the versatility and efficacy of CNNs in modern image processing tasks.39

In this paper, we introduce a CNN-based approach for demosaicing images captured with a hexachromatic imaging sensor (Fig. 1). This sensor incorporates two distinct types of pixelated spectral filters arranged in a checkerboard pattern: one set designed to transmit visible light wavelengths from 400 to 700 nm and another set tailored to permit NIR light wavelengths from 700 to 1000 nm. Given the pixelated design of the sensor, the spatial resolution in both the visible and NIR spectra is effectively halved. To address this, we developed a CNN model specifically trained to restore the diminished spatial resolution inherent to our sensor’s unique architecture. The performance of our CNN-based demosaicing method is evaluated against traditional interpolation techniques using a benchmark dataset (referred to as the Waterloo image dataset) and a selection of images captured with a vertically stacked image sensor. In addition, we present a series of pre-clinical and clinical images to demonstrate the practical application and effectiveness of our approach.

Fig. 1

Schematic overview of our bioinspired sensor with spectral sensitivity and a residual CNN for demosaicing color and NIR images: panels (a)–(c) depict the configuration of our hexachromatic image sensor, incorporating pixelated filters and vertically stacked photodiodes. Panel (d) details the quantum efficiency of our bioinspired sensor. Panel (e) illustrates the design of the residual CNN, featuring two separate networks for demosaicing the color and NIR channels.

JBO_29_7_076005_f001.png

2.

Deep Learning-Based Demosaicing Implementation

In this work, we utilized CNN tailored specifically for the task of demosaicing color and NIR images captured by our novel bioinspired hexachromatic imaging sensor. Drawing upon the foundational principles of deep residual learning41 and leveraging insights from super-resolution techniques40 and CNN-based demosaicing research,39 we optimized a very deep CNN within this framework. The architectural blueprint of our CNN demosaicing model is depicted in Fig. 2, showcasing a structure designed to overcome the common pitfalls associated with deep network training.

Fig. 2

Our deep residual CNN is constructed with 20 layers, each layer consisting of a convolutional segment that utilizes 64 filters with 3 × 3 receptive fields, complemented by batch normalization and a SELU activation function. This network is adept at learning residuals, which are subsequently integrated with a bilinearly interpolated image to produce the final demosaiced image. The model incorporates 3D convolution kernels in both the initial and concluding layers, and we provide illustrations of both the intermediate results from the CNN and the ultimate image.

JBO_29_7_076005_f002.png

Deep CNNs, characterized by their numerous layers, operate under the premise that increased depth should theoretically yield higher accuracy in tasks such as image demosaicing. However, empirical observations reveal a paradox where, beyond a certain depth, the model’s performance plateaus and may even deteriorate, an issue not attributable to overfitting.42 This degradation in performance with increased depth highlights challenges in network optimization rather than model complexity per se.

To address the challenges of optimizing deep learning models for image demosaicing, we integrate the residual learning framework into our strategy, leveraging its proven success in the fields of super-resolution and color image demosaicing.39,40 The essence of this approach is its focus on high-frequency details, regarded as residuals, which the network learns to predict directly from the data, rather than reconstructing the entire image anew. This method is especially suited to our application, given the separate spectral information provided by the color and NIR channels and their independent needs for interpolation.

The process begins with both color and NIR images at a resolution that is half of what the sensor is capable of capturing (Fig. 2). To accurately mirror the pixelated filter pattern on the image sensor, the initial pixel of both the color and NIR images is offset by one in the horizontal direction. Initially, missing pixels are estimated using bilinear interpolation, a method chosen for its computational efficiency despite its tendency to introduce zipper artifacts and color inaccuracies. However, these drawbacks are somewhat mitigated in our hexachromatic imager compared with traditional Bayer color filter arrays. The interpolated image serves as the input for the CNN, which is tasked with refining the accuracy of the pixel estimates to achieve full resolution. Importantly, this bilinearly interpolated image retains the low-frequency content of the eventual full-resolution output, given that half of the pixels in both images are identical. Thus, the CNN is specifically trained to discern and learn the high-frequency details that differ, facilitating a more focused and efficient learning process.

This approach not only streamlines the model’s training by emphasizing the learning of high-frequency information over the entire image but also enhances the depth and learning rate capabilities of the CNN. Consequently, the final full-resolution image emerges from the combination of the bilinear interpolated base and the high-frequency residuals refined by CNN.

Our CNN architecture consists of 20 layers, each incorporating a convolutional layer, a batch normalization, and a scaled exponential linear unit (SELU) activation function. Every convolutional layer employs 64 filters with 3 × 3 receptive fields and one-pixel padding to maintain the spatial dimensions of the input, except for the final layer, which uses three filters also with 3×3 receptive fields, tailored for the demosaicing output. The examples of the receptive fields from both the first and last layers are displayed in Fig. 2. The mathematical framework for this demosaicing process is as follows:

Eq. (1)

Fn(Y)=selu(Y*Wn+Bn),  n=1N1,

Eq. (2)

SELU(x)={λxif  x>0λα(ex1)if  x0,

Eq. (3)

F(Y)=FN1(Y)*WN+BN.

In Eq. (1), Y is the input patch; Fn is the output feature map of the n’th layer; Wn and Bn represent the filters and the bias vector of the n’th layer, respectively; and * is the convolution operator. This equation represents a convolution operation and the extraction of useful local features in the image. To maintain the spatial dimensions of the input, we utilized 64 filters with 3×3 receptive fields and one-pixel padding in all layers except the final one. For the final layer [Eq. (3)], we used three filters with 3×3 receptive fields, specifically designed for the demosaicing output. Each layer consists of a convolutional layer, followed by a batch normalization layer, and a SELU activation function. Furthermore, λ and α are constants as defined in the SELU activation function literature. The optimization of the network utilizes a modified L2-norm for the loss function

Eq. (4)

L(Θ)=1ni=1n(F(Yi;Θ)+Y^i)Xi2,
where F(Yi;θ) represents the predicted high-frequency components or residuals that, when added to the bilinearly interpolated image Yi, approximate the ground truth high-resolution image Xi. The network initialization follows the MSRA policy, and the Adam optimizer is employed for gradient updates with a learning rate set to 1×105. This setup ensures that the network effectively learns the residuals necessary for reconstructing the high-resolution image from its lower-resolution counterpart.

In our study, we utilized a dataset comprising 4744 images from the Waterloo Exploration Dataset (WED).43 This dataset was chosen for its diverse range of color scenes and intricate high-frequency patterns, offering the true color values at every pixel. Such variability in color and spatial detail was instrumental in training the CNN to identify features under varying conditions and accurately predict the missing pixel values. From each image, we extracted 100 patches of 50×50  pixels to serve as the training data for the network. The rationale for selecting this patch size is grounded in the nature of demosaicing as a process that predicts missing pixel values based on nearby information; thus, local context is more critical for this task than global context, which might be more important for other computer vision tasks such as object recognition. To enhance our dataset further, we applied data augmentation techniques to each patch by rotating them four times at 90-deg intervals and applying horizontal flips, thereby enriching the training data and improving the robustness of the network. Of the WED images, 4644 are used for training purposes, and 100 are used for testing.

To determine the optimal number of layers for the residual CNN model, we trained models with 5, 10, and 15 layers using the WED dataset. The average peak signal-to-noise ratio (PSNR) for image reconstruction on 100 WED test images was 33.5 dB for the 5-layer model, 33.8 dB for the 10-layer model, and 34.5 dB for the 15-layer model. All of these performances were lower than the 20-layer model, which achieved a PSNR of 35.81 dB. Given the balance between training resources and model complexity, the 20-layer model is considered optimal.

The architecture of our CNN is designed to process both color and NIR images with the same underlying structure. Given the minimal spectral overlap between the color and NIR spectrums, we posit that the interrelation between these pixel sets is negligible, allowing for their independent analysis. We anticipate that the NIR channels will display high-frequency structures akin to those found in color images. An example of this is seen with NIR fluorescence emitted by indocyanine green, which not only illuminates lymph nodes but also outlines the lymphatic vessels.29 These vessels, often only a few pixels in width, significantly contribute to the high-frequency detail in the images. Therefore, it is reasonable to expect that NIR images will contain high-frequency information comparable to that of color images. The principal distinction lies in the spectral sensitivity, which differs markedly between our NIR and visible channels in comparison with the Bayer color imaging sensor. Addressing this disparity to optimize the six-channel spectral estimation for full-resolution image reconstruction is a goal for future development.

3.

Evaluation of Residual CNNs for Demosaicing

We evaluate the performance of our demosaicing algorithm using several datasets: Waterloo color images captured with a Bayer color filter array; color images obtained with the Foveon X3 sensor (Fig. 3); the “UIUC Color NIR X3” dataset consisting of color and NIR scenes from the University of Illinois campus captured with vertically stacked photodiodes; the “UIUC NIR Preclinical” dataset featuring animal models of breast cancer imaged with vertically stacked photodiodes and tagged with NIR fluorescent probes; and the “UIUC NIR Clinical” dataset, which includes clinical images from patients with lung or breast cancer, also tagged with NIR fluorescent probes and captured with vertically stacked photodiodes.

Fig. 3

Visual comparison of demosaiced images obtained through bilinear interpolation and the residual CNN framework reveals significant differences. Close-up views highlight how the CNN method successfully avoids zig-zag and false color artifacts, which are prominently visible in images processed by bilinear interpolation.

JBO_29_7_076005_f003.png

The UIUC Color NIR X3 and UIUC NIR Preclinical datasets provide full-resolution images across six spectral channels per pixel: three in the visible spectrum and three in the NIR spectrum. The datasets were captured using a bare image sensor equipped with vertically stacked detectors, omitting any pixelated filter array. Two sets of images were acquired: one using a low-pass filter that blocks all NIR wavelengths above 700 nm and another with a high-pass filter that blocks the transmission of visible photons below 700 nm. This imaging approach simulates the spectral sensitivity profile of our pixelated hexachromatic image sensor. Conversely, the UIUC NIR Clinical dataset lacks ground truth data, containing only pixels with either visible or NIR information as captured by our bioinspired sensor. Given the critical importance of this dataset for assessing our sensor’s clinical performance, we developed an accuracy metric inspired by analogous methodologies in the color imaging domain. Finally, the Waterloo dataset is used to evaluate the demosaicing algorithms on color data collected with a Bayer color sensor. Because we use 90% of this dataset to train our CNN model, we use the remaining 10% to evaluate its demosaicing results. The downside of this evaluation is that the difference in spectral response between the two cameras can provide slight biases in the CNN toward Bayer-type sensors.

3.1.

Color Demosaicing on Images Obtained with Foveon-Sigma Camera

To evaluate the performance of our CNN-based demosaicing routine on images with diverse colors and textures, we utilized an unmosaiced reference image taken by a commercial camera (DP1x, Sigma, Marietta, Georgia, United States) without a color filter array. A short-pass filter, allowing only the visible spectrum, was attached atop the camera to capture a still image of 2640×1760  pixels. This reference image, depicted in Fig. 3(a), showcases the Electrical and Computer Engineering Building at the University of Illinois at Urbana-Champaign.

We then applied a color–NIR filter pattern to the reference image to simulate the visible pixels captured by a hexachromatic camera. This mosaiced color image served as the input to our model. To assess our model’s accuracy, we compared the demosaiced images with the original unmosaiced reference image, treating it as the ground truth for evaluating the image reconstruction quality. As a comparative baseline, a one-dimensional (1D) bilinear interpolation was implemented alongside our model.

In Fig. 3(b), two 200×200 patches, one of a canopy and the other of a façade, illustrate the comparison. The bilinear demosaicing of the canopy exhibits numerous artificial artifacts on the leaves and visible mosaicing, whereas our model significantly reduces these imperfections, resulting in a reconstruction closer to the ground truth. Similarly, for the façade, our model produces smoother reconstructions of high-contrast elements, such as wires, outperforming the rougher appearance generated by bilinear interpolation.

For the quantitative evaluation of the demosaicing performance, we employed four metrics: PSNR, mean squared error (MSE), the 95th percentile of the structural dissimilarity index measure (DSSIM), and the 95th percentile color difference (ΔE). Our CNN model showed substantial improvements across these metrics compared with bilinear interpolation, with increases in PSNR across the color channels and significant reductions in MSE, indicating overall enhanced image quality. Specifically, the PSNR improved by an average of 2.25 dB, and the MSE saw a 40.6% reduction (see Table 1). Using an independent two-sample t test on the residuals of CNN and 1D bilinear interpolation with 100 images from the Waterloo dataset, we obtained p values of 0.00027, 0.00003, and 0.00004 for the three color channels. These results indicate a statistically significant difference in PSNR between the two methods.

Table 1

Effectiveness of our residual CNN compared with bilinear interpolation using the Waterloo dataset, along with full images and sub-images from the Foveon X3, which demonstrates that the residual CNN consistently surpasses the bilinear demosaicing method in all evaluated metrics and datasets.

MetricWaterloo 100Foveon X3 full resFoveon X3 canopyFoveon X3 façade
1D bilinearResidual CNN1D bilinearResidual CNN1D bilinearResidual CNN1D bilinearResidual CNN
PSNR–Vis R (dB)33.22 ± 3.5735.97 ± 3.7029.1431.3924.8026.1326.3834.09
PSNR–Vis G (dB)33.06 ± 3.0535.81 ± 3.2829.1231.3224.7626.1228.6635.78
PSNR–Vis B (dB)32.99 ± 2.9835.67 ± 3.2829.6431.9825.4526.9029.7535.79
MSE–Vis R0.387 ± 0.3160.215 ± 0.1970.6600.3932.2121.6261.4710.249
MSE–Vis G0.360 ± 0.2810.201 ± 0.1940.7060.4252.2501.6441.6890.327
MSE–Vis B0.402 ± 0.3900.234 ± 0.3000.6630.3872.4991.7901.6010.399
% ΔE4.203 ± 1.9333.120 ± 1.5097.3285.07613.07211.48310.3493.988
% DSSIM0.138 ± 0.0600.096 ± 0.0580.1720.1400.2160.1790.0680.025

Focusing on the reconstruction quality of smaller image patches with detailed colors and textures, we analyzed two 200 × 200 pixel patches: one of the building façade and one of the canopy. For these detailed sections, our model still outperformed the baseline but with slightly reduced margins compared with the full-image analysis. Nevertheless, the performance in reconstructing high-contrast details, such as the building façade, was particularly noteworthy, with our model achieving significantly better PSNR, a 79.5% reduction in MSE, and marked improvements in DSSIM and ΔE, underscoring our model’s superior capability in handling images with high-contrast patterns (see Table 1). Comparable improvements in image quality were also noted on the Waterloo image dataset, as detailed in Table 1.

3.2.

Color and NIR Demosaicing on Hexachromatic Images

To broaden the scope of our model to encompass multi-spectrum imaging, we assessed our demosaicing approach using 20 images from the UIUC Color NIR X3 dataset, published by Blair and Gruev.31 These images were captured with a custom camera equipped with both short-pass and long-pass filters, selectively blocking the NIR and visible spectrums, respectively. This setup, featuring a sensor with three vertically stacked photodiodes per pixel, enables the generation of unmosaiced ground truth images in both the visible and NIR spectrums by alternating the filters at a 700 nm threshold.

Our CNN model significantly outperforms bilinear interpolation in reconstructing images across the visible spectrum. Specifically, the model achieves a PSNR increase of 2.41, 2.50, and 2.50 dB for the blue, green, and red channels, respectively, leading to an average PSNR improvement of 2.47 dB for the visible channels. The model also reduces the MSE by 41.4% compared with the baseline. In terms of similarity, the 95% DSSIM is improved by 24.0% with our model. In addition, the color difference across the three channels, measured by 95% ΔE, shows a 14.2% improvement with our model (see Table 2).

Table 2

Our residual CNN against bilinear interpolation across various datasets, including UIUC Color NIR X3, UIUC NIR Preclinical, and UIUC NIR Clinical images for breast and lung cancer, which clearly shows that the CNN method outperforms bilinear demosaicing across all metrics and datasets consistently.

MetricUIUC Color NIR X3UIUC NIR PreclinicalBreast cancer clinical imageLung cancer clinical image
1D bilinearResidual CNN1D bilinearResidual CNN1D bilinearResidual CNN1D bilinearResidual CNN
PSNR–Vis R (dB)38.84 ± 3.1941.25 ± 2.7837.60 ± 1.2439.76 ± 1.5643.9546.5538.4944.06
PSNR–Vis G (dB)38.89 ± 2.8641.39 ± 2.8437.17 ± 1.2139.41 ± 1.3544.3947.1638.5843.92
PSNR–Vis B (dB)39.54 ± 2.7542.04 ± 3.0239.49 ± 1.5141.49 ± 1.8243.9645.8438.1844.53
MSE–Vis R0.198 ± 0.1660.111 ± 0.0870.179 ± 0.0540.116 ± 0.0490.0420.0230.1200.033
MSE–Vis G0.189 ± 0.1460.110 ± 0.0970.183 ± 0.0500.115 ± 0.0450.0370.0190.1120.032
MSE–Vis B0.174 ± 0.1450.106 ± 0.1020.145 ± 0.0510.099 ± 0.0470.0340.0220.1360.031
% ΔE–Vis2.698 ± 0.7722.315 ± 0.4692.373 ± 0.4411.979 ± 0.3581.2911.1732.0671.975
% DSSIM–Vis0.107 ± 0.0290.081 ± 0.0280.102 ± 0.0520.081 ± 0.0410.0220.0160.0520.048
PSNR–NIR 1 (dB)34.04 ± 2.4835.84 ± 2.5647.05 ± 0.4647.83 ± 1.2351.6654.8647.5248.58
PSNR–NIR 2 (dB)37.44 ± 3.2439.11 ± 3.1654.10 ± 1.8754.27 ± 1.6351.6154.1051.3052.01
PSNR–NIR 3 (dB)40.91 ± 3.4642.28 ± 3.2455.43 ± 1.1755.54 ± 1.7138.7642.8952.8453.06
MSE–NIR 10.360 ± 0.2360.231 ± 0.1400.033 ± 0.0080.020 ± 0.0040.0950.0730.0970.076
MSE–NIR 20.302 ± 0.1790.201 ± 0.0670.072 ± 0.0250.077 ± 0.0270.0890.0500.2100.178
MSE–NIR 30.195 ± 0.1060.137 ± 0.0690.112 ± 0.0710.116 ± 0.0750.0930.0520.0770.074
% ΔE–NIR3.506 ± 1.0292.923 ± 0.7960.787 ± 0.0870.785 ± 0.0870.5960.3780.8130.794
% DSSIM–NIR0.116 ± 0.0290.104 ± 0.0320.039 ± 0.0150.036 ± 0.0170.0480.0290.0320.031

For the NIR spectrum images, our model’s performance demonstrates an enhancement, with PSNR increases of 1.80, 1.67, and 1.37 dB across the NIR channels, culminating in an average PSNR gain of 1.61 dB. The MSE is reduced by 33.0%, and the 95% DSSIM sees a 10.8% increase. The color difference improvement in the NIR channels, indicated by 95% ΔE, is 16.6%. These results highlight our model’s superior demosaicing capabilities over bilinear methods for both visible and NIR spectrums (see Table 2). However, the performance on NIR images suggests room for improvement, likely due to the model being primarily trained on visible spectrum images.

3.3.

Preclinical Evaluation of Residual CNNs for Color and NIR Image Demosaicing

Expanding our model to include pre-clinical animal studies, we evaluated it using a publicly available dataset featuring three pairs of unmosaiced visible and NIR in vivo images of female mice with breast tumors (4T1, American Type Culture Collection, Rockville, Maryland, United States), introduced subcutaneously and grown to a diameter of 1 cm. To highlight the breast tumors with NIR fluorescence, each mouse received an injection of IRDye 800CW Maleimide (100  μL at 11.91  μg per mL in phosphate-buffered saline) into the retro-orbital sinus, allowing for tumor accumulation over 24 h. Image collection was performed using a custom camera equipped with three stacked photon detectors and an excitation filter to omit NIR excitation light. Visible images were captured under white light, and NIR images were obtained under infrared illumination (I0785MU6000M4S, Innovative Photonic Solutions, Plainsboro, New Jersey, United States).

Applying our model to this dataset yielded notable enhancements across various metrics for both visible and NIR images (see Fig. 4). The PSNR saw increases of 2.15, 2.26, and 2.00 dB across the color channels, with an average PSNR boost of 2.13 dB for the trio. The CNN model’s reconstruction reduced the MSE by 34.7% compared with the baseline. Similarly, the average DSSIM showed a 37.0% improvement, and the color difference (ΔE) was enhanced by 13.6% (see Table 2).

Fig. 4

Side-by-side visual evaluation of demosaicing using bilinear interpolation and the residual CNN approach on an animal model with breast cancer illustrates the advantages. The comparison of color and NIR fluorescence images demonstrates how the CNN demosaicing method uncovers superior high-resolution details compared with the bilinear technique.

JBO_29_7_076005_f004.png

In the NIR spectrum, the PSNR improvement for each channel was 0.78, 0.15, and 0.09 dB due to the low-frequency information of NIR fluorescence and lower sensitivity in the NIR 2 and 3 channels. The model’s MSE reduction in this spectrum was 1.8%, and the 95% DSSIM and NIR color differences show very limited enhancement (see Table 2).

3.4.

Clinical Evaluation of Residual CNNs for Color and NIR Image Demosaicing

Clinical data were acquired using our hexachromatic image sensor in two distinct scenarios (Fig. 5). First, during breast cancer surgery, patients received a peritumoral injection of indocyanine green to delineate the sentinel lymph nodes. The image sensor was mounted above the operation area, enabling the real-time capturing of both color and NIR fluorescence videos as surgeons conducted lymph node mapping. In the second scenario, lung cancer patients undergoing surgical procedures received VGT-309, a cathepsin-targeting indocyanine green (ICG) agent designed to illuminate the tumor microenvironment through fluorescence in areas with heightened cathepsin activity. Ex vivo tissues were subsequently imaged directly in the operating room using our hexachromatic sensor, facilitating the detailed observation of tumor-specific fluorescence signals.

Fig. 5

(a), (b) Color and NIR images taken in vivo in the operating room from a patient during breast cancer surgery. (c), (d) Color and NIR images taken ex vivo on the back table in the operating room from a patient undergoing surgery for lung cancer. The arrows point out the zipper artifacts at the edges in both imaging types when processed through bilinear interpolation, which are significantly reduced by employing our CNN methodology.

JBO_29_7_076005_f005.png

Figures 5 and 6 display the original images, those processed by bilinear interpolation, and images enhanced using CNN interpolation. The images enhanced by the CNN display sharper edges and fewer zipper artifacts compared with those processed with bilinear interpolation, as indicated by the arrows on the images in Fig. 5. In addition, Fourier transform analyses reveal variations in the frequency content between the two demosaicing methods relative to the original images (Fig. 6). As anticipated, the bilinear method exhibits a greater discrepancy from the original in terms of high-frequency content due to its inherent limitations. Conversely, the CNN approach, by learning and accurately reconstructing the high-frequency details, significantly reduces errors in these areas of the reconstructed images. This improvement is clearly observable both in the fast Fourier transform (FFT) difference visuals and in the single-line plot across the vertical of the FFT representation. An analysis along the vertical frequencies highlights a 20 dB improvement in the CNN method over bilinear processing at higher frequencies. The images shown in Fig. 6(e) further demonstrate the difference between the original and demosaiced images using bilinear and CNN approaches, respectively. As indicated in the FFT images in Figs. 6(c) and 6(d), the images in Fig. 6(e) show that the CNN method reconstructs the edges more accurately than the bilinear interpolation method.

Fig. 6

(a) Ex vivo color images of lung cancer tissue processed using both bilinear and CNN demosaicing techniques. (b) The magnitude of the Fourier-transformed images reveals that the high-frequency content of the CNN-demosaiced image closely matches that of the original. (c) The comparison between the original and demosaiced images, either by bilinear or CNN methods, further demonstrates the superior spatial reconstruction achieved with the CNN approach. An analysis along the vertical frequencies highlights a 20-dB improvement in the CNN method over bilinear processing at higher frequencies.

JBO_29_7_076005_f006.png

The data shown in Fig. 6 demonstrate that the CNN method more closely aligns with the original image in terms of FFT magnitude, highlighting its superior performance in image reconstruction. Table 2 and Fig. 7 present a summary of the performance metrics, including PSNR, MMSE, DSSIM, and ΔE, highlighting the CNN’s superior performance over the bilinear method in processing these clinical images.

Fig. 7

Bar chart comparison of key metrics between bilinear and CNN demosaicing methods across different datasets summarized in (a) Table 1, (b) Table 2 visible channel, and (c) Table 2 NIR channel. Our residual CNN approach consistently outperforms the bilinear demosaicing method across all datasets and metrics.

JBO_29_7_076005_f007.png

4.

Conclusion

In conclusion, our study demonstrates the significant advancements achieved in the field of image demosaicing for medical imaging through the application of a residual CNN. By comparing our CNN model against traditional bilinear interpolation methods across various datasets, including the challenging settings of pre-clinical animal studies and clinical trials, we evidenced marked improvements in image quality metrics, such as PSNR, MSE, DSSIM, and ΔE. Notably, the application of our model to hexachromatic sensor data in the context of fluorescent cancer imaging not only enhanced the clarity and accuracy of both color and NIR images but also minimized high-frequency artifacts, thereby facilitating a more precise identification and analysis of cancerous tissues.

The clinical utility of our hexachromatic image sensor, combined with the CNN-based demosaicing approach, was further underscored in two distinct surgical scenarios: breast cancer lymph node mapping with indocyanine green and lung cancer surgery using a cathepsin-binding ICG agent. These applications highlighted the sensor’s ability to provide real-time, enhanced visualization of tumor environments and sentinel lymph nodes, significantly aiding surgical decisions and outcomes.

Moving forward, the continued refinement and application of CNNs in image demosaicing hold the promise of not only advancing medical imaging technologies but also contributing to more accurate diagnoses, tailored surgical interventions, and ultimately improved patient care. Our findings advocate for the integration of advanced computational methods such as CNNs into the development of next-generation medical imaging devices, aiming to bridge the gap between current limitations and the potential for high-fidelity, multi-spectral imaging in clinical practice.

Disclosures

All authors declare that they have no known financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Code and Data Availability

The code used to train, validate, and test the neural network was obtained from https://github.com/yifeij7/Residual-CNN and is publicly available.

Ethics

The clinical study was approved by the institutional review boards at the University of Illinois Urbana Champaign and the University of Pennsylvania. All animal studies were approved by the Institutional Animal Care and Use Committee (IACUC) at the University of Illinois Urbana-Champaign.

Acknowledgments

This work was funded by grants from the U.S. Air Force Office of Scientific Research (Grant No. FA9550-24-1-0112), the National Science Foundation (Grant Nos. 2030421 and 2344460), the Office of Naval Research (Grant No. N00014-21-1-2177), and the National Institutes of Health (Grant No. 1P01CA254859).

References

1. 

R. L. Siegel et al., “Cancer statistics, 2021,” CA Cancer J. Clin., 71 (1), 7 –33 https://doi.org/10.3322/caac.21654 CAMCAM 0007-9235 (2021). Google Scholar

2. 

H. Sung et al., “Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries,” CA: A Cancer J. Clin., 71 (3), 209 –249 https://doi.org/10.3322/caac.21660 (2021). Google Scholar

3. 

D. Crosby et al., “Early detection of cancer,” Science, 375 (6586), eaay9040 https://doi.org/10.1126/science.aay9040 SCIEAS 0036-8075 (2022). Google Scholar

4. 

L. Wyld, R. A. Audisio and G. J. Poston, “The evolution of cancer surgery and future perspectives,” Nat. Rev. Clin. Oncol., 12 (2), 115 –124 https://doi.org/10.1038/nrclinonc.2014.191 (2015). Google Scholar

5. 

R. Sullivan et al., “Global cancer surgery: delivering safe, affordable, and timely cancer surgery,” Lancet Oncol., 16 (11), 1193 –1224 https://doi.org/10.1016/S1470-2045(15)00223-5 LOANBN 1470-2045 (2015). Google Scholar

6. 

A. L. Vahrmeijer et al., “Image-guided cancer surgery using near-infrared fluorescence,” Nat. Rev. Clin. Oncol., 10 (9), 507 –518 https://doi.org/10.1038/nrclinonc.2013.123 (2013). Google Scholar

7. 

G. Hong, A. L. Antaris and H. Dai, “Near-infrared fluorophores for biomedical imaging,” Nat. Biomed. Eng., 1 (1), 0010 https://doi.org/10.1038/s41551-016-0010 (2017). Google Scholar

8. 

Y. Chen, S. Wang and F. Zhang, “Near-infrared luminescence high-contrast in vivo biomedical imaging,” Nat. Rev. Bioeng., 1 (1), 60 –78 https://doi.org/10.1038/s44222-022-00002-8 (2023). Google Scholar

9. 

A. V. DSouza et al., “Review of fluorescence guided surgery systems: identification of key performance capabilities beyond indocyanine green imaging,” J. Biomed. Opt., 21 (8), 080901 https://doi.org/10.1117/1.JBO.21.8.080901 JBOPFO 1083-3668 (2016). Google Scholar

10. 

T. Hollon et al., “Artificial-intelligence-based molecular classification of diffuse gliomas using rapid, label-free optical imaging,” Nat. Med., 29 (4), 828 –832 https://doi.org/10.1038/s41591-023-02252-4 1078-8956 (2023). Google Scholar

11. 

P. S. Low, S. Singhal and M. Srinivasarao, “Fluorescence-guided surgery of cancer: applications, tools and perspectives,” Curr. Opin. Chem. Biol., 45 64 –72 https://doi.org/10.1016/j.cbpa.2018.03.002 COCBF4 1367-5931 (2018). Google Scholar

12. 

J. S. D. Mieog et al., “Fundamentals and developments in fluorescence-guided cancer surgery,” Nat. Rev. Clin. Oncol., 19 (1), 9 –22 https://doi.org/10.1038/s41571-021-00548-3 (2022). Google Scholar

13. 

G. M. Van Dam et al., “Intraoperative tumor-specific fluorescence imaging in ovarian cancer by folate receptor-α targeting: first in-human results,” Nat. Med., 17 (10), 1315 –1319 https://doi.org/10.1038/nm.2472 1078-8956 (2011). Google Scholar

14. 

E. I. Sega and P. S. Low, “Tumor detection using folate receptor-targeted imaging agents,” Cancer Metastasis Rev., 27 655 –664 https://doi.org/10.1007/s10555-008-9155-6 CMRED4 (2008). Google Scholar

15. 

B. L. Smith et al., “Intraoperative fluorescence guidance for breast cancer lumpectomy surgery,” NEJM Evid., 2 (7), EVIDoa2200333 https://doi.org/10.1056/EVIDoa2200333 (2023). Google Scholar

16. 

P. Bou-Samra et al., “Intraoperative molecular imaging: 3rd biennial clinical trials update,” J. Biomed. Opt., 28 (5), 050901 https://doi.org/10.1117/1.JBO.28.5.050901 JBOPFO 1083-3668 (2023). Google Scholar

17. 

E. R. Fossum and D. B. Hondongwa, “A review of the pinned photodiode for CCD and CMOS image sensors,” IEEE J. Electron. Devices Soc., 2 33 –43 https://doi.org/10.1109/JEDS.2014.2306412 (2014). Google Scholar

18. 

P. L. Dillon, D. M. Lewis and F. G. Kaspar, “Color imaging system using a single CCD area array,” IEEE J. Solid-State Circuits, 13 (1), 28 –33 https://doi.org/10.1109/JSSC.1978.1050992 IJSCBC 0018-9200 (1978). Google Scholar

19. 

S. Achilefu et al., “Goggle imaging systems and methods,” (2019). Google Scholar

20. 

M. Garcia et al., “Bio-inspired imager improves sensitivity in near-infrared fluorescence image-guided surgery,” Optica, 5 (4), 413 –422 https://doi.org/10.1364/OPTICA.5.000413 (2018). Google Scholar

21. 

B. Bayer, “Color imaging array,” (1976). Google Scholar

22. 

J. V. Frangioni, “In vivo near-infrared fluorescence imaging,” Curr. Opin. Chem. Biol., 7 (5), 626 –634 https://doi.org/10.1016/j.cbpa.2003.08.007 COCBF4 1367-5931 (2003). Google Scholar

23. 

S. L. Troyan et al., “The FLARE™ intraoperative near-infrared fluorescence imaging system: a first-in-human clinical trial in breast cancer sentinel lymph node mapping,” Ann. Surg. Oncol., 16 2943 –2952 https://doi.org/10.1245/s10434-009-0594-2 (2009). Google Scholar

24. 

B. Zhu and E. Sevick-Muraca, “A review of performance of near-infrared fluorescence imaging devices used in clinical studies,” Br. J. Radiol., 88 (1045), 20140547 https://doi.org/10.1259/bjr.20140547 (2015). Google Scholar

25. 

S. B. Mondal et al., “Real-time fluorescence image-guided oncologic surgery,” Adv. Cancer Res., 124 171 –211 https://doi.org/10.1016/B978-0-12-411638-2.00005-7 ACRSAJ 0065-230X (2014). Google Scholar

26. 

Y. Liu et al., “Near-infrared fluorescence goggle system with complementary metal–oxide–semiconductor imaging sensor and see-through display,” J. Biomed. Opt., 18 (10), 101303 https://doi.org/10.1117/1.JBO.18.10.101303 JBOPFO 1083-3668 (2013). Google Scholar

27. 

J. V. Frangioni, “New technologies for human cancer imaging,” J. Clin. Oncol., 26 (24), 4012 https://doi.org/10.1200/JCO.2007.14.3065 JCONDN 0732-183X (2008). Google Scholar

28. 

P. B. Garcia-Allende et al., “Towards clinically translatable NIR fluorescence molecular guidance for colonoscopy,” Biomed. Opt. Express, 5 (1), 78 –92 https://doi.org/10.1364/BOE.5.000078 BOEICL 2156-7085 (2014). Google Scholar

29. 

S. Blair et al., “Hexachromatic bioinspired camera for image-guided cancer surgery,” Sci. Transl. Med., 13 (592), eaaw7067 https://doi.org/10.1126/scitranslmed.aaw7067 (2021). Google Scholar

30. 

M. B. George et al., “Bioinspired color-near infrared endoscopic imaging system for molecular guided cancer surgery,” J. Biomed. Opt., 28 (5), 056002 https://doi.org/10.1117/1.JBO.28.5.056002 JBOPFO 1083-3668 (2023). Google Scholar

31. 

S. Blair and V. Gruev, “Interpolant-based demosaicing routines for dual-mode visible/near-infrared imaging systems,” Opt. Express, 30 (19), 34201 –34217 https://doi.org/10.1364/OE.458101 OPEXFF 1094-4087 (2022). Google Scholar

32. 

S. Blair et al., “Decoupling channel count from field of view and spatial resolution in single-sensor imaging systems for fluorescence image-guided surgery,” J. Biomed. Opt., 27 (9), 096006 https://doi.org/10.1117/1.JBO.27.9.096006 JBOPFO 1083-3668 (2022). Google Scholar

33. 

D. Menon and G. Calvagno, “Color image demosaicking: an overview,” Signal Process. Image Commun., 26 (8-9), 518 –533 https://doi.org/10.1016/j.image.2011.04.003 SPICEF 0923-5965 (2011). Google Scholar

34. 

Jr. J. E. Adams, “Interactions between color plane interpolation and other image processing functions in electronic photography,” Proc. SPIE, 2416 144 –151 https://doi.org/10.1117/12.204825 PSISDG 0277-786X (1995). Google Scholar

35. 

Y. Monno et al., “Adaptive residual interpolation for color image demosaicking,” in IEEE Int. Conf. Image Process. (ICIP), 3861 –3865 (2015). https://doi.org/10.1109/ICIP.2015.7351528 Google Scholar

36. 

J. Mairal, M. Elad and G. Sapiro, “Sparse representation for color image restoration,” IEEE Trans. Image Process., 17 (1), 53 –69 https://doi.org/10.1109/TIP.2007.911828 IIPRE4 1057-7149 (2007). Google Scholar

37. 

A. Krizhevsky, I. Sutskever and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Adv. Neural Inf. Process. Syst. 25, (2012). Google Scholar

38. 

K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” (2014). Google Scholar

39. 

N.-S. Syu, Y.-S. Chen and Y.-Y. Chuang, “Learning deep convolutional networks for demosaicing,” (2018). Google Scholar

40. 

J. Kim, J. K. Lee and K. M. Lee, “Accurate image super-resolution using very deep convolutional networks,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 1646 –1654 (2016). https://doi.org/10.1109/CVPR.2016.182 Google Scholar

41. 

K. He et al., “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 770 –778 (2016). https://doi.org/10.1109/CVPR.2016.90 Google Scholar

42. 

K. He and J. Sun, “Convolutional neural networks at constrained time cost,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 5353 –5360 (2015). https://doi.org/10.1109/CVPR.2015.7299173 Google Scholar

43. 

K. Ma et al., “Waterloo exploration database: new challenges for image quality assessment models,” IEEE Trans. Image Process., 26 (2), 1004 –1016 https://doi.org/10.1109/TIP.2016.2631888 IIPRE4 1057-7149 (2017). Google Scholar

Biography

Yifei Jin is a graduate student in the electrical and computer engineering program at the University of Illinois at Urbana-Champaign. His research interest includes design and development of multispectral imaging systems for image guided cancer surgeries.

Borislav Kondov is a thoracic surgeon at the University Clinic Hospital in North Macedonia. The heart of his research and clinical practice in North Macedonian hospitals is the improvement of surgical techniques for patients with breast and lung cancer, including pioneering the first fluorescent-guided and radiocolloid-guided surgeries for breast cancer patients in the country.

Goran Kondov is a distinguished thoracic surgeon at the University Clinic Hospital in North Macedonia. Central to his research and clinical practice is the enhancement of surgical procedures for breast and lung cancer patients within North Macedonian hospital environments. Notably, he pioneered the first fluorescent-guided and radiocolloid-guided surgeries for breast cancer patients in the country.

Sunil Singhal is a thoracic surgeon who is nationally recognized for his expertise in lung cancer, mediastinal tumors, and mesothelioma. He is the William Maul Measey Professor in the Department of Surgery at the University of Pennsylvania of Perelman School of Medicine, a chief of the Division of Thoracic Surgery, and the vice chair of Translational Research for the Department of Surgery.

Shuming Nie is the Grainger Distinguished Chair in Engineering, a professor of bioengineering, chemistry, materials science and engineering, electrical and computer engineering at the University of Illinois at Urbana-Champaign, and founding dean of the College of Engineering and Applied Sciences of Nanjing University, China. His academic research is primarily in the areas of cancer nanomedicine, image-guide surgery, and cancer immunotherapy. His scholarly work has been cited well over 90,000 times.

Viktor Gruev is the Wendell and Rita Dunning Faculty Scholar Professor at the University of Illinois at Urbana-Champaign. His research focuses on bio-inspired sensory technology, particularly, the visual system of the mantis shrimp. He has developed a range of bioinspired imaging sensors, including color-polarization, UV-enhanced cameras, and multispectral sensors. His imaging technology is used in diverse applications, from medical fields, such as intraoperative cancer detection to underwater polarization-based geolocalization.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Yifei Jin, Borislav Kondov, Goran Kondov, Sunil Singhal, Shuming Nie, and Viktor Gruev "Convolutional neural network advances in demosaicing for fluorescent cancer imaging with color–near-infrared sensors," Journal of Biomedical Optics 29(7), 076005 (23 July 2024). https://doi.org/10.1117/1.JBO.29.7.076005
Received: 26 March 2024; Accepted: 17 June 2024; Published: 23 July 2024
Advertisement
Advertisement
KEYWORDS
Near infrared

Interpolation

Image sensors

Optical filters

Tunable filters

Image restoration

Image filtering

Back to Top