Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.12188/34744
Title: Hybrid CNN-transformer demosaicing for bioinspired single-chip color-near-infrared fluorescence imaging in oncologic surgery
Authors: Jin Y
Yang J
Kondov B 
Kondov G 
Singhal S
Forsyth D
Cunningham BT
Nie S
Gruev V
Keywords: bioinspired sensors
cancer surgery
convolutional neural network
demosaicing
image guided surgery
near infrared imaging
transformers
Issue Date: Oct-2025
Publisher: J Biomed Opt
Source: Jin Y, Yang J, Kondov B, Kondov G, Singhal S, Forsyth D, Cunningham BT, Nie S, Gruev V. Hybrid CNN-transformer demosaicing for bioinspired single-chip color-near-infrared fluorescence imaging in oncologic surgery. J Biomed Opt. 2025 Oct;30(10):106008. doi: 10.1117/1.JBO.30.10.106008. Epub 2025 Oct 28. PMID: 41164265; PMCID: PMC12561625.
Journal: Journal of Biomedical Optics
Abstract: Significance: Single-chip multispectral imaging sensors with vertically stacked photodiodes and pixelated spectral filters enable advanced, real-time visualization for image-guided cancer surgery. However, their design inherently reduces spatial resolution. We present a convolutional neural network (CNN)-transformer demosaicing algorithm, validated on both clinical and preclinical datasets that effectively doubles spatial resolution and improves image quality-substantially enhancing intraoperative cancer visualization. Aim: We present a CNN-transformer-based demosaicing approach specifically optimized for reconstructing high-resolution color and NIR images acquired by a hexachromatic imaging sensor. Approach: A hybrid CNN-transformer demosaicing model was developed and trained on color-image datasets, then rigorously evaluated on color and NIR images to demonstrate superior reconstruction quality compared with conventional bilinear interpolation and residual CNN methods. Results: Our CNN-transformer demosaicing method achieves an average mean squared error (MSE) reduction of ∼85% for color images and 76% for NIR images and improves structural dissimilarity by roughly 72% and 79%, respectively, compared with state-of-the-art CNN-based demosaicing algorithms in preclinical datasets. In clinical datasets, our approach similarly demonstrates significant reductions in MSE and structural dissimilarity, substantially outperforming existing CNN-based methods, particularly in reconstructing high-frequency image details. Conclusions: We demonstrate improvements in spatial resolution and image fidelity for color and NIR images obtained from hexachromatic imaging sensors, achieved by integrating convolutional neural networks with transformer architectures. Given recent advances in GPU computing, our CNN-transformer approach offers a practical, real-time solution for enhanced multispectral imaging during cancer surgery.
URI: http://hdl.handle.net/20.500.12188/34744
Appears in Collections:Faculty of Medicine: Journal Articles

Files in This Item:
File SizeFormat 
JBO-030-106008.pdf6.35 MBAdobe PDFView/Open
Show full item record

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.