Please use this identifier to cite or link to this item:
http://hdl.handle.net/20.500.12188/34744| Title: | Hybrid CNN-transformer demosaicing for bioinspired single-chip color-near-infrared fluorescence imaging in oncologic surgery | Authors: | Jin Y Yang J Kondov B Kondov G Singhal S Forsyth D Cunningham BT Nie S Gruev V |
Keywords: | bioinspired sensors cancer surgery convolutional neural network demosaicing image guided surgery near infrared imaging transformers |
Issue Date: | Oct-2025 | Publisher: | J Biomed Opt | Source: | Jin Y, Yang J, Kondov B, Kondov G, Singhal S, Forsyth D, Cunningham BT, Nie S, Gruev V. Hybrid CNN-transformer demosaicing for bioinspired single-chip color-near-infrared fluorescence imaging in oncologic surgery. J Biomed Opt. 2025 Oct;30(10):106008. doi: 10.1117/1.JBO.30.10.106008. Epub 2025 Oct 28. PMID: 41164265; PMCID: PMC12561625. | Journal: | Journal of Biomedical Optics | Abstract: | Significance: Single-chip multispectral imaging sensors with vertically stacked photodiodes and pixelated spectral filters enable advanced, real-time visualization for image-guided cancer surgery. However, their design inherently reduces spatial resolution. We present a convolutional neural network (CNN)-transformer demosaicing algorithm, validated on both clinical and preclinical datasets that effectively doubles spatial resolution and improves image quality-substantially enhancing intraoperative cancer visualization. Aim: We present a CNN-transformer-based demosaicing approach specifically optimized for reconstructing high-resolution color and NIR images acquired by a hexachromatic imaging sensor. Approach: A hybrid CNN-transformer demosaicing model was developed and trained on color-image datasets, then rigorously evaluated on color and NIR images to demonstrate superior reconstruction quality compared with conventional bilinear interpolation and residual CNN methods. Results: Our CNN-transformer demosaicing method achieves an average mean squared error (MSE) reduction of ∼85% for color images and 76% for NIR images and improves structural dissimilarity by roughly 72% and 79%, respectively, compared with state-of-the-art CNN-based demosaicing algorithms in preclinical datasets. In clinical datasets, our approach similarly demonstrates significant reductions in MSE and structural dissimilarity, substantially outperforming existing CNN-based methods, particularly in reconstructing high-frequency image details. Conclusions: We demonstrate improvements in spatial resolution and image fidelity for color and NIR images obtained from hexachromatic imaging sensors, achieved by integrating convolutional neural networks with transformer architectures. Given recent advances in GPU computing, our CNN-transformer approach offers a practical, real-time solution for enhanced multispectral imaging during cancer surgery. | URI: | http://hdl.handle.net/20.500.12188/34744 |
| Appears in Collections: | Faculty of Medicine: Journal Articles |
Files in This Item:
| File | Size | Format | |
|---|---|---|---|
| JBO-030-106008.pdf | 6.35 MB | Adobe PDF | View/Open |
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.