Abstract.
Significance
Oral cancer surgery requires accurate margin delineation to balance complete resection with post-operative functionality. Current in vivo fluorescence imaging systems provide two-dimensional margin assessment yet fail to quantify tumor depth prior to resection. Harnessing structured light in combination with deep learning (DL) may provide near real-time three-dimensional margin detection.
Aim
A DL-enabled fluorescence spatial frequency domain imaging (SFDI) system trained with in silico tumor models was developed to quantify the depth of oral tumors.
Approach
A convolutional neural network was designed to produce tumor depth and concentration maps from SFDI images. Three in silico representations of oral cancer lesions were developed to train the DL architecture: cylinders, spherical harmonics, and composite spherical harmonics (CSHs). Each model was validated with in silico SFDI images of patient-derived tongue tumors, and the CSH model was further validated with optical phantoms.
Results
The performance of the CSH model was superior when presented with patient-derived tumors (). The CSH model could predict depth and concentration within 0.4 mm and , respectively, for in silico tumors with depths less than 10 mm.
Conclusions
A DL-enabled SFDI system trained with in silico CSH demonstrates promise in defining the deep margins of oral tumors.
Keywords: molecular-guided surgery, fluorescence imaging, spatial frequency domain imaging, optical tomography, deep learning, oral cancer surgery
1. Introduction
Of the 10 most common solid tumors, oral cancer has the highest prevalence of positive margins following surgery.1 This has a negative impact on patient outcomes, as clear surgical margins are a consistent prognostic factor for local control and overall survival in oral cancer.2,3 Surgical intervention is required for most oral cancer patients, yet tumor delineation remains a significant challenge.4,5 With these procedures, a balance is required between complete tumor resection and conservation of healthy tissue to prevent tumor recurrence and preserve functionality.6 Although palpation, visual cues, pathology, and anatomical imaging modalities are common intraoperative techniques to define the tumor, these methods are limited in their precision; consequently, the incidence of positive margins in oral cancer has been reported as high as 85%.7,8
Fluorescence-guided surgery is emerging as an intraoperative imaging technique for cancer surgery, as it provides high-resolution contrast to differentiate tumor from healthy tissue.9 A growing number of fluorescence contrast agents are approved or currently under investigation across a wide spectrum of surgical specialities.10 In oral cancer, recent clinical trials are establishing fluorescence imaging for intraoperative assessment of surgical margins.11,12 The use of near-infrared (NIR) fluorophores (650 to 900 nm) enables visualization of sub-surface structures due to lower absorption in the biological imaging window.13 Although this can provide surgical guidance in many clinical scenarios, one key limitation is that clinical fluorescence devices do not quantify the depth of sub-surface structures.14,15 This is particularly important for oral cancer, as determining the depth of tumor invasion is often the key challenge leading to positive margins.16 The sentinel margin technique developed by van Keulen et al. is an important advance to help identify the closest sub-surface margin based on relative fluorescence measurements of the ex vivo surgical specimen,17,18 but as yet this lacks quantitative capabilities. Results from a phase II fluorescence guidance trial demonstrate the challenge of oral cancer depth assessment.11 Specifically, de Wit et al. report that the accuracy of margin classification degraded as margin thickness increased (i.e., from 1 to 3 mm to 3 to 5 mm of overlaying tissue in post-resection surgical specimens). These same authors also suggest that fluorescence depth detection might improve results, citing the potential for variable aperture (VA) and spatial frequency domain imaging (SFDI) techniques.11
Several imaging techniques are under development for intraoperative fluorescence depth detection. Using VA, the dual-aperture fluorescence ratio technique recently reported by Rounds et al.19 accurately classified close versus clear margins in ex vivo surgical specimens of an initial patient cohort. For depth quantification using SFDI, techniques have been developed to quantify the depth to sub-surface inclusions based on the fluorescence decay rate across spatial frequencies.20,21 Alternative techniques using hyperspectral acquisition (i.e., multi-excitation and/or multi-emission wavelength bands) and ratiometric techniques have also been validated in pre-clinical neurosurgery models of buried tumors.22–24 All of these examples focus on identifying the top surface of buried tumors lying below the tissue surface. In these cases, which we refer to as a “submarine” model, depth is defined as the thickness of healthy tissue covering a fluorescent inclusion. These techniques have been motivated by clinical applications including guidance of neurosurgical procedures that involve “piecemeal” resection of layers of tissue overlying a buried tumor20–24 or for imaging the basal surface of an ex vivo surgical specimen.19,25 As oral cancer tends to originate at the mucosal surface and infiltrate deeper into healthy tissue,16 here, we focus on identifying the bottom surface of infiltrative tumors, which we refer to as an “iceberg” model. In this case, depth is the thickness of the fluorescent tumor, as would be encountered during in vivo imaging of the mucosal surface in oral cancer.
For depth quantification, we are using SFDI acquisition in combination with deep learning (DL) reconstruction. SFDI captures wide-field reflectance images while projecting patterns of light to enable quantification of the optical scattering and absorption properties of the tissue.26,27 SFDI fluorescence images provide depth information: high-frequency patterns capture shallow information, and low-frequency patterns obtain deeper information.28 Many existing fluorescence depth quantification approaches use ratiometric or analytical methods with approximations to simplify the inverse problem, such as assumptions of semi-infinite mediums or fluorescence originating from point-source inclusions.20,22–24 DL techniques offer the potential to use learning-based techniques to relax the requirement for mathematical simplicity or complex numerical inversion.29,30 Furthermore, although time-intensive to train, DL methods rapidly generate inferences after image acquisition, which lends itself to intraoperative applications. Here, we adopt an in silico training approach that directly incorporates the physics of light propagation into the training process.31,32 The lack of SFDI images of patient tumors and the requirement for a comprehensive training dataset presented a need for synthetic images. Here, we investigate a custom “Siamese” convolutional neural network (CNN) for tumor delineation and corresponding in silico model for training. In this initial investigation, we employ 10 patient contours from pre-operative radiological imaging to generate optical testing data for simulations and phantom experiments.
2. Methods
2.1. Spatial Frequency Domain Imaging System
The prototype SFDI system under development is in Fig. 1. A light engine with six light-emitting diodes (LEDs) (Spectra X, Lumencor, Beaverton, Oregon, United States) that has 20-nm bandwidth sources centered at 390, 438, 475, 512, 586, and 632 nm is the light source for the system. A 3-mm liquid light guide (LGG0338, Thorlabs, Newton, New Jersey, United States) couples the light engine to the digital light projector (DLP), a spatial modulator (DLi6500 1080p Optics Bundle, DLi, Austin, Texas, United States) based on a DLP development kit (LightCrafter 6500, Texas Instruments, Dallas, Texas, United States). All images are captured with a 1392 × 1040-pixel, 14-bit, monochrome charge-coupled device (CCD) camera (Pixelfly USB, PCO AG, Kelheim, Germany) that is coupled to a 25-mm focal length lens (model 67715, Edmund Optics, Barrington, New Jersey, United States). A synchronization cable couples the projector and the camera so that images are acquired sequentially at each spatial frequency. In this study, the SFDI system collects fluorescence and reflectance images across six spatial frequencies using the 632-nm light source. Reflectance images are required to obtain the optical properties of the tissue; absorption and scattering are computed from reflectance images at two separate spatial frequencies using SFDI lookup tables.26,33 A 650-nm-long pass filter (FELH0650, Thorlabs, Newton, New Jersey, United States) is used when capturing fluorescence images of protoporphyrin IX (PpIX). To remove specular reflection, the system includes crossed linear polarizers (LPVISE2X2, Thorlabs, Newton, New Jersey, United States). Prior to data acquisition, the SFDI system is calibrated with a polyurethane-based phantom containing titanium dioxide and carbon black (Biomimic PB0317, INO, Quebec City, Quebec, Canada).
2.2. Deep Learning Approach for Fluorescence Topography
We investigate a DL model for obtaining fluorescence concentration and depth topography from SFDI images. The Siamese CNN architecture shown in Fig. 2 was adapted from the work done by Smith et al.31 in which DL was used to predict depth to the top surface of a submerged inclusion (“submarine” model) and the corresponding fluorescence lifetime from SFDI-derived optical properties and macroscopic fluorescence lifetime imaging. Our Siamese CNN differs as it uses optical properties and SFDI fluorescence images as separate inputs to predict the depth to the bottom surface of an infiltrative inclusion (“iceberg” model) and the corresponding fluorescence concentration. In addition to changing the model inputs and outputs, we also modified a few of the internal model details to improve performance, most notably changing kernel sizes to for improved deblurring and including residual blocks (ResBlocks) in both arms prior to concatenation.
As shown in Fig. 2, optical properties go through a series of two-dimensional (2D) convolutions with filters, stride, ReLU activation functions, and zero padding to maintain image dimensionality. Before concatenation, the images are fed into a 2D ResBlock in which two additional 2D convolutions are performed with a skip connection in which the input also bypasses the two sets of convolutions. Fluorescence images are passed to a single three-dimensional (3D) convolution with 3 × 3× 6 filters, stride, a ReLU activation function, and zero padding to maintain dimensionality. These images are then passed to two 2D convolutional layers and a 2D ResBlock. As concentration and depth predictions depend on the optical properties and fluorescence images, the fluorescence data are reshaped and concatenated with the optical properties. A separable convolution and a 2D ResBlock are applied to the concatenated data. The concatenated data are split into two arms and a series of 2D convolutions that reduce the number of feature maps while maintaining image dimensionality and are applied to produce the fluorescence concentration map and depth map.
All DL implementation is in Amazon Web Services (AWS) SageMaker with a ml.g4dn.2xlarge instance type (1 NVIDIA T4 GPU, 8 vCPU). During training, 920,514 parameters are learned with Adam as the optimizer and learning rate decay (initial ). Training continued for a maximum of 100 epochs unless the early stopping criteria were met. Training the CNN took approximately 8.6 h at for the in silico datasets described below.
2.3. In Silico Training Data
Figure 3 compares the three shape model iterations used to mimic oral cancer tumors: cylinders, spherical harmonics, and composite spherical harmonics (CSHs). A set of 10,000 of each shape model is generated in MATLAB. Cylinders were chosen as the first shape model as a proof of concept to assess the potential for DL in combination with fluorescence SFDI. Cylinders are generated with randomly chosen widths = 10 to 40 mm and depths = 1 to 10 mm, and to mimic the “iceberg” tumor model, the top surface of the cylinder is set to (i.e., the tissue surface). Spherical harmonics were chosen as the second shape model as they could be quickly generated and introduced more complex geometric features. Spherical harmonics are originally generated with a radius of 10 mm and then scaled to have widths = 10 to 40 mm and depths = 1 to 10 mm. To create perturbations in the shape and diversity in the training set, the order and degree of the spherical harmonic are randomly selected with order = 2 to 20 and degree = order-20. To achieve the “iceberg” model, the section of the spherical harmonic protruding above is cut. CSHs were developed to add asymmetry to the tumor model. The CSH model was intended to have organic features that varied significantly within the dataset. A CSH comprised four spherical harmonics joined as separate quadrants based on two randomly generated reciprocal functions in the plane as shown in Fig. 4. A series of transformations were included when developing the shape model to add regions of buried tumor for the primarily “iceberg” tumor. The CSH is rotated in the , , and directions, shifted in the direction, scaled separately along and axis to have widths = 10 to 40 mm, and scaled in the direction to have depths = 1 to 10 mm. The section of the CSH that is protruding over is cut so that the tumor is flush with the background tissue.
All shape models were passed to an in-house tissue simulator based on a diffusion theory light propagation model in MATLAB to generate in silico images for DL training. Specifically, the simulator was based on a perturbation approach to model the tumor as a single-region heterogeneity using steady-state Green’s functions calculated on a 3D Cartesian grid.34,35 The tissue simulator produced in silico images over a field of view of ( resolution) with Poisson additive noise. A set of reflectance and fluorescence images across spatial frequencies, optical property maps, and true depth and concentration maps were generated for each synthetic tumor shape. The fluorescence and reflectance images were generated for six spatial frequencies using PpIX properties for the fluorophore and a range of optical properties within tissue: absorption to and scattering to . Optical property maps were calculated using SFDI lookup tables with in silico reflectance images at . Fluorophore concentration was considered homogeneous and randomly chosen between 1 and , with an assumed PpIX quantum efficiency ().36 Depth was calculated at each pixel; for areas in which the top surface of the tumor was at/above the healthy tissue (), depth was calculated as the distance between the top surface and the bottom surface; for regions where the top surface of the tumor was buried (), depth was the distance from to the bottom surface of the tumor. The cylinder, spherical harmonic, and CSH models used during in silico testing were trained with fluorescence solely in the tumor. A fourth training set based on CSH was developed for the phantom study and included a randomized amount of fluorescence in the healthy background tissue ranging from 0.1% to 50% of the fluorescence concentration of the tumor. Although the phantoms did not contain fluorescence in the background, this model provided better agreement with the experimental data that are subject to fluorescence filter leakage, camera electronic noise, and SFDI calibration errors not currently included in the training process. These factors led to a mismatch in background fluorescence between experimental and in silico SFDI images.
2.4. In Silico Implementation
To assess the feasibility of DL for predicting depth and concentration from SFDI fluorescence images and optical properties, in silico test sets were developed. A test set of 100 shapes was generated for each respective model (cylinder, spherical harmonic, and CSH). The geometric and optical properties remained within the range of the training data for all three sets. Images were passed to each model to confirm that the networks could make accurate predictions for new cases of the same data type as the training set.
Meshes of patient-derived oral cancer were used to evaluate the ability of the DL training data to capture complex features seen in real tumors. After institutional ethics board approval for retrospective patient data access (University Health Network REB #22-5471), 10 pre-operative magnetic resonance imaging (MRI) of tongue cancer surgical patients were contoured to produce corresponding meshes (average depth = 8.3 mm) as displayed in Fig. 5. Tongue cancer was selected as an initial model for evaluation as these are the most prevalent oral cancer subsite.37 The contoured tumors had widths and depths that fell within the geometric range of the training data. To increase the number of test samples, the 10 contours were scaled in each plane (widths = 10 to 40 mm and depths = 1 to 10 mm) and randomly rotated around the axis ( to 180 deg). These transformations provided a second validation set of 100 patient-derived tumors (average depth = 5.4 mm). As the invasion depth of late-stage oral cavity cancer (e.g., T3–T4 tumors exceeding 1 cm depth) can exceed the penetration depths of near-infrared light, a third evaluation case was included based on an additional patient case (depth = 12.5 mm) to assess DL performance in out-of-range cases. The tumor exceeded the functional depth of our system but had widths that remained within the imaging field of view.
Both test sets were passed to the tissue simulator twice to produce a test set without background fluorescence and a test set with a randomized amount of fluorescence in the background (0.1% to 50% of the fluorescence concentration of the tumor). The sets with background fluorescence were used to assess the in silico performance of the model trained with CSH and background fluorescence. The tissue simulator produced synthetic fluorescence images (PpIX ), reflectance images, optical properties (, ), concentration maps, and depth maps for each test set. The patient tumors protruded above the plane, the depth for each pixel was calculated as the difference between the bottom surface of the tumor, and the top surface, which was either the plane or the top surface of the protruding tumor. Images were passed to each DL model to assess how well the training data could capture the complex geometry of real patient tumors.
2.5. Phantom Implementation
Optical phantoms were used to assess the performance of the DL model when experimentally collected images were passed to a model trained with in silico images. Figure 6 illustrates each stage within the phantom process. The contoured meshes used for in silico validation were 3D printed to create imprints for agar phantoms. These phantoms required two solutions: (i) non-fluorescent agar solution to represent the background tissue and (ii) fluorescent agar solution to represent the tumor. In both the background and tumor agar, India ink ( and ) was the absorbing agent, and intralipid ( and ) was the scatterer. The fluorescent agar solution included PpIX () as the fluorophore and iohexol as a contrast agent for computed tomography (CT) scanning. The background solution was poured into a dish, and the 3D-printed tumors were held at the surface of the solution to create a negative mold. Once the background agar solidified, the 3D-printed tumor was removed, leaving behind a negative imprint for the fluorescent tumor solution. The fluorescent solution was poured into the negative imprint to create a fluorescent tumor.
To model the true geometry, cone beam CT scans (Cios Spin, Siemens Healthineers, Erlangen, Germany) of the phantoms were taken to get the geometry and orientation of the tumor to the background agar. The CT scans were contoured in the 3D-Slicer software,38 to get a 3D mesh that was manually rotated to align with the images taken by the SFDI system. The calibrated mesh from the CT was then passed to our tissue simulator to obtain the ground truth for depth and concentration maps.
The training data for deep learning assumed the background healthy tissue was uniform and spanned the entire imaging field of view. To accommodate this, the background of the phantoms was designed to exceed the field of view. Although the images were originally the correct field of view, subsequent manual cropping was required to remove artifacts in the periphery of the images due non-uniform fluorescence filtering. The SFDI images were collected at a resolution of and had varying dimensions based on the manual crop. If the manual crop created a region smaller than the required field of view, images required padding. To resize the image to have at least a field of view, the images were padded with the values within one standard deviation of the mean background intensity of non-fluorescent agar. As the pixel size at the imaging plane was not an even multiple of the pixel size of the training data ( images at ), an interpolation step was required to convert camera images. Specifically, to obtain images of at a resolution of , the padded images were cropped to at the original camera resolution (), and box interpolation was used to reduce the resolution to . This process was repeated for all reflectance, fluorescence, absorption, and scattering images. Processed images were automatically uploaded to an AWS S3 bucket for DL depth and concentration predictions.
3. Results
3.1. In Silico Results
To confirm that the architecture and parameters of the DL model were suitable, the models were tested with 100 images derived from the same data type as the training data: the cylinder model was tested with cylinders, the spherical harmonic model was tested with spherical harmonics, and the CSH model was tested with CSH. As shown in Fig. 7, each model made reasonable predictions for both depth and concentration. The cylinder model had an average (standard deviation) depth error of 0.36 (0.55) mm and a concentration error of 0.091 . The spherical harmonic model had depth errors of 0.72 (1.04) mm and concentration errors of 0.18 . The CSH model had the highest errors with an average depth error of 0.90 (1.28) mm and an average concentration error of 0.29 .
In silico SFDI fluorescence images and optical properties of the original and augmented patient-derived tongue tumors were passed to all three models. Although the architecture and parameters remained constant across all models, the model predictions differed significantly depending on the training data. Figure 8 illustrates DL predictions across models for a sample from the unscaled in silico set and a sample from the augmented set. Table 1 includes the predictive errors associated with each model when validated with the original patient-derived data (). The model trained with cylinders had the overall highest error when testing with patient-derived tumor meshes. Complex geometric features within the testing set were not captured in the cylinder training set and the model could not accurately make predictions as shown in Fig. 8. The depth and concentration predictions had significant errors. The spherical harmonic model had significantly lower errors in concentration predictions (concentration ) but did not demonstrate a significant improvement in depth predictions (depth ) when compared with the cylinder model. Due to the lack of asymmetry in this training set, many of the predictions had features intrinsic to spherical harmonics: deepest in the center with a gradual radial decrease in depth. The model trained with CSH had significantly lower errors than the cylinder model (depth and concentration ) and spherical harmonic model (depth and concentration ). As the complexity of the shape model increased and the data became more diverse, the model could learn geometric features intrinsic to real tumors.
Table 1.
Original tumors () | Augmented tumors () | |||
---|---|---|---|---|
Depth error (mm) | Concentration error () | Depth error (mm) | Concentration error () | |
Cylinders | 4.2 ± 4.9 | 2.8 ± 2.1 | 1.62 ± 2.5 | 1.9 ± 2.0 |
Spherical harmonics | 3.3 ± 3.6 | 1.78 ± 1.3 | 1.2 ± 2.0 | 0.9 ± 1.1 |
CSH | 1.0 ± 1.5 | 0.5 ± 0.8 | 0.4 ± 0.6 | 0.4 ± 0.6 |
The improvement in DL predictions as training data complexity increased was apparent when the models were validated with the augmented patient-derived data (). The cylinder model had significantly higher errors than both the spherical harmonic model (depth and concentration ) and the CSH model (depth and concentration ). The CSH model continued to have significantly lower error than the spherical harmonic model (depth and concentration ), indicating that the CSH had superior predictive performance when working with patient-derived data. The difference in errors between the two validation sets was likely due to the average depth of the data. As the average depths of the augmented tumors matched the average depths of the training set, the errors were consistently lower in the augmented validation set as shown in Table 1.
The unscaled and augmented tumor sets were also used to assess the effects of adding background fluorescence. The CSH model trained with background fluorescence was validated with in silico images of the unscaled and augmented test sets containing background fluorescence. The respective depth and concentration error for the unscaled set was and , and the augmented set was and . In comparison with the CSH model trained and tested with the same tumor sets without background fluorescence, there was no significant difference between depth and concentration estimates for the unscaled (depth , concentration ) and augmented (depth , concentration ) tumor sets.
Figure 9 illustrates a patient-derived case with maximum depth (12.5 mm) that exceeds the ranges of depths (1 to 10 mm) included in the CSH training model. This case was specifically selected to assess what happens when the DL model encounters “out-of-range” depths (i.e., is there evidence of “AI hallucinations”?), with additional cases required in the future to assess this fully. Here, the depth predictions truncate at and the error increases as depth increases.
3.2. Phantom Results
As the CSH model had the lowest errors in silico, it was used as the primary model for phantom studies. Figure 10 illustrates the predictions from this model on two patient-derived phantoms (maximum depths of 5.70 and 5.25 mm) made from separate batches of agar-based material. Phantoms 1 and 2 had average depth errors of 0.62 and 0.75 mm, respectively, and average concentration errors of 1.84 and . One possible source of error in the concentration map is the choice of PpIX fluorescence quantum efficiency used to train the DL model; here, we used for PpIX,36 but a number of values are reported in the literature.39 Batch-to-batch variability in phantom preparation may also contribute to concentration estimate uncertainties.
4. Discussion and Conclusion
This pre-clinical study demonstrates the potential to use CSH as an in silico oral cancer tumor model to train a DL-enabled SFDI system for fluorescence depth quantification. Simulation studies using 10 patient tongue tumor contours were used for initial validation, as well as to generate an augmented data set of 100 simulated cases for more comprehensive testing. Although initially focused on oral cancer, this work also establishes a generalizable deep learning architecture for SFDI-based fluorescence depth quantification to support future developments for other surgical oncology applications (e.g., breast, brain, and ovary).
Without a large database containing 3D meshes of patient-derived oral tumors for deep learning training, CSHs serve as a promising alternative. Training with patient-derived tumors would be optimal; however, contouring MRI images of oral cancer is time-consuming and prone to errors. In contrast, automatically generating synthetic tumor shapes offers a fast method of creating a diverse training set. Cylinders and spherical harmonics as tumor models were vital in the preliminary stages of developing a shape model as they served as a proof-of-concept for initial studies and led to the development of CSH as tumor surrogates. Although the primary concern associated with such synthetic datasets is that they are unable to capture the complexities of real data, the CSH shape model demonstrated its ability to capture geometric features intrinsic to patient-derived oral tumors in an initial test of 10 subjects.
Although the geometry of patient-derived tumors is captured by the training data, there are multiple limitations associated with this initial study, beyond the limited number of test samples. First, although the CSH model demonstrates its ability to capture the bulk geometries of real tumors, the shape model does not account for the complex infiltrative boundaries of oral tumors. Although in practice oral tumors are also defined on the bulk geometry, large margins are required to account for the infiltrative tumor boundaries. Pathology reveals that these boundaries contain irregular “finger-like” protrusions, satellite tumor nodules, and tumor buds.40 As diffuse optical imaging techniques have a limited ability to capture microscopic features, surgical margins will still be required. Future work should involve integrating small, localized regions of tumor to the shape model to mimic the infiltrative boundaries of real tumors. This addition is also required to assess our system’s limits in terms of fluorophore accumulation and diffuse optical imaging in small nodules. Second, the light propagation model assumed a homogeneous distribution of the fluorophore completely localized to the tumor. This simplified the calculations required for generating the fluorescence images and optical property maps; however, this assumption does not align with in vivo studies as tumors tend to have heterogeneous distributions of the fluorophore. Third, to simplify the training data, we assumed that the tumors were completely flush with the background and that the tissue surface was level. This approximation will not hold for real patient cases as tumors and healthy tissue have some degree of topography. To address this, phase-shifted profilometry will be integrated into the SFDI workflow to correct for irregular surfaces,41 in combination with Monte Carlo simulations of light propagation.32 Further assessment of the optical properties and reproducibility of the phantom recipe is required. Finally, only two phantoms were included here as an initial proof-of-concept demonstration, but future work will include results from additional patient cases and investigate other fluorophores of known quantum efficiencies.
One limitation of our approach is the assumption of a known fluorophore quantum efficiency () to relate fluorescence yield () to concentration (). Although this is feasible in a controlled laboratory setting (e.g., PpIX in a known solvent), in the in vivo setting, it is known that PpIX depends on unknown local environmental factors (e.g., pH, oxygen, and temperature).39 In this case, it may only be possible to recover PpIX concentration estimates within a linear scale factor of the known concentration, as previous authors have suggested for PpIX fluorescence reconstruction.36 Future studies will be required to assess how quantum efficiency variations affect depth and concentration estimates under more realistic conditions, not only for PpIX but potentially for other NIR fluorophores that may exhibit different in vivo behavior.
The exact depth at which DL estimates start to degrade due to the limited penetration of near-infrared light is subject to more extensive evaluation. The simulation and phantom test cases reported here were specific to nominal absorption coefficients ( at 632 nm), showing error for tumors with depth . Additional testing is required across a broader range of conditions, including cases with higher absorption coefficients. Certainly, accurate depth estimation will not be feasible in reflectance mode for all oral cancers, particularly late-stage cases with depths in the centimeters (e.g., T3 and T4).42 Nonetheless, although this paper focused on depth quantification suitable for in vivo imaging of mucosal oral cancer tumors (i.e., “iceberg scenario” of depth to bottom tumor surface), in the future, this system will be applied to quantify margin thickness of “inverted” tumors (i.e., “submarine scenario” of depth to top tumor surface), in which depth quantification is required down to a 5-mm depth to confirm margin status.11 For both in vivo and ex vivo scenarios, future pre-clinical experiments in animal models with complex, heterogeneous optical properties and fluorophore distributions will be essential to validate this technology under more realistic clinical conditions.
Acknowledgments
The work in this paper is an extension of SPIE proceedings.43 This work is supported by the Raymond Ng and Wendy Chui Foundation for Innovation in Otolaryngology–Head and Neck Surgery (Department of Otolaryngology–Head and Neck Surgery, University of Toronto), the Terry Fox Research Institute (TFRI) New Frontiers Program Project (Grant No. 1075), and the Princess Margaret Cancer Foundation (Garron Family GTx Surgery – Engineering Fund, Lucky Bean Foundation GTx STEM Research Scholarship Program). The authors thank Dr. Jason Smith for the helpful discussions on his open-source deep learning code (Rensselaer Polytechnic Institute; PI: Dr. Xavier Intes), as cited, that this work built upon. Thanks also to the undergraduate co-op students whose previous work helped advance this project including Jacqueline Fleisig, Ashley Gilbank, Murtuza Rajkotwala, Yasmeen El-Rayyes, Anjolaoluwa Adewale, Jerry Wan, and Elise Schwarz.
Biography
Biographies of the authors are not available.
Contributor Information
Natalie J. Won, Email: natalie.won@uhn.ca.
Mandolin Bartling, Email: mandolin.bartling@mail.utoronto.ca.
Josephine La Macchia, Email: Josephine.lamacchia@uhn.ca.
Stefanie Markevich, Email: markevis@mcmaster.ca.
Scott Holtshousen, Email: Scott.Holtshousen@uhn.ca.
Arjun Jagota, Email: Arjun.Jagota@uhn.ca.
Christina Negus, Email: Christina.Negus@uhn.ca.
Esmat Najjar, Email: Esmat.Najjar@uhn.ca.
Brian C. Wilson, Email: Brian.wilson@uhnresearch.ca.
Jonathan C. Irish, Email: jonathan.irish@uhn.on.ca.
Michael J. Daly, Email: michael.daly@uhn.ca.
Disclosures
The authors declare no relevant financial or non-financial interests to disclose.
Code and Data Availability
Available upon reasonable request to the corresponding author.
References
- 1.Orosco R. K., et al. , “Positive surgical margins in the 10 most common solid cancers,” Sci. Rep. 8(1), 5686 (2018). 10.1038/s41598-018-23403-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Chen T.-C., et al. , “The impact of pathologic close margin on the survival of patients with early stage oral squamous cell carcinoma,” Oral Oncol. 48(7), 623–628 (2012). 10.1016/j.oraloncology.2012.01.015 [DOI] [PubMed] [Google Scholar]
- 3.Solomon J., et al. , “The impact of close surgical margins on recurrence in oral squamous cell carcinoma,” J. Otolaryngol. Head Neck Surg. 50(1), 9 (2021). 10.1186/s40463-020-00483-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Pagedar N. A., “Better visualization of oral cancer margins—a struggle of cancer and technology,” JAMA Otolaryngol. Head Neck Surg. 146(12), 1156–1157 (2020). 10.1001/jamaoto.2020.3262 [DOI] [PubMed] [Google Scholar]
- 5.Rosenthal E. L., et al. , “The status of contemporary image-guided modalities in oncologic surgery,” Ann. Surg. 261(1), 46–55 (2015). 10.1097/SLA.0000000000000622 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Mehanna H., et al. , “Head and neck cancer--part 2: treatment and prognostic factors,” BMJ 341, c4690 (2010). 10.1136/bmj.c4690 [DOI] [PubMed] [Google Scholar]
- 7.Luryi A. L., et al. , “Positive surgical margins in early stage oral cavity cancer: an analysis of 20,602 cases,” Otolaryngol. Head Neck Surg. 151(6), 984–990 (2014). 10.1177/0194599814551718 [DOI] [PubMed] [Google Scholar]
- 8.Smits R. W., et al. , “Resection margins in oral cancer surgery: room for improvement,” Head Neck 38(Suppl. 1), E2197–E2203 (2016). 10.1002/hed.24075 [DOI] [PubMed] [Google Scholar]
- 9.Vahrmeijer A. L., et al. , “Image-guided cancer surgery using near-infrared fluorescence,” Nat. Rev. Clin. Oncol. 10(9), 507–518 (2013). 10.1038/nrclinonc.2013.123 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Bou-Samra P., et al. , “Intraoperative molecular imaging: 3rd biennial clinical trials update,” J. Biomed. Opt. 28(5), 050901 (2023). 10.1117/1.JBO.28.5.050901 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.de Wit J. G., et al. , “EGFR-targeted fluorescence molecular imaging for intraoperative margin assessment in oral cancer patients: a phase II trial,” Nat. Commun. 14(1), 4952 (2023). 10.1038/s41467-023-40324-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Voskuil F. J., et al. , “Exploiting metabolic acidosis in solid cancers using a tumor-agnostic pH-activatable nanoprobe for fluorescence-guided surgery,” Nat. Commun. 11(1), 3257 (2020). 10.1038/s41467-020-16814-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Keereweer S., et al. , “Optical image-guided cancer surgery: challenges and limitations,” Clin. Cancer Res. 19(14), 3745–3754 (2013). 10.1158/1078-0432.CCR-12-3598 [DOI] [PubMed] [Google Scholar]
- 14.DSouza A. V., et al. , “Review of fluorescence guided surgery systems: identification of key performance capabilities beyond indocyanine green imaging,” J. Biomed. Opt. 21(8), 80901 (2016). 10.1117/1.JBO.21.8.080901 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Pogue B. W., et al. , “Perspective review of what is needed for molecular-specific fluorescence-guided surgery,” J. Biomed. Opt. 23(10), 100601 (2018). 10.1117/1.JBO.23.10.100601 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Brouwer de Koning S. G., et al. , “Assessment of the deep resection margin during oral cancer surgery: a systematic review,” Eur. J. Surg. Oncol. 47(9), 2220–2232 (2021). 10.1016/j.ejso.2021.04.016 [DOI] [PubMed] [Google Scholar]
- 17.van Keulen S., et al. , “The Sentinel margin: intraoperative ex vivo specimen mapping using relative fluorescence intensity,” Clin. Cancer Res. 25(15), 4656–4662 (2019). 10.1158/1078-0432.CCR-19-0319 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Krishnan G., et al. , “Fluorescent molecular imaging can improve intraoperative sentinel margin detection in oral squamous cell carcinoma,” J. Nucl. Med. 63(8), 1162–1168 (2022). 10.2967/jnumed.121.262235 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Rounds C. C., et al. , “Improved intraoperative identification of close margins in oral squamous cell carcinoma resections using a dual aperture fluorescence ratio approach: first in-human results,” J. Biomed. Opt. 29(1), 016003 (2024). 10.1117/1.JBO.29.1.016003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Sibai M., et al. , “Quantitative subsurface spatial frequency-domain fluorescence imaging for enhanced glioma resection,” J. Biophotonics 12, e201800271 (2018). 10.1002/jbio.201800271 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Konecky S. D., et al. , “Spatial frequency domain tomography of protoporphyrin IX fluorescence in preclinical glioma models,” J. Biomed. Opt. 17(5), 056008 (2012). 10.1117/1.JBO.17.5.056008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Kolste K. K., et al. , “Macroscopic optical imaging technique for wide-field estimation of fluorescence depth in optically turbid media for application in brain tumor surgical guidance,” J. Biomed. Opt. 20(2), 26002 (2015). 10.1117/1.JBO.20.2.026002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Jermyn M., et al. , “Macroscopic-imaging technique for subsurface quantification of near-infrared markers during surgery,” J. Biomed. Opt. 20(3), 036014 (2015). 10.1117/1.JBO.20.3.036014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Kim A., et al. , “Topographic mapping of subsurface fluorescent structures in tissue using multiwavelength excitation,” J. Biomed. Opt. 15(6), 066026 (2010). 10.1117/1.3523369 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Vonk J., et al. , “Improving oral cavity cancer diagnosis and treatment with fluorescence molecular imaging,” Oral Dis. 27, 21–26 (2020). 10.1111/odi.13308 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Cuccia D. J., et al. , “Quantitation and mapping of tissue optical properties using modulated imaging,” J. Biomed. Opt. 14(2), 024012 (2009). 10.1117/1.3088140 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Gioux S., Mazhar A., Cuccia D. J., “Spatial frequency domain imaging in 2019: principles, applications, and perspectives,” J. Biomed. Opt. 24(7), 071613 (2019). 10.1117/1.JBO.24.7.071613 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Hayakawa C. K., et al. , “Optical sampling depth in the spatial frequency domain,” J. Biomed. Opt. 24(7), 071603 (2018). 10.1117/1.JBO.24.7.071603 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Smith J. T., et al. , “Deep learning in macroscopic diffuse optical imaging,” J. Biomed. Opt. 27(2), 020901 (2022). 10.1117/1.JBO.27.2.020901 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Tian L., et al. , “Deep learning in biomedical optics,” Lasers Surg. Med. 53(6), 748–775 (2021). 10.1002/lsm.23414 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Smith J. T., et al. , “Macroscopic fluorescence lifetime topography enhanced via spatial frequency domain imaging,” Opt. Lett. 45(15), 4232–4235 (2020). 10.1364/OL.397605 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Nizam N. I., et al. , “Monte Carlo-based data generation for efficient deep learning reconstruction of macroscopic diffuse optical tomography and topography applications,” J. Biomed. Opt. 27(8), 083016 (2022). 10.1117/1.JBO.27.8.083016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Angelo J., et al. , “Ultrafast optical property map generation using lookup tables,” J. Biomed. Opt. 21(11), 110501 (2016). 10.1117/1.JBO.21.11.110501 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Jacques S. L., Pogue B. W., “Tutorial on diffuse light transport,” J. Biomed. Opt. 13(4), 041302 (2008). 10.1117/1.2967535 [DOI] [PubMed] [Google Scholar]
- 35.Ostermeyer M. R., Jacques S. L., “Perturbation theory for diffuse light transport in complex biological tissues,” J. Opt. Soc. Am. 14(1), 255–261 (1997). 10.1364/JOSAA.14.000255 [DOI] [PubMed] [Google Scholar]
- 36.Flynn B. P., et al. , “White light-informed optical properties improve ultrasound-guided fluorescence tomography of photoactive protoporphyrin IX,” J. Biomed. Opt. 18(4), 046008 (2013). 10.1117/1.JBO.18.4.046008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Justesen M. M., et al. , “Impact of tumor subsite on survival outcomes in oral squamous cell carcinoma: a retrospective cohort study from 2000 to 2019,” Oral Oncol. 149, 106684 (2024). 10.1016/j.oraloncology.2024.106684 [DOI] [PubMed] [Google Scholar]
- 38.Kikinis R., Pieper S. D., Vosburgh K. G., “3D slicer: a platform for subject-specific image analysis, visualization, and clinical support,” in Intraoperative Imaging and Image-Guided Therapy, Jolesz F. A., Ed., pp. 277–289, Springer, New York: (2014). [Google Scholar]
- 39.Myrzakhmetov B., et al. , “Photophysical properties of protoporphyrin IX, pyropheophorbide-a and photofrin® in different conditions,” Pharmaceuticals 14(2), 138 (2021). 10.3390/ph14020138 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Jang J. Y., Choi N., Jeong H.-S., “Surgical extent for oral cancer: emphasis on a cut-off value for the resection margin status: a narrative literature review,” Cancers 14(22), 5702 (2022). 10.3390/cancers14225702 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Gioux S., et al. , “Three-dimensional surface profile intensity correction for spatially modulated imaging,” J. Biomed. Opt. 14(3), 034045 (2009). 10.1117/1.3156840 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Lydiatt W. M., et al. , “Head and neck cancers-major changes in the American Joint Committee on Cancer eighth edition cancer staging manual,” CA Cancer J. Clin. 67(2), 122–137 (2017). 10.3322/caac.21389 [DOI] [PubMed] [Google Scholar]
- 43.Won N. J., et al. , “Deep learning architectures for spatial-frequency 3D fluorescence in oral cancer surgery models,” Proc. SPIE 12825, 128250D (2024). 10.1117/12.3002880 [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Available upon reasonable request to the corresponding author.