Abstract
Quantitative photoacoustic tomography (QPAT) is a valuable tool in characterizing ovarian lesions for accurate diagnosis. However, accurately reconstructing a lesion’s optical absorption distributions from photoacoustic signals measured with multiple wavelengths is challenging because it involves an ill-posed inverse problem with three unknowns: the Grüneisen parameter , the absorption distribution, and the optical fluence . Here, we propose a novel ultrasound-enhanced Unet model (US-Unet) that reconstructs optical absorption distribution from PAT data. A pre-trained ResNet-18 extracts the US features typically identified as morphologies of suspicious ovarian lesions, and a Unet is implemented to reconstruct optical absorption coefficient maps, using the initial pressure and US features extracted by ResNet-18. To test this US-Unet model, we calculated the blood oxygenation saturation values and total hemoglobin concentrations from 655 regions of interest (ROIs) (421 benign, 200 malignant, and 34 borderline ROIs) obtained from clinical images of 35 patients with ovarian/adnexal lesions. A logistic regression model was used to compute the ROC, the area under the ROC curve (AUC) was 0.94, and the accuracy was 0.89. To the best of our knowledge, this is the first study to reconstruct quantitative PAT with PA signals and US-based structural features.
Keywords: Quantitative photoacoustic tomography, Blood oxygenation saturation, Unet, Ovarian cancer
1. Introduction
Photoacoustic imaging (PAI) is an emerging imaging modality for non-invasive, non-ionizing, real-time measurement of the optical properties of biological tissue [1]. Compared with other optical imaging modalities, PAI can image deeper because acoustic scattering in tissue is an order of magnitude smaller than that of optical scattering [2]. Additionally, PAI inherently possesses ultrasound resolution because received photoacoustic waves are used to form images [3], [4], [5]. Photoacoustic tomography (PAT) uses a broad laser beam for illumination and an array of ultrasound transducers to measure the photoacoustic waves generated by the targeted biological tissue [6]. Imaging is typically done either by standard delay-and-sum beamforming if a linear array is used as a receiver or by tomographic reconstruction if a circular receiver array is employed [7]. For clinical applications, linear array transducers in the frequency range of 3–10 MHz are often used because they can easily access an organ through a limited biological window and are widely available at a low cost [8], [9]. In this transducer frequency range, the depth of penetration can reach several centimeters [10]. Oncology applications of PAT include imaging and diagnosis of breast cancer [11], [12], ovarian cancer [13], [14], cervical cancer [15], and thyroid cancer [16].
The PA effect is induced by short laser pulses which generate initial pressure rises inside the tissue. The generated initial pressure () in the tissue is proportional to Grüneisen parameter , target absorption coefficient and light fluence [17], [18], [19], [20]. It is a technical challenge to quantitatively recover or reconstruct the optical absorption coefficient using PAT data because the received pressure is the product of tissue absorption and light fluence. Quantitative PAT approaches involve an ill-posed, non-unique, and non-linear inverse problem which is difficult to solve reliably for clinical applications [21], [22]. Moreover, PAT data can be noisy and background tissue absorption can produce artifacts, both of which make the reconstruction more challenging [23].
Blood oxygenation saturation (%sO2) and total hemoglobin concentration (HbT) are functional biomarkers of the malignancy of ovarian tissue [24], [25], [26]. A %sO2 map can be calculated point by point from the ratio of the estimated relative oxygenated hemoglobin to the relative total hemoglobin concentration, assuming the unknown laser fluence at each imaging point at all wavelengths used is a constant and can be canceled out. However, the accuracy of %sO2 depends on the PAT data quality received at all wavelengths.
There have been several approaches to reconstructing photoacoustic images, for example, iterative methods [27], [28], and machine learning-based approaches [29], [30], [31], [32]. Cai used an end-to-end machine learning model to achieve fast reconstruction [33]. Luke et al. proposed a double Unet (named ‘O-net’) to simultaneously estimate the oxygen saturation in blood vessels and segment the vessels from the surrounding background tissue [34], and Bench achieved 3D %sO2 reconstruction of blood vessels under the skin by a 3D fully convolutional network [35].
However, these studies about machine learning-based PAT reconstruction methods were only using photoacoustic signals as input. Here, based on the knowledge of ovarian lesions, structural morphologies seen by US images were characteristics to classify benignancy or malignancy [41], [42], [43] and can be incorporated into the PAT U-net to improve the diagnostic accuracy of ovarian lesions.
Machine learning models, such as the ResNet model and Unet model, have demonstrated the ability to classify or reconstruct biomedical images. ResNet-18 is one of the most popular machine learning models used in image classification [36], [37]. The developers of ResNet-18 pre-trained it on the ImageNet dataset [38]. Transfer learning can be adapted to train a ResNet-18 model for ovarian lesion feature extraction, because ImageNet objects such as plates (round shapes), stones (masses), and flags (edges) have features similar to ovarian lesions, such as nodules (round shapes) and solid masses with edges. The Unet model is suitable for PAT reconstruction because it can maintain PA high resolution by using its skip-connection feature to copy the input to output. In contrast to Unet, CNN models are less suitable for PAT reconstruction because they use pooling layers that downgrade the resolution [39], [40].
In this manuscript, based on the observed characteristics of ovarian lesions in US images, structural morphologies seen in US images are extracted by a ResNet-18 model. These US features are incorporated into a Unet model structure to perform quantitative PAT reconstruction of absorption coefficient maps and then the calculation of total hemoglobin and %so2, referred to as US enhanced Unet (US-Unet). To the best of our knowledge, this is the first report of co-registered US being used to enhance the machine learning performance of PAT reconstruction. Both phantom experiments and clinical studies showed the superiority of the proposed method.
2. Methodology and materials
2.1. Ultrasound-enhanced Unet model
To extract ovarian tissue morphology features, a pre-trained ResNet-18 model was implemented, as shown in Fig. 1B. Ultrasound images of ovarian lesions were used to fine-tune the ResNet-18. Since we only had 1200 ultrasound images after augmentation, we chose to use a pre-trained ResNet-18 to avoid training from scratch with random-initial weights. Three cross-validations were used to evaluate the performance of the model.
To reconstruct PAT images, we designed a US-enhanced Unet model whose structure is shown in Fig. 1. Fig. 1A shows the Unet model used to reconstruct the optical absorption coefficient maps. Fig. 1B shows the ResNet-18 model implemented to extract features from US images. To enhance the Unet model with extracted morphological features, we implemented the ResNet-18 model with 18 convolutional layers. After the final convolutional layer, we changed the fully connected layer to 256 neurons, which was first resized to 16 × 16, then concatenated to a bottleneck block in the Unet model (as indicated by the dashed red arrow in Fig. 1.) [45]. Thus, the US-enhanced Unet utilizes the US image features extracted by ResNet and the initial pressure calculated from our co-registered US and PAT system [24], to quantitatively reconstruct optical absorption coefficient maps.
The Unet model was trained using simulation and phantom data to learn the PAT reconstruction process, implementing the mean square error loss function. The ResNet-18 model was trained with clinical US images for classification, utilizing binary cross entropy loss. After training for 100 epochs, the ResNet-18 model could extract morphology features from US images well and classify lesions accurately. The features extracted from the ResNet-18 model were incorporated into the Unet model to perform PAT reconstruction. Thus, the US-enhanced Unet model could reconstruct PAT images with ovarian morphology features incorporated.
The model was implemented in Pytorch and trained for a total of 200 epochs on an Nvidia 2080 Ti GPU, using the ADAM optimizer with a learning rate of 1e-4. A batch size of 256 images was used, and the training process was monitored for early stopping. To speed up training, the initial pressure maps and US images were down sampled from 700 × 1400–128 × 128. We used 2000 simulation data and 480 phantom data for training, 540 blood tube datasets for validation, and 621 clinical datasets for testing.
2.2. %sO2 and total hemoglobin concentration (HbT) calculation
The ultrasound-enhanced Unet model reconstructed the optical absorption coefficient from raw data collected by the transducer. Then, using Eq. (1), we could compute the molar concentrations of oxy- and deoxy- hemoglobin [46]:
(1) |
Here, represents the wavelength. Further, denotes the optical absorption coefficient, and and are the molar extinction coefficients of oxy- and deoxy-hemoglobin, respectively. and are the molar concentrations of oxy- and deoxy-hemoglobin. For spectrum unmixing, we used the non-negative least square method.
The HbT and the oxygen saturation (%sO2) can be computed from Eq. (2):
(2) |
2.3. Co-registered US and PAT system
To collect initial pressure in this study, we used the real-time co-registered PAT and US system demonstrated in Ref. [47]. The co-registered PAT system has three parts, 1) a tunable Ti:sapphire laser (700–900 nm) pumped by a Q-switched Nd: YAG laser (Symphotics, Camarillo, CA), which generates 10 ns laser pulses at a 15 Hz repetition rate; 2) an optical fiber-based light delivery system [48]; and 3) a commercial US system (EC-12R, Alpinion Medical Systems, Republic of Korea). In this study, the system acquired PAT data at four wavelengths (730, 780, 800, 830 nm) [49]. In principle, two optical wavelengths are needed to compute and , and ideally, these two wavelengths are on each side of isosbestic wavelength at ∼800 nm. However, two wavelengths of 730 nm and 780 nm give a more robust estimate , and the use of 800 nm and 830 nm give a more robust estimate of. Also, these four wavelengths are close to one other, and so the change of wavelength-dependent fluence is minimal for spectral unmixing.
2.4. Simulations and phantoms (Unet training)
Simulation data were used to train the U-net model. To generate the simulation data, we defined digital phantoms with a homogenous background and targets with different shapes and optical properties (absorption and scattering coefficients). The optical fluence inside each digital phantom was calculated using Monte Carlo simulation [50]; the maximum depth was 6 cm, the grid and voxel sizes were 0.1 × 0.1 × 0.1 cm, and the total number of photons used was 1e7. We then computed the initial pressure value at each pixel of the phantom by multiplying the Grüneisen parameter at that pixel by its absorption coefficient and optical fluence. This initial pressure map was then used as the input of the k-wave toolbox [51], which generated the raw photoacoustic signal at each transducer element. The transducer configuration was the same as our co-registered US and PAT system. The simulation sampling frequency was 41 MHz, and input grid size was 700 × 1400, which was 6cmx12cm in real size. Finally, the delay-and-sum algorithm was employed to reconstruct the initial pressure from the raw data, here the Grüneisen parameter was assumed to be 1 [52]. When training the US-Unet model, we used the reconstructed initial pressure map as the input and the optical absorption coefficient map as the output.
Finally, the delay-and-sum algorithm was employed to reconstruct the initial pressure from the raw data. In simulations, the Grüneisen parameter was assumed as 1 [52]. In phantom studies, the Grüneisen parameter was assumed to be constant and the light fluence profiles (ϕ (r)) were assumed to be wavelength independent. When training the US-Unet model, we used the reconstructed initial pressure map as the input and the optical absorption coefficient map as the output.
We simulated the initial pressure with two different target shapes (a sphere and an ellipsoid), 25 different target absorption coefficients (ranging from 0.1 cm−1 to 4.9 cm−1, with a step of 0.2 cm−1), three different target depths (2.5, 3, and 3.5 cm), three different radii (1, 1.5, and 2 cm), and four different wavelengths (730, 780, 800, and 830 nm). In total, 1800 simulation data were generated.
Five ovary-mimicking gelatin-based phantoms were made, all with the same amount of gelatin, diluted intralipid (to provide the desired optical scattering), and glass beads (to enhance ultrasound scattering). Different amounts of ink were added to each phantom to achieve absorption coefficients of 0.2, 0.4, 0.6, 0.8, and 1.0 . After being refrigerated for 24 h, each of the five solidified phantoms was placed by turns at three different imaging depths (1 cm, 1.5 cm, and 2 cm) in an intralipid solution with calibrated absorption and reduced scattering coefficients of 0.02 and 6 cm−1, similar to the optical properties of soft tissue, and PAT data were acquired at each depth. Two sets of experiments were done for each phantom at each depth, and each measurement contained 4 frames, which generated 480 data sets at the previously mentioned four wavelengths, three depths, and five absorption coefficients.
Besides gelatin phantoms, we made blood tubes to mimic the blood vessels in tissues with different %sO2 contrast. Five calibrated %sO2 blood samples of 24.9 %, 44.2 %, 64.9 %, 83.9 %, and 97.6 % were made by controlling the amount of oxygen and nitrogen in the tube. The blood within the tube was collected from volunteers and mixed with saline water in a temperature/humidity-controlled chamber. The blood tube %sO2 values were calibrated with an ABL90 FLEX Radiometer. More details can be found in Ref. 24. These blood tubes were placed at 9 different depths (1 cm, 1.5 cm, 2 cm, 2.5 cm, 3 cm, 3.5 cm, 4 cm, 4.5 cm, and 5 cm from the transducer) in a tank filled with intralipid solution. The experiments were repeated three times, generating 540 blood tube datasets in total.
2.5. Patients and ovarian lesion data
The study protocol was approved by the institutional review board of the Washington University Medical School and was compliant with the Health Insurance Portability and Accountability Act. A total of 35 patients with ovarian/ adnexal masses signed the informed consent and participated in this study from February 2017 to November 2018 [24]. The patients’ ages ranged from 33 to 87, and the ovarian lesions’ diameters ranged from 1.8 cm to 11.5 cm. The final data set consisted of 10 malignant ovarian/adnexal masses, 36 benign masses, and normal ovaries, and 3 borderline tumors of low malignancy potential. A total of 655 regions of interest (ROIs) (421 benign, 200 malignant, and 34 borderline tumor ROSs) were selected and used in this study (Table 1).
Table 1.
Malignant | Benign | Borderline | |
---|---|---|---|
Number of Patients | 6 | 26 | 3 |
Number of Ovarian Lesions | 10 | 36 | 3 |
Number of Data Sets | 200 | 421 | 34 |
Ultrasound images from a commercial US machine were collected from the same group of patients. We selected 600 ovarian lesion images, and after augmentation (rotation and flipping), the resulting 1200 ultrasound images were used in ResNet-18 fine-tune training and validation.
2.6. Ovarian cancer morphologies revealed in US
As shown in Fig. 2, there were five typical morphologies in the malignant ovarian cancers characterized by US images [53]. Some ovarian cancers appeared as solid papillary projections from the wall of cystic lesions (Fig. 2A and B). Other ovarian cancers had significant blood content, as shown by Doppler (Fig. 2C). Still other ovarian cancers appeared as solid lesions with irregular contours (Fig. 2D and E).
Compared with benign lesions, malignant lesions often have more numerous solid nodules growing from the cyst wall, more irregular solid components both inside and/or at the lesion periphery, and more blood vessels revealed by Doppler US. Although these features are predictive of malignancy, however, they are not sensitive and specific [54]. Here, we sought to improve overall ovarian cancer diagnosis by using additional functional information obtained from co-registered photoacoustic imaging, such as the %sO2 contrast and the quantitative hemoglobin concentration.
3. Results and analysis
3.1. Phantom results
As shown in Fig. 3, the reconstructed µa was calculated for the phantom experiments, each box represents the reconstructed µa distribution from the proposed method, and the dashed line is the ground truth. The mean square estimated error for the phantoms, with absorption coefficients of 0.2, 0.4, 0.6, 0.8, and 1.0 cm-1, were 0.018, 0.011, 0.019, 0.011, and 0.013 cm-1, respectively.
3.2. Blood tube validation results
After we trained the Unet model with simulation and phantom data, blood tube data was used to validate the performance of the US-Unet model. Blood tube reconstruction examples are shown in Fig. 4, Fig. 5. Both the calculated %sO2 and HbT maps were in the similar range of calibrated values, and the shapes were close to the tube shape.
All the blood tube reconstructions were aggregated for quantitative display. As shown in Fig. 6, each box represents the calculated mean %sO2 for a tube at a different depth, from 1 cm to 5 cm, in steps of 0.5 cm. The dashed line is the ground truth based on calibration.
In our method, the machine learning model was trained with the ground truth, and the blood tubes were homogenous. Therefore, the model could reconstruct the real shape and size of the targets. In other words, the machine learning model can mitigate boundary buildup due to interference [55]. Besides, we downsampled the reconstruction images, which also could yield a more homogenous result.
The normalized mean square estimated error for blood tubes with %sO2 of 24.9 %, 44.2 %, 64.9 %, 83.9 %, and 97.6 % were 2.9 %, 3.5 %, 2.7 %, 2.6 %, and 3.0 %, respectively. Compared with standard delay-and-sum results of 5.6 %, 3.8 %, 9.8 %, 8.7 % and 17.7 %, respectively [24], as shown in Fig. 6, the errors were much smaller. Fig. 7 shows the Unet reconstructed quantitative HbT data of blood tube. The calibrated values were given in black circles.
3.3. ResNet-18 results
Three cross-validations were used to test the accuracy of the ResNet-18 model. The accuracies were 0.80 ± 0.02, 0.76 ± 0.04, and 0.81 ± 0.04, which suggests that about 80 % accuracy can be achieved with morphology features only.
We implemented Grad-CAM (gradient-weighted class activation mapping) to learn what features the three ResNet-18 models had learned. Grad-CAM uses the “gradients of any target” concept, back-flowing into the final convolutional layer to produce a coarse localization map highlighting important regions in the image for predicting the diagnostic result [56], [57]. In other words, if we have a trained model, we can use specific data and its label to perform back propagation to find the regions the model is “looking at” to make its prediction.
From Fig. 8, we can see the ResNet-18 model learned the “nodules” and boundaries (hot spots) that were essential to benign or malignant lesion classification based on ultrasound images. The results also suggest that, although morphological features could distinguish between benign and malignant lesions, they are not specific or sensitive (moderate accuracy). So, to improve the diagnostic results, these features extracted from ResNet-18 could be incorporated into the US-Unet model.
3.4. Clinical results
For clinical tests, total hemoglobin and %sO2 are important biomarkers in analyzing ovarian lesions. Total hemoglobin and %sO2 were calculated from the optical absorption coefficient maps of the US-enhanced Unet model output.
Fig. 9 shows the %sO2 and the HbT results for a malignant and a benign lesion. In Fig. 9A, the mean %sO2 of the ROI is 30.3 %, which is much lower than the mean %sO2 of 80.6 % in the benign lesion shown in Fig. 9B. Besides the mean %sO2 value, we could also see peripheral %sO2 distribution around the ultrasonically seen lesion. And according to mean HbT value, the malignant case (377 μM) is 1.5 times higher than the benign case (255 μM).
value of boxes.
To test the performance of the US-Unet model, three reconstruction methods were implemented: US-Unet, delay-and-sum reconstruction, and Unet reconstruction. As seen in Fig. 10, all three methods could distinguish benign from malignant lesions, with contrasts of 1.68, 1.33, 1.44, respectively. However, the US-Unet model was superior to the other two methods in contrast. As for contrast between borderline tumors and benign lesions, the delay-and-sum is better than that US-Unet and Unet alone, however, due to the smaller sample size of borderline tumors, we could not make any conclusion at this point.
Fig. 11 shows the reconstructed quantitative total hemoglobin concentrations using the US-Unet (A), delay-and-sum (B), and Unet (C). For US-Unet, the median HbT of the malignant cases is 2.15 times higher than that of the benign cases. For delay-and-sum beamforming, the relative HbT is 1.51 times higher. For Unet, it is 1.71 times higher. The HbT contrast between borderline tumors and benign lesions is different for the delay-and-sum than that of US-Unet and Unet alone. Again, due to the smaller sample size of borderline tumors, we could not make any conclusion at this point. More data are needed.
Logistic regression was implemented to evaluate the diagnostic performance of %sO2 and HbT data and %sO2 data only. The %sO2 values, with and without HbT testing data from the output of the US-Unet model, with the input of PAT data and US images, were used to calculate the receiver operating characteristic (ROC) curves and the area under the curve (AUC) (See Fig. 12). Accuracies were calculated for the US-enhanced Unet model, the delay-and-sum model, and the Unet model (Table 2). Here, we distinguished benign and malignant cases only, because the three borderline cases are not large enough for a category to be classified. The results show the superiority of US-enhanced Unet with %sO2 and HbT data.
Table 2.
US-enhaced Unet |
Unet |
Dealy-and-Sum |
||||
---|---|---|---|---|---|---|
accuracy | AUC | accuracy | AUC | accuracy | AUC | |
%sO2 and HbT | 0.89 | 0.94 | 0.62 | 0.68 | 0.71 | 0.82 |
%sO2 | 0.80 | 0.89 | 0.61 | 0.66 | 0.66 | 0.72 |
4. Summary and discussion
In this study, a novel ultrasound-enhanced Unet model was proposed to reconstruct the optical coefficient maps of ovarian/adnexal lesions from photoacoustic data. The experimental results showed that the ultrasound-enhanced Unet model reconstructed the target’s %sO2 and HbT maps more accurately than the delay-and-sum method and provided an improved diagnosis of ovarian cancers from benign lesions.
Ovarian lesion morphologies from US images were predictive of malignancy, therefore, a ResNet-18 was implemented to extract relative features from US images. Grad-CAM demonstrated that the ResNet-18 classification model focused on a pertinent area to make its predictions, which also suggested that the model had extracted related features useful for classification. Compared with both a Unet model without ultrasound extracted features and with the standard delay-and-sum method, the US-enhanced Unet demonstrated superior diagnostic contrast and also achieved diagnostic performance of AUC of 0.94 and accuracy of 0.90. Classification results using reconstructions from the Unet model underperformed, even compared with the delay-and-sum method. Without ovarian morphology features, the Unet model could learn only from simulation and phantom data. There is a big difference between clinical and simulation/phantom data, because ovarian tissue is heterogeneous, with complex shapes. This factor could lead to an inferior prediction if using the Unet model only. The delay-and-sum beamforming method directly reconstructed the relative HbT from PAT raw data, with no learning process, and the errors resulted from the tissue heterogeneity, and the inaccurate assumption that the light fluence was wavelength independent.
The US-enhanced Unet has several limitations. The US-enhanced Unet model was limited by low resolution and small reconstructed image size, due to down sampling. However, using a large size of initial pressure will increase the training time exponentially. Additionally, although we reconstructed the quantitative oxy- and deoxy-hemoglobin concentrations, there were outliers in the reconstructions due to ill-posed problems and the diagnostic performance of HbT only was lower than %sO2. Finally, the absorption coefficients used in simulations and gelatin phantoms in training are our best estimates because we found very limited literature reporting ovarian tissue absorption coefficients. One study, Ref. 44, used diffuse optical tomography to image excised ovarian lesions. Among 24 samples measured at 780 nm, 6 malignant samples had a mean absorption coefficient of 0.153 cm−1, and benign samples had a mean coefficient of 0.093 cm−1. However, 4 out of 6 malignant samples were pieces of large malignant lesions, and there was significant blood loss from the samples before imaging. Additionally, the diffuse optical tomography underestimated the target absorption because of the linear assumption of the target perturbation and target absorption using Born approximation. Thus, we believed the ovarian tissue optical absorption coefficients reported in Ref 44 were underestimated. In another ex vivo ovarian specimen study using spatial frequency domain imaging, the measured absorption coefficient of ovarian tissue ranged from 0.1 cm−1 to 0.7 cm−1 [58]. We also measured the absorption coefficient for the blood tubes used by spectrometry and found the absorption coefficients for the four wavelengths ranged from 1.11 cm−1 to 2.06 cm−1, more details in Appendix B. From the literature, the optical absorption coefficient of blood is in the range of 1–5 cm−1 [59], [60]. Because malignant ovarian lesions consist of mixed blood vessels and stromal tissues, we used estimated absorption coefficients in the range of 0.2 − 1 cm−1 for gelatin phantoms and 0.1–4.9 cm−1 in simulations.
The %sO2 and HbT data shown in Fig. 10, Fig. 11 reveal that delay-and-sum beamforming performed differently from US-Unet and Unet for three borderline tumor cases. Delay-and-sum showed borderline tumors had lower %sO2 and lower relative HbT than benign lesions, while the US-enhanced Unet and Unet both showed that the %sO2 and HbT values of borderline cases were similar to those of benign lesions. Borderline ovarian lesions are slow-growing tumors with low malignant potential. The tumor angiogenesis development could be slow, and the lesion morphology could overlap between large benign and malignant lesions. There is no reported %sO2 data to support the suggestion that borderline ovarian/adnexal lesions should have the same or lower values than benign lesions. Nevertheless, it is likely that Unet and US-enhanced Unet did not see enough of these cases to robustly predict %sO2 and HbT. It is also likely that delay-and-sum was not sufficiently robust to estimate %sO2 when PAT signals were low. More borderline ovarian/adnexal lesions will be acquired from the ongoing trial to further evaluate the performance of US-Unet and beamforming algorithms in predicting borderline ovarian/adnexal lesions.
For clinical data, the quality of reconstruction was highly affected by the system’s SNR. For low SNR data, the reconstruction will be less robust. Future work will be focused on building a larger database from phantoms to enhance training and reduce the effect of a low SNR. Currently, a clinical trial is on-going with the goal of acquiring US and PAT data of diverse ovarian/adnexal lesions of malignancies of various types, benign solid and cystic lesions, and mixed solid and cystic lesions, as well as normal ovaries. This large in vivo database will be used to further validate the initial finding that US-enhanced Unet has the superior diagnostic performance to standard delay-and-sum beamforming.
Declaration of Competing Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Acknowledgments
This work was supported by National Institute of Health (R01 CA237664 and R01 CA228047). We thank the entire GYN oncology group for helping identify patients and the Radiology team for helping with US scans. Finally, we thank Professor James Ballard for manuscript editing.
Biographies
Yun Zou received B.S. and M.S. in Engineering Physics from Tsinghua University. He is a PhD candidate in Biomedical Engineering Department of Washington University in St. Louis. His research interests are diffuse optical tomography and photoacoustic tomography with novel machine learning methods. He is also interested in classification and detection using Ultrasound images for cancer diagnosis.
Eghbal Amidi received his BS and MS in Electrical engineering. In 2021, He graduated with his Ph.D. in Biomedical Engineering from Washington University in St. Louis. He is currently working as a Data Scientist at Caris Life Sciences, Irving, TX. His research interests include computer vision, image processing, and machine learning in radiology and histology images.
Hongbo Luo received his B.S. in Opto-Information Science and Technology from Huazhong University of Science and Technology, China and M.S. in Optical Engineering from the same university. He is currently pursuing his PhD in Electrical and System Engineering at Washington University in St. Louis, USA. His research interest includes Optical Coherence Tomography and Photoacoustic imaging.
Quing Zhu is the Edwin H. Murty Professor of Biomedical Engineering, Washington University in St. Louis. She is also an associate faculty in Radiology, Washington University in St. Louis. Professor Zhu has been named a Fellow of OSA, a Fellow of SPIE, a Fellow of AIMBE. Her research interests include multi-modality ultrasound, diffuse light, photoacoustic imaging, and Optical Coherence Tomography for breast, ovarian, colorectal cancer applications.
Footnotes
Supplementary data associated with this article can be found in the online version at doi:10.1016/j.pacs.2022.100420.
Appendix A. Supplementary material
.
Data Availability
Data will be made available on request.
References
- 1.Wang L.V., Gao L. Photoacoustic microscopy and computed tomography: from bench to bedside. Annu. Rev. Biomed. Eng. 2014;16:155–185. doi: 10.1146/annurev-bioeng-071813-104553. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Beard P. Biomedical photoacoustic imaging. Interface Focus. 2011;1.4:602–631. doi: 10.1098/rsfs.2011.0028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Attia A.B.E., Balasundaram G., Moothanchery M., Dinish U.S., Bi R., Ntziachristos V., Olivo M. A review of clinical photoacoustic imaging: current and future trends. Photoacoustics. 2019;16 doi: 10.1016/j.pacs.2019.100144. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Xia J., Yao J., Wang L.V. Photoacoustic tomography: principles and advances. Electromagn. Waves. 2014;147:1. doi: 10.2528/PIER14032303. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Jeon S., Kim J., Lee D., Baik J.W., Kim C. Review on practical photoacoustic microscopy. Photoacoustics. 2019;15 doi: 10.1016/j.pacs.2019.100141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Zhou Y., Yao J., Wang L.V. Tutorial on photoacoustic tomography. J. Biomed. Opt. 2016;21.6 doi: 10.1117/1.JBO.21.6.061007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Manwar R., Kratiewicz K., Avanaki K. Overview of ultrasound detection technologies for photoacoustic imaging. Micromachines. 2020;11.7:692. doi: 10.3390/mi11070692. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Szabo T.L., Lewin P.A. Ultrasound transducer selection in clinical imaging practice. J. Ultrasound Med. 2013;32.4:573–582. doi: 10.7863/jum.2013.32.4.573. [DOI] [PubMed] [Google Scholar]
- 9.Montilla L.G., Olafsson R., Bauer D.R., Witte R.S. Real-time photoacoustic and ultrasound imaging: a simple solution for clinical ultrasound systems with linear arrays. Phys. Med. Biol. 2012;58.1:N1. doi: 10.1088/0031-9155/58/1/N1. [DOI] [PubMed] [Google Scholar]
- 10.Wang L.V. Tutorial on photoacoustic microscopy and computed tomography. IEEE J. Sel. Top. Quantum Electron. 2008;14.1:171–179. doi: 10.1109/JSTQE.2007.913398. [DOI] [Google Scholar]
- 11.Li X., Heldermon C.D., Yao L., Xi L., Jiang H. High resolution functional photoacoustic tomography of breast cancer. Med. Phys. 2015;42.9:5321–5328. doi: 10.1118/1.4928598. [DOI] [PubMed] [Google Scholar]
- 12.Xi L., Li X., Yap L., Grobmyer S., Jiang H. Design and evaluation of a hybrid photoacoustic tomography and diffuse optical tomography system for breast cancer detection. Med. Phys. 2012;39.5:2584–2594. doi: 10.1118/1.3703598. [DOI] [PubMed] [Google Scholar]
- 13.Nandy S., Mostafa A., Hagemann I.S., Powell M.A., Amidi E., Robinson K., Mutch D.G., Siegel C., Zhu Q. Evaluation of ovarian cancer: initial application of coregistered photoacoustic tomography and US. Radiology. 2018;289.3:740–747. doi: 10.1148/radiol.2018180666. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Salehi H.S., Li H., Merkulov A., Kumavor P.D., Vavadi H., Sanders M., Kueck A., Brewer M.A., Zhu Q. Coregistered photoacoustic and ultrasound imaging and classification of ovarian cancer: ex vivo and in vivo studies. J. Biomed. Opt. 2016;21.4 doi: 10.1117/1.JBO.21.4.046006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Valluru K.S., Willmann J.K. Clinical photoacoustic imaging of cancer. Ultrasonography. 2016;35.4:267. doi: 10.14366/usg.16035. https://doi.org/10.14366.16035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Dogra V.S., Chinni B.K., Valluru K.S., Moalem J., Giampoli E.J., Evans K., Rao N.A. Preliminary results of ex vivo multispectral photoacoustic imaging in the management of thyroid cancer. Am. J. Roentgenol. 2014;202.6:W552–W558. doi: 10.2214/AJR.13.11433. [DOI] [PubMed] [Google Scholar]
- 17.Xu M., Wang L.V. Photoacoustic imaging in biomedicine. Rev. Sci. Instrum. 2006;77(4) doi: 10.1063/1.2195024. [DOI] [Google Scholar]
- 18.Cox B.T., Laufer J.G., Beard P.C., Arridge S.R. Quantitative spectroscopic photoacoustic imaging: a review. J. Biomed. Opt. 2012;17.6 doi: 10.1117/1.JBO.17.6.061202. [DOI] [PubMed] [Google Scholar]
- 19.Luriakose M., Borden M.A. Microbubbles and nanodrops for photoacoustic tomography. Curr. Opin. Colloid Interface Sci. 2021;55 doi: 10.1016/j.cocis.2021.101464. [DOI] [Google Scholar]
- 20.Nguyen V.P., Paulus Y.M. Photoacoustic ophthalmoscopy: principle, application, and future directions. J. Imaging. 2018;4.12:149. doi: 10.3390/jimaging4120149. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Shao P., Cox B., Zemp R.J. Estimating optical absorption, scattering, and grueneisen distributions with multiple-illumination photoacoustic tomography. Appl. Opt. 2011;50(19):3145–3154. doi: 10.1364/AO.50.003145. [DOI] [PubMed] [Google Scholar]
- 22.Haltmeier M., Nguyen L.V. Analysis of iterative methods in photoacoustic tomography with variable sound speed. SIAM J. Imaging Sci. 2017;10.2:751–781. doi: 10.1137/21m1463409. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Kazakeviciute A., Ho C.J.H., Olivo M. Multispectral photoacoustic imaging artifact removal and denoising using time series model-based spectral noise estimation. IEEE Trans. Med. Imaging. 2016;35(9):2151–2163. doi: 10.1109/TMI.2016.2550624. [DOI] [PubMed] [Google Scholar]
- 24.Amidi E., Yang G., Uddin K.M.S., Luo H., Middleton W., Powell M., Siegel C., Zhu Q. Role of blood oxygenation saturation in ovarian cancer diagnosis using multi‐spectral photoacoustic tomography. J. biophotonics. 2021;14(4) doi: 10.1002/jbio.202000368. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Mallidi Srivalleesha, Luke Geoffrey P., Emelianov Stanislav. Photoacoustic imaging in cancer detection, diagnosis, and treatment guidance. Trends Biotechnol. 2011;29.5:213–221. doi: 10.1137/16M1104822. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Su J.L., Wang B., Wilson K.E., Bayer C., Chen Y., Kim S., Homan K.A., Emelianov S.Y. Advances in clinical and biomedical applications of photoacoustic imaging. Expert Opin. Med. Diagn. 2010;4.6:497–510. doi: 10.1517/17530059.2010.529127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Cox B.T., Arridge S.R., Kostli K.P., Beard P.C. Two-dimensional quantitative photoacoustic image reconstruction of absorption distributions in scattering media by use of a simple iterative method. Appl. Opt. 2006;45(8):1866–1875. doi: 10.1364/ao.45.001866. [DOI] [PubMed] [Google Scholar]
- 28.Huang C., Wang K., Nie L., Wang L.V., Ananstasio M.A. Full-wave iterative image reconstruction in photoacoustic tomography with acoustically inhomogeneous media. IEEE Trans. Med. Imaging. 2013;32.6:1097–1110. doi: 10.1109/TMI.2013.2254496. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Guan S., Hsu K., Eyassu M., Vhitnis P.V. Dense dilated UNet: deep learning for 3D photoacoustic tomography image reconstruction. arXiv Prepr. arXiv. 2021 doi: 10.48550/arXiv.2104.03130. 2104.03130. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.C. Yang, H. Lan, H. Zhong, and F. Gao, Quantitative photoacoustic blood oxygenation imaging using deep residual and recurrent neural network, 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019). IEEE, 2019. 10.1109/ISBI.2019.8759438. [DOI]
- 31.Awasthi N., Jain G., Kalva S.E., Pramanik M., Yalavarthy P.K. Deep neural network-based sinogram super-resolution and bandwidth enhancement for limited-data photoacoustic tomography. IEEE Trans. Ultrason. Ferroelectr., Freq. Control. 2020;67.12:2660–2673. doi: 10.1109/TUFFC.2020.2977210. [DOI] [PubMed] [Google Scholar]
- 32.Anas E.M.A., Zhang H.K., Kang J., Boctor E. Enabling fast and high quality LED photoacoustic imaging: a recurrent neural networks based approach. Biomed. Opt. Express. 2018;9.8:3852–3866. doi: 10.1364/BOE.9.003852. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Cai C., Deng K., Ma C., Luo J. End-to-end deep neural network for optical inversion in quantitative photoacoustic imaging. Opt. Lett. 2018;43.12:2752–2755. doi: 10.1364/OL.43.002752. [DOI] [PubMed] [Google Scholar]
- 34.Luke G.P., Hawlik K.H., Namen A.C.V., Shang R. O-Net: a convolutional neural network for quantitative photoacoustic image segmentation and oximetry. arXiv Prepr. arXiv. 2019 doi: 10.48550/arXiv.1911.01935. 1911.01935. 1911.01935. [DOI] [Google Scholar]
- 35.Bench C., Hauptmann A., Cox B.T. Toward accurate quantitative photoacoustic imaging: learning vascular blood oxygen saturation in three dimensions. J. Biomed. Opt. 2020;25.8 doi: 10.1117/1.JBO.25.8.085003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Liang G., Zheng L. A transfer learning method with deep residual network for pediatric pneumonia diagnosis. Comput. Methods Prog. Biomed. 2020;187 doi: 10.1016/j.cmpb.2019.06.023. [DOI] [PubMed] [Google Scholar]
- 37.A. Mahajan, S. Chaudhary, Categorical image classification based on representational deep network (RESNET), 2019 3rd International conference on Electronics, Communication and Aerospace Technology (ICECA). IEEE, 2019. 10.1109/ICECA.2019.8822133. [DOI]
- 38.J. Deng, W. Dong, R. Socher, L.J. Li, K. Li, F. Li, Imagenet: A large-scale hierarchical image database, 2009 IEEE conference on computer vision and pattern recognition. Ieee, 2009. 10.1109/CVPR.2009.5206848. [DOI]
- 39.Ronneberger O., Fischer P., Brox T. Springer; Cham: 2015. U-net: Convolutional networks for biomedical image segmentation, International Conference on Medical image computing and computer-assisted intervention. [Google Scholar]
- 40.Drozdzal M., Vorontsov E., Chartrand G., Kadoury S., Pal C. Springer; Cham: 2016. The importance of skip connections in biomedical image segmentation, Deep learning and data labeling for medical applications; pp. 179–187. [DOI] [Google Scholar]
- 41.Menonm U., Maharaj A.G., Hallett R., Ryan A., Burnell M., Shama A., Lewis S., Davies S., Philpott S., Lopes A., Godfrey K., Oram D., Herod J., Williamson K., Seif M.W., Scott I., Would T., Woolas R., Murdoch J., Dobbs S., Amso N.N., Leeson S., Cruickshank D., Fallowfield L., Singh N., Dawnay A., Skates S.J., Parmar M., Jacobs I. Sensitivity and specificity of multimodal and ultrasound screening for ovarian cancer, and stage distribution of detected cancers: results of the prevalence screen of the UK Collaborative Trial of Ovarian Cancer Screening (UKCTOCS) Lancet Oncol. 2009;10.4:327–340. doi: 10.1016/S1470-2045(09)70026-9. [DOI] [PubMed] [Google Scholar]
- 42.Timmerman D., Testa A.C., Bourne T., Ameye L., Jurkovic D., Holsbeke C.V., Paladini D., Calster B.V., Vergote I., Huffel S.V., Valentin L. Simple ultrasound‐based rules for the diagnosis of ovarian cancer. Ultrasound Obstet. Gynecol. 2008;31.6:681–690. doi: 10.1002/uog.5365. [DOI] [PubMed] [Google Scholar]
- 43.Kamal R., Hamed S., Mansour S., Mounir Y., Sallam S.A. Ovarian cancer screening—ultrasound; impact on ovarian cancer mortality. Br. J. Radiol. 2018;91.1090:20170571. doi: 10.1259/bjr.20170571. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Aguirre A., Ardeshirpour Y., Sanders M.M., Brewer M., Zhu Q. Potential role of coregistered photoacoustic and ultrasound imaging in ovarian cancer detection and characterization. Transl. Oncol. 2011;4(1):29–37. doi: 10.1593/tlo.10187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition. 2016.
- 46.Liu H., Boas D.A., Zhang Y., Yodh A.G., Chance B. Determination of optical properties and blood oxygenation in tissue using continuous NIR light. Phys. Med. Biol. 1995;40.11:1983. doi: 10.1088/0031-9155/40/11/015. [DOI] [PubMed] [Google Scholar]
- 47.Nandy S., Mostafa A., Hagemann I.S., Powell M.A., Amidi E., Robinson K., Mutch D.G., Siegel C., Zhu Q. Evaluation of ovarian cancer: initial application of coregistered photoacoustic tomography and US. Radiology. 2018;289(3):740–747. doi: 10.1148/radiol.2018180666. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Salehi H.S., Wang T., Kumavor P.D., Li H., Zhu Q. Design of miniaturized illumination for transvaginal coregistered photoacoustic and ultrasound imaging. Biomed. Opt. Express. 2014;5(9):3074–3079. doi: 10.1364/BOE.5.003074. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Yang G., Amidi E., Champman W., Nandy S., Mostafa A., Abdelal H., Alipour Z., Chatterjee D., Mytch M., Zhu Q. Co-registered photoacoustic and ultrasound real-time imaging of colorectal cancer: ex-vivo studies. Photons Ultrasound.: Imaging Sens. 2019;Vol. 10878 doi: 10.1117/12.2507638. International Society for Optics and Photonics, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.S. Yan, and Q. Fang, Hybrid mesh and voxel based Monte Carlo algorithm for accurate and efficient photon transport modeling in complex bio-tissues, Biomed. Opt. Express, 11(11) pp. 6262–6270. 10.1364/BOE.409468. [DOI] [PMC free article] [PubMed]
- 51.Treeby B.E., Cox B.T. k-wave: Matlab toolbox for the simulation and reconstruction of photoacoustic wave fields. J. Biomed. Opt. 2010;15(2) doi: 10.1117/1.3360308. [DOI] [PubMed] [Google Scholar]
- 52.Klemm M., Crddock I.J., Leendertz J.A., Preeece A., Benjamin R. Improved delay-and-sum beamforming algorithm for breast cancer detection. Int. J. Antennas Propag. 2008;2008 doi: 10.1155/2008/761402. [DOI] [Google Scholar]
- 53.Andreotti R.F., Timmerman D., Strachowski L.M., Froyman W., Benacerragf B.R., Bennett G.L., Bourne T., Brown D.L., Coleman B.G., Frates M.C., Goldstein S.R., Hamper U.M., Horrow M.M., Schulman M.H., Reinhold C., Rose S.L., Whitcomb B.P., Wolfman W.L., Glanc P. O-RADS US risk stratification and management system: a consensus guideline from the ACR Ovarian-Adnexal Reporting and Data System Committee. Radiology. 2020;294.1:168–185. doi: 10.1148/radiol.2019191150. [DOI] [PubMed] [Google Scholar]
- 54.Shetty M. Vol. 40. WB Saunders,; 2019. (Imaging and differential diagnosis of ovarian cancer, Seminars in Ultrasound, CT and MRI). [DOI] [PubMed] [Google Scholar]
- 55.Guo Z., Li L., Wang L.V. On the speckle-free nature of photoacoustic tomography. Med Phys. 2009;36(9):4084–4088. doi: 10.1118/1.3187231. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.R.R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, Grad-cam: Visual explanations from deep networks via gradient-based localization, Proceedings of the IEEE international conference on computer vision. 2017.
- 57.L. Chen, J. Chen, H. Hajimirsadeghi, and G. Mori, Adapting grad-cam for embedding networks, Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 2020.
- 58.Nandy S., Mostafa A., Kumavor P.D., Sanders M., Brewer M., Zhu Q. Characterizing optical properties and spatial heterogeneity of human ovarian tissue using spatial frequency domain imaging. J. Biomed. Opt. 2016;21(10) doi: 10.1117/1.JBO.21.10.101402. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Roggan A., Friebel M., Dörschel K., Hahn A., A, Mueller G.J. Optical properties of circulating human blood in the wavelength range 400–2500 nm. J. Biomed. Opt. 1999;4(1):36–46. doi: 10.1117/1.429919. [DOI] [PubMed] [Google Scholar]
- 60.Bosschaart N., Edelman G.J., Aalders M.C., van Leeuwen T.G., Faber D.J. A literature review and novel theoretical approach on the optical properties of whole blood. Lasers Med. Sci. 2014;29(2):453–479. doi: 10.1007/s10103-013-1446-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data will be made available on request.