Skip to main content
IEEE Sponsored Documents logoLink to IEEE Sponsored Documents
. 2021 Jul 19;6(4):446–453. doi: 10.1109/TRPMS.2021.3098448

Deep Learning Signal Discrimination for Improved Sensitivity at High Specificity for CMOS Intraoperative Probes

Joshua Moo 1,, Paul Marsden 1, Kunal Vyas 2, Andrew J Reader 1
PMCID: PMC8991998  PMID: 35419499

Abstract

The challenge in delineating the boundary between cancerous and healthy tissue during cancer resection surgeries can be addressed with the use of intraoperative probes to detect cancer cells labeled with radiotracers to facilitate excision. In this study, deep learning algorithms for background gamma ray signal rejection were explored for an intraoperative probe utilizing CMOS monolithic active pixel sensors optimized toward the detection of internal conversion electrons from Inline graphicTc. Two methods utilizing convolutional neural networks (CNNs) were explored for beta-gamma discrimination: 1) classification of event clusters isolated from the sensor array outputs (SAOs) from the probe and 2) semantic segmentation of event clusters within an acquisition frame of an SAO which provides spatial information on the classification. The feasibility of the methods in this study was explored with several radionuclides including 14C, 57Co, and Inline graphicTc. Overall, the classification deep network is able to achieve an improved area under the curve (AUC) of the receiver operating characteristic (ROC), giving 0.93 for 14C beta and Inline graphicTc gamma clusters, compared to 0.88 for a more conventional feature-based discriminator. Further optimization of the lower left region of the ROC by using a customized AUC loss function during training led to an improvement of 31% in sensitivity at low false positive rates compared to the conventional method. The segmentation deep network is able to achieve a mean dice score of 0.93. Through the direct comparison of all methods, the classification method was found to have a better performance in terms of the AUC.

Keywords: Cancer surgeries, CMOS, convolutional neural network (CNN), internal conversion (IC) electrons, intraoperative beta probes

I. Introduction

One of the main challenges faced in cancer resection surgeries is the difficulty in delineating the boundary between cancerous and healthy tissue, leading to suboptimal surgical outcomes due to the overestimation of the extent of tumor margins in order to achieve complete tumor removal, which inadvertently causes the excessive removal of healthy tissue. On the other hand, in cases where tumor excision is incomplete, subsequent resection operations and extensive postoperative adjuvant radiotherapy are required which in turn will reduce patient survival rates [1]. Current margin evaluation methods involve the pathological analysis of biopsies taken during surgery, where the results can take up to weeks to be able to inform if re-excision is required [2]. Therefore, there is a need for the real-time detection of cancerous tissue during surgery which can be achieved through intraoperative technology.

Optical image-guided cancer surgery which involves the detection of photons emitted by tumor-specific fluorescence agents was also presented as a solution for real-time tumor margin delineation. However, the need for the development of these tumor-specific agents has limited clinical adoption due to: 1) the efficiency of the agents in being able to reach and remain at the target location; 2) degradation in signal localization due to absorption and scattering of light in tissue; and 3) lack of financial incentives for the production of these agents for diagnostic purposes [3].

The use of gamma and beta emitting radiopharmaceuticals which accumulates within cancer cells in conjunction with intraoperative probes presents a potential solution for tumor localization during surgery. Initially, several gamma sensitive intraoperative probes were designed to target Inline graphic emitting radiotracers due to the large range of gamma rays in tissue, which allows the detection of tumors that are located deep under the tissue surface. However, these attempts were found to be suboptimal due to the susceptibility of the gamma probes to the highly penetrating gamma rays from distant organs due to the nonspecific uptake of the radiotracer, degrading the tumor-to-background signal ratio which leads to the ambiguity of tumor location [4], [5].

Research interests have then shifted toward beta sensitive intraoperative probes due to the short range of electrons and positrons in tissue, which presents as an advantage over gamma probes since signal is detected only when placed in close proximity to the source. Moreover, the effects associated with the nonspecific radiotracer uptake are diminished due to lower signal interference from distant sources, which improves the spatial selectivity of the probe [6], [7]. Several beta intraoperative probes have been proposed with designs that are optimized toward the rejection of background gamma rays by subtraction or coincidence detection through the implementation of multiple detectors within a single probe which imposes physical limitations on the design of the probe [6], [8], [9]. For example, Yamamoto et al. [9] proposed a beta intraoperative probe involving a phoswich detector composed of a plastic scintillator and a high Z scintillator, bismuth germanate (BGO). The probe employs pulse shape discrimination for positron detection and rejection of background gamma events through coincident detection of a positron and one of its corresponding 511-keV annihilation photons in conjunction with the different decay times of the scintillators. Therefore, the discrimination method is only restricted to 18F-fluorodeoxyglucose (FDG) guided surgeries. On top of that, shielding of the probe is required to reduce the susceptibility to background gamma rays which would inadvertently compromise the sensitivity of the probe for positrons.

Using CMOS sensors for the detection of ionizing radiation have been investigated in [10], and more specifically beta emissions in [11]. These studies have shown the feasibility of using CMOS sensors for the detection of Inline graphic emissions from 90Sr and 90Y (pure Inline graphic emitting sources) phantoms under realistic clinical conditions through the implementation of a clustering algorithm for event detection, which considers a neighborhood of pixels around high intensity seed pixels. Although the sensitive (epitaxial) layer of CMOS sensors is generally sufficiently thin (orders of a few micrometers) to ensure low gamma ray interaction probabilities, there is still a finite detection efficiency which means there will be background gamma ray signal contamination. Therefore, there is still a need for the active rejection of background gamma ray of lower energies through the analysis of the event clusters detected by the CMOS sensor. The use of convolutional neural networks (CNNs): 1) utilizes their ability to learn and extract appropriate features from training examples for the specific classification of beta and gamma clusters and 2) obviates the need for hand-crafted features which relies on the incorrect assumption that such features are able to capture all relevant information for the classification task [12].

In this study, two deep learning architectures for background gamma ray rejection are proposed for an intraoperative probe utilizing CMOS monolithic active pixel sensors optimized toward the detection of internal conversion (IC) electrons from Inline graphicTc, which is more readily available and cost effective. On top of that, Inline graphicTc also has a longer half-life of 6 h which allows for more flexible scheduling of surgeries. The charge distribution over several detector pixels caused by the emissions can be isolated (Fig. 1) and the rejection of gamma signal is possible with only a single detector if discrimination between gamma and beta clusters can be achieved. Therefore, methods utilizing CNNs were explored for beta–gamma discrimination: 1) classification of event clusters isolated from the sensor array outputs (SAOs) and 2) semantic segmentation of event clusters within each SAO acquisition frame.

Fig. 1.

Fig. 1.

(a) Example of beta cluster from 14 C and (b) gamma cluster from Inline graphicTc. The event clusters are cropped based on the size of their bounding box and zero padded.

II. Methods and Materials

A. Detector

A prototype intraoperative probe (Fig. 2) developed by Lightpoint Medical utilizes a CMOS sensor ( Inline graphic 640 pixels, pixel size Inline graphic, and a sensitive area of 3.84 mm Inline graphic mm). The thickness of the sensitive layer, which is the epitaxial layer of the CMOS sensor is approximately Inline graphic (a similar model was used in [11]) and is capable of achieving 60 frames/s. The detection efficiency of a CMOS sensor with a similar sensitive layer thickness is reported to be 2% for gamma rays with energies higher than 20 keV [11]. The probe is connected to a PC via USB for image acquisition, transmission, processing, and analysis. To account for the dark current or electronic noise of each pixel, several blank frames are initially acquired to obtain an average dark image which is then subtracted from each image acquired in the presence of a radioactive source.

Fig. 2.

Fig. 2.

Prototype CMOS intraoperative probe developed by lightpoint medical.

B. Sources and Cluster Detection

A labeled dataset, where the identity of the clusters is known, is required for training the CNNs for beta–gamma discrimination. Therefore, SAOs from pure Inline graphic (14C) and Inline graphic (57Co) emitting radionuclides were obtained which allows the clusters to be identified. SAOs containing only Inline graphicTc gammas were acquired by using a 2-mm PMMA film to shield against IC electrons. The emission types and probabilities of the radionuclides used in this study are shown in Table I. Inline graphic clusters from 14C was used as a surrogate for Inline graphicTc IC electrons due to their similar emission energies.

TABLE I. List of Emissions for Radionuclides Used.

Radioisotope Emission type Emission probability (%) Energy (keV)
14C Inline graphic 100 156 (max)
57Co Inline graphic 91 136
Inline graphic 9 122
99mTc Inline graphic 88 140
IC electron 12 119–140*
*

Energies are discrete and dependent upon the binding energy of the electron shell

A threshold is applied to the dark current-corrected image which results in a binary image containing the clusters. Connected components labeling (8-connectivity) is then applied to the binary image which allows the position and bounding box size of each event cluster within a frame to be determined. Using this information, the clusters are isolated from the original dark current-corrected image which is then saved to form a dictionary. SAOs with 14C were used to generate a beta cluster dictionary, whereas SAOs from 57Co and Inline graphicTc with PMMA were used to generate the respective gamma cluster dictionaries. Overall, 123 107 14C beta clusters, 20 359 57Co gamma clusters, and 9887 Inline graphicTc gamma clusters were obtained.

C. Classification (Models C1 and C2)

The individual clusters in the dictionaries were initially zero padded to an image patch of Inline graphic pixels in order to account for the variable cluster size which allows for a consistent input to the network. The clusters patches were also normalized to values between 0 and 1. The overall model architecture (Fig. 3) consists of three convolutional layers with kernels of size Inline graphic, followed by a rectified linear unit (ReLU, a commonly used nonlinear activation function which sets negative values to zero while retaining all positive values [13]) activation at each layer. Each convolutional layer is also followed by a max pooling layer with Inline graphic kernel size for feature dimension reduction. The filter maps after these feature extraction operations are then flattened to provide a 1-D feature vector to 4 fully connected layers with 128, 64, 32, and 2 nodes, respectively. A softmax activation (which normalizes a vector into a probability distribution [14]) is applied at the output layer to provide a probabilistic score for each class and the loss function used is the categorical cross-entropy (CE)

C.

which is calculated between the output prediction, Inline graphic, and one-hot encoded ground truth, Inline graphic, across the number of samples, Inline graphic, and the number of classes, Inline graphic. The classification architecture is adapted from LeNet-5 [12] which was originally designed for handwriting recognition. Given that there are only two classes for this application, the number of layers and channels was reduced from the original LeNet-5 architecture to the facilitate training time. 60% of the total clusters were used for training, 20% for validation, and 20% for testing. Model C1 is trained and tested with 14C beta and 57Co gamma clusters and model C2 is trained and tested with 14C beta and Inline graphicTc gamma clusters. The reason for training the two different models is to evaluate the ability of the network to discriminate between beta and gamma clusters from pure sources as well as when Inline graphicTc is used, which is the radiotracer of interest in this application. Due to the difficulty in obtaining SAOs containing only IC electrons from Inline graphicTc, the beta clusters from 14C were used in model C2 since the energies of these betas are comparable to the IC electrons. A receiver operating characteristic (ROC) curve was generated for both models where the true positive rate (TPR) is plotted against the false positive rate (FPR). The positive class in the ROC curves corresponds to the beta class.

Fig. 3.

Fig. 3.

Model architecture for classification of beta and gamma clusters.

The area under the ROC curves (AUCs) is used as a measurement for the performance of the classifiers, where an AUC of 1 corresponds to a perfect classifier. The AUC of the models were also compared to a feature-based lookup table classifier developed by Lightpoint Medical which is implemented in MATLAB (The Mathworks, Inc.). Initially, 2-D histograms for beta [Fig. 4(a)] and gamma [Fig. 4(b)] clusters based on their mean intensity and area (the product of which is proportional to the total energy deposited) were generated, which are then normalized by the total number of events to give a probability distribution. The beta-events histogram is divided by the gamma-events histogram to obtain a selectivity map, where the lookup table [Fig. 4(c)] can be generated such that clusters that fit into regions on the map that shows selectivity above a threshold will be classified as betas. 14C beta and Inline graphicTc gamma clusters were also tested on model C1 to evaluate the possibility of applying transfer learning during the training process such that a model trained on pure beta and gamma emitting sources is able to distinguish between gamma and IC electron signals from Inline graphicTc.

Fig. 4.

Fig. 4.

2-D histograms of mean intensity and area of (a) beta and (b) gamma events. (c) Lookup table generated from both histograms.

In this application, misclassifying gamma clusters as beta clusters would falsely indicate the presence of cancerous tissue within the region of interest. Therefore, the cost of false positives is higher than false negatives which means a decision threshold will be selected such that the classifiers will be operating at a low FPR or high specificity. Given that the CE loss function does not take into account these misclassification costs during training, there is a need for further optimization of the classifier operating at high specificity, which corresponds to the lower left region of the ROC curve. Therefore, an alternative loss function which utilizes the equivalency of the Wilcoxon-Mann-Whitney (WMW) statistic and the AUC was used to train the model to directly maximize the AUC of the lower left region of the ROC curve [15]. The WMW statistic is approximated by the differentiable function

C.

where Inline graphic and Inline graphic are the probabilistic scores for the positive and negative class, respectively. The hyperparameters Inline graphic and Inline graphic are set to 0.7 and 2, respectively, which corresponds to the optimal values determined through grid search in [15]. The loss function Inline graphic can then be formulated as

C.

where Inline graphic and Inline graphic are the number of positive and negative samples, respectively. Positive samples with scores that are higher than negative samples by a margin Inline graphic have no contribution to the loss function. The loss function can then be extended to focus on the lower left region of the ROC curve by mapping the probabilistic scores Inline graphic obtained from the final softmax layer of the classification network function

C.

where Inline graphic is the mean of the probabilistic scores and the hyperparameter Inline graphic is set to 1.1, which is again found to be optimal through grid search. Thus, the network is then trained using the loss function Inline graphic

C.

where training is only focused on samples with scores higher than the mean score. This concept of explicitly selecting data that is within the region of interest on the ROC curve for training is also demonstrated in [16]. In this case, clusters assigned with a high probabilistic score by the network are either easily identifiable beta clusters or difficult gamma clusters. Learning to distinguish between these clusters would contribute the most to improving the AUC at the lower left region of the ROC.

D. Segmentation (Models S1 and S2)

Semantic segmentation was also explored as a method to directly label the beta and gamma clusters within an acquisition frame of an SAO in one step by assigning a class label to each pixel. The lack of ground-truth SAOs containing both beta and gamma clusters motivated the synthesis of SAOs using clusters from measured dictionaries. Initially, the clusters are added to a zero image and the remaining zero values in the resulting image are then replaced with values sampled from a fitted normal distribution of background intensities where the mean and variance are determined from real images acquired by the detector. The synthesized SAOs are of size Inline graphic (smaller than the native SAO to reduce computational costs). Finally, the images are then normalized to values between 0 and 1 which is then used as the input to the segmentation network.

The network architecture for segmentation is shown in Fig. 5, consisting of three convolutional layers with 32 kernels of size Inline graphic with ReLU activation at each layer. This is then followed by a convolutional layer consisting of three kernels of size Inline graphic with softmax activation. The output consists of an image with three channels where each pixel contains three probabilistic scores for the background, beta, and gamma class. The class with the highest score is then assigned to the pixel. The targets are Inline graphic images where each channel corresponds to a binary image for each class and the loss function used is the generalized dice loss (GDL) which is reported to have better performance over the more conventionally used CE loss in the presence of class imbalance [17], which in this case there are more pixels in the SAO belonging to the background class compared to either the beta/gamma class

D.

with

D.

where Inline graphic corresponds to the ground truth or gold standard segmentation labels and Inline graphic corresponding to the predicted probabilistic values, calculated across Inline graphic number of pixels and Inline graphic number of classes. The contribution to the loss function from each class is weighted by the inverse of its area Inline graphic to provide invariance to class imbalance. Model S1 is trained and tested with SAOs containing 14C beta and 57 Co gamma clusters and Model S2 is trained and tested with SAOs containing 14 C beta and Inline graphicTc gamma clusters. A total of 500 images were synthesized with 60% used for training, 20% for validation, and 20% for testing. To be able to provide a performance comparison in terms of beta–gamma discrimination with the classification methods mentioned above, an ROC curve was generated based on the scores of the clusters. To achieve this, the beta and gamma scores for each cluster are extracted and a softmax activation is applied. The mean of the resulting beta scores of each cluster are then used to generate the ROC curve. To evaluate whether the number of patches used in this study is sufficient to optimize the discrimination task, model C2 was retrained at different fractions of the total number of patches obtained experimentally and the AUC metric on the test set was recorded.

Fig. 5.

Fig. 5.

Model architecture for semantic segmentation. A frame of the SAO is fed into the network where the event clusters will be labeled and segmented.

III. Results

A. Classification

The training and validation losses for models C1 and C2 are shown in Fig. 6. The ROC curves for both models are shown in Fig. 7 where the AUCs for model C1 and C2 are 0.988 and 0.936, respectively. Fig. 8 shows a comparison between the CE and WMW loss functions with model C2. Overall, model C2 trained with the WMW loss achieved the lowest AUC [Fig. 8(a)]. However, if we consider the ROC up to 2% FPR (lower-left region of the ROC), WMW loss achieved the best performance in terms of the AUC when compared to either using CE or the lookup table [Fig. 8(b)]. Fig. 9 shows the ROC curves for all classifiers tested with 14C beta and Inline graphicTc gamma clusters. A lower AUC is observed with model C1 compared to model C2, however, it is still comparable to the lookup table trained with 14C beta and Inline graphicTc gamma clusters. The AUC score as a function of training set size with model C2 is reported in Fig. 10.

Fig. 6.

Fig. 6.

Training and validation loss as a function of the number of epochs. Early stopping was determined through monitoring the validation loss with a patience of 10 epochs.

Fig. 7.

Fig. 7.

Comparison of ROC curves between CNN classifier and lightpoint lookup table classifier. (a) ROC for Model C1, trained and tested with 14C beta and 57Co gamma clusters. (b) ROC for Model C2, trained and tested with 14C beta and Inline graphicTc gamma clusters.

Fig. 8.

Fig. 8.

(a) Comparison of overall ROC between categorical CE loss and WMW loss with model C2. (b) Comparison of lower left part of the ROC. AUC values reported are at 2% FPR.

Fig. 9.

Fig. 9.

Comparison of ROC for all classifiers tested on 14C beta and Inline graphicTc gamma clusters. All classifiers were trained on 14C beta clusters and the gamma clusters depicted in the square brackets for each classifier.

Fig. 10.

Fig. 10.

Training set size grid search conducted with model C2 trained using CE loss. The mean and standard deviation values are calculated across five repeats.

B. Segmentation

The dice scores of each individual class were monitored as the performance metric and are shown in Fig. 11. The dice scores for the background class are close to unity for both models. However, lower dice scores for the gamma and beta class are found with model S2 when compared to model S1. The ROC curve generated based on the mean beta scores of the clusters from segmentation with model S2 is compared with model C2 in Fig. 12. A lower AUC is observed with model S2.

Fig. 11.

Fig. 11.

Comparison of dice scores for each class between models S1 and S2.

Fig. 12.

Fig. 12.

Comparison of ROC between models C2 and S2.

IV. Discussion and Conclusion

Both deep learning methods have shown promising results for beta–gamma discrimination. Better performance in terms of the AUC is observed with both models C1 and C2 compared to the lookup table classifier, as the classification network learns many features for itself, compared to the two handcrafted features which are the mean intensity and area of the clusters. The lower performance observed with Model C2 compared to C1 could be due to the presence of IC electrons in the Inline graphicTc gamma dictionary and a higher degree of class imbalance between 14C beta and Inline graphicTc gamma clusters. The performance with model C1 when tested on Inline graphicTc gamma clusters shows the possibility of implementing transfer learning during the training process such that the classification network can be applied to discriminating the emissions from Inline graphicTc. This would be especially beneficial due to the difficulty in isolating the IC electron signals from Inline graphicTc for training the classifier.

The need to optimize the lower left region of the ROC curve is essential since misclassifying gamma signals as beta signals would lead to unhelpful ambiguity of the tumor location. The use of the WMW loss allowed the direct optimization of the AUC as well as improvement in sensitivity at high specificity. The model trained with WMW loss achieved approximately an improvement of 31% in AUC compared to the model trained with CE loss up to an FPR of 0.02, which would be within the typical range of operation in a surgical environment. At approximately 0.03 FPR, the model trained with CE loss begins to outperform the WMW loss in terms of sensitivity which is expected due to the mapping applied to the samples in (4). Therefore, this presents a tradeoff between optimization at high and low specificity.

Increasing the training set size shows an improvement in performance in terms of the AUC is seen in Fig. 10. With a training set size above 60 000, the AUC begins to plateau which shows a point of diminishing returns has been reached and further increase in training set size will yield a smaller increase in performance with a longer training time. This also shows that the total number of patches used in this study is sufficient for the discrimination task.

Examples of misclassifications and each type of clusters are shown in Fig. 13. These cases are primarily due to the clusters having similar appearances and physical properties to clusters in the other class. This could be due to the presence of noise in the training data labels, which causes the ambiguity of the identity of the cluster. An apparent example of a beta cluster in Fig. 13(d) (highlighted in green) was misclassified by the lookup table but not the CNN classifier. This demonstrates that the features learned by the CNN for the classification were able to outperform the hand-crafted features used to generate the lookup table.

Fig. 13.

Fig. 13.

(a) and (b) Examples of misclassifications made by both CNN and lookup table classifiers. (c) and (d) Examples of misclassifications made by the lookup table but not the CNN. (e) and (f) Examples of beta and gamma clusters, respectively. The gamma cluster in row 2, column 3 in (a) has similar appearance and physical properties to the beta cluster in row 1, column 1 of (e) (highlighted in red). A similar comparison can be made between the beta cluster in row 2, column 1 of (b) and the gamma cluster in row 1, column 4 of (f) (highlighted in yellow). An apparent example of a beta cluster in (d) was misclassified by the lookup table but not the CNN classifier (highlighted in green).

The dice scores with semantic segmentation for the background class demonstrates the ability of the networks to segment the clusters from the SAOs. The dice scores for the beta and gamma class have also shown that discrimination between the two can be achieved in one step. However, better performance in terms of the AUC is observed for the classification method because more information is available, requiring only one classification of many pixels, compared to requiring classification for each and every pixel in the SAO for the segmentation network. Overall, through the direct comparison of both methods, the classification network yielded better performance in beta–gamma discrimination.

In conclusion, the deep learning-based gamma signal rejection methods in this study have shown improved performances compared to a deterministic machine-learned method as well as further improvement at low FPRs which is essential for clinical situations.

Acknowledgment

For the purpose of open access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission.

Funding Statement

This work was supported in part by the Wellcome Trust through the Engineering and Physical Sciences Research Council (EPSRC), part of the EPSRC DTP under Grant EP/R513064/1, and in part by the Lightpoint Medical Ltd., under the CASE Studentship Scheme by core funding from the Wellcome/EPSRC Centre for Medical Engineering under Grant WT203148/Z/16/Z.

References

  • [1].Jeevan R.et al. , “Reoperation rates after breast conserving surgery for breast cancer among women in England: Retrospective study of hospital episode statistics,” Brit. Med. J., vol. 345, Jul. 2012, Art. no. e4505. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Boughey J. C., Keeney G. L., Radensky P., Song C. P., and Habermann E. B., “Economic implications of widespread expansion of frozen section margin analysis to guide surgical resection in women with breast cancer undergoing breast-conserving surgery,” J. Oncol. Pract., vol. 12, no. 4, pp. e413–e422, Feb. 2016. [DOI] [PubMed] [Google Scholar]
  • [3].Keereweer S.et al. , “Optical image-guided cancer surgery: Challenges and limitations,” Clin. Cancer Res., vol. 19, no. 14, pp. 3745–3754, 2013. [Online]. Available: https://doi.org/10.1158/1078-0432.CCR-12-3598 [DOI] [PubMed] [Google Scholar]
  • [4].Davidson B. R., Waddington W. A., Short M. D., and Boulos P. B., “Intraoperative localization of colorectal cancers using radiolabelled monoclonal antibodies,” Brit. J. Surg., vol. 78, no. 6, pp. 664–670, Jun. 1991. [DOI] [PubMed] [Google Scholar]
  • [5].Piert M.et al. , “Positron detection for the intraoperative localisation of cancer deposits,” Eur. J. Nucl. Med. Mol. Imag., vol. 34, no. 10, pp. 1534–1544, 2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Daghighian F.et al. , “Intraoperative beta probe: A device for detecting tissue labeled with positron or electron emitting isotopes during surgery,” Med. Phys., vol. 21, no. 1, pp. 153–157, Jan. 1994. [DOI] [PubMed] [Google Scholar]
  • [7].Raylman R. R., Fisher S. J., Brown R. S., Ethier S. P., and Wahl R. L., “Fluorine-18-fluorodeoxyglucose-guided breast cancer surgery with a positron-sensitive probe: Validation in preclinical studies,” J. Nucl. Med., vol. 36, no. 10, pp. 1869–1874, Oct. 1995. [PubMed] [Google Scholar]
  • [8].Hoffman E. J., Tornai M. P., Janecek M., Patt B. E., and Iwanczyk J. S., “Intraoperative probes and imaging probes,” Eur. J. Nucl. Med., vol. 26, no. 8, pp. 913–935, 1999. [DOI] [PubMed] [Google Scholar]
  • [9].Yamamoto S., Matsumoto K., Sakamoto S., Tarutani K., Minato K., and Senda M., “An intra-operative positron probe with background rejection capability for FDG-guided surgery,” Ann. Nucl. Med., vol. 19, no. 1, p. 23, 2005. [DOI] [PubMed] [Google Scholar]
  • [10].Servoli L., Biagetti D., Passeri D., and Gattuso E. S., “Characterization of standard CMOS pixel imagers as ionizing radiation detectors,” J. Instrum., vol. 5, no. 7, pp. P07003–P07003, 2010. [Online]. Available: https://doi.org/10.1088/1748-0221/5/07/p07003 [Google Scholar]
  • [11].Solestizi L. A.et al. , “Feasibility study on the use of CMOS sensors as detectors in radioguided surgery with β–- emitters,” Appl. Radiat. Isotopes, vol. 165, Nov. 2020, Art. no. 109347. [Online]. Available: https://doi.org/10.1016/j.apradiso.2020.109347 [DOI] [PubMed] [Google Scholar]
  • [12].Lecun Y., Bottou L., Bengio Y., and Haffner P., “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278–2324, Nov. 1998. [Online]. Available: https://doi.org/10.1109/5.726791 [Google Scholar]
  • [13].Reader A. J., Corda-D’Incan G., Mehranian A., da Costa-Luis C., Ellis S., and Schnabel J., “Deep learning for PET image reconstruction,” IEEE Trans. Radiat. Plasma Med. Sci., vol. 5, no. 3, pp. 1–25, Jan. 2021. [Online]. Available: https://doi.org/10.1109/TRPMS.2020.3014786 [Google Scholar]
  • [14].Goodfellow I. J., Bengio Y., Courville A., “6.2.2.3 softmax units for multinoulli output distributions,” in Deep Learning. Cambridge, MA, USA: MIT Press, 2016, pp. 180–184. [Google Scholar]
  • [15].Yan L., Dodier R., Mozer M. C., and Wolniewicz R., “Optimizing classifier performance via an approximation to the wilcoxon-mann-whitney statistic,” in Proc. 20th Int. Conf. Int. Conf. Mach. Learn., 2003, pp. 848–855. [Google Scholar]
  • [16].Yih W., Goodman J., and Hulten G., “Learning at low false positive rates,” in Proc. CEAS, 2006. [Google Scholar]
  • [17].Sudre C. H., Li W., Vercauteren T., Ourselin S., and Cardoso M. J., “Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support (Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 10553, Heidelberg, Germany: Springer, 2017, pp. 240–248. [Online]. Available: https://doi.org/10.1007/978-3-319-67558-9_28 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Ieee Transactions on Radiation and Plasma Medical Sciences are provided here courtesy of Institute of Electrical and Electronics Engineers

RESOURCES