Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Dec 1.
Published in final edited form as: Pediatr Blood Cancer. 2018 Sep 10;65(12):e27417. doi: 10.1002/pbc.27417

Computer-assisted Curie scoring for metaiodobenzylguanidine (MIBG) scans in patients with neuroblastoma

Elizabeth A Sokol 1, Roger Engelmann 2, Wenjun Kang 3, Navin Pinto 4, Adam Starkey 2, Hollie Lai 5, Helen Nadel 6, Barry L Shulkin 7, Yonglin Pu 2, Daniel Appelbaum 2, Gregory A Yanik 8, Susan L Cohn 1, Samuel G Armato III 2,#, Samuel Volchenboum 1,3,#
PMCID: PMC6317352  NIHMSID: NIHMS995857  PMID: 30198643

Abstract

Background:

Radiolabeled metaiodobenzylguanidine (MIBG) is sensitive and specific for detecting neuroblastoma. The extent of MIBG-avid disease is assessed using Curie scores. Although Curie scoring is prognostic in patients with high-risk neuroblastoma, there is no standardized method to assess the response of specific sites of disease over time. The goal of this study was to develop approaches for Curie scoring to facilitate the calculation of scores and comparison of specific sites on serial scans.

Procedure:

We designed three semiautomated methods for determining Curie scores, each with increasing degrees of computer assistance. Method A was based on visual assessment and tallying of MIBG-avid lesions. For method B, scores were tabulated from a schematic that associated anatomic regions to MIBG-positive lesions. For method C, an anatomic mesh was used to mark MIBG-positive lesions with automatic assignment and tallying of scores. Five imaging physicians experienced in MIBG interpretation scored 38 scans using each method, and the feasibility and utility of the methods were assessed using surveys.

Results:

There was good reliability between methods and observers. The user-interface methods required 57 to 110 seconds longer than the visual method. Imaging physicians indicated that it was useful that methods B and C enabled tracking of lesions. Imaging physicians preferred method B to method C because of its efficiency.

Conclusions:

We demonstrate the feasibility of semiautomated approaches for Curie score calculation. Although more time was needed for strategies B and C, the ability to track and document individual MIBG-positive lesions over time is a strength of these methods.

Keywords: Curie score, MIBG, neuroblastoma

1 |. INTRODUCTION

Neuroblastoma is an embryonal tumor of the sympathetic nervous system responsible for 15% of pediatric cancer deaths in the United States.1 It displays genetic and clinical heterogeneity. Based on clinical and biologic variables, patients are assigned to risk groups and treatment regimens.2 Despite excellent outcomes for some, survival remains poor for high-risk patients despite intensive, multimodal therapies.35 Survival has been shown to be superior for those who respond to induction therapy.69 123I-MIBG whole-body scintigraphy is a powerful imaging technique for detecting neuroblastoma and evaluating treatment response.

The current standard in the Children’s Oncology Group (COG) for comparing successive 123I- MIBG scans to assess treatment response is the Curie method,10 as detailed in a recent consensus report from the International Neuroblastoma Risk Group (INRG) Task Force.11 The scoring algorithm divides the body into nine skeletal sections with a tenth soft-tissue section. The 10 sections are graded for extent of MIBG avidity on a 0 to 3 scale: 0 = no involvement, 1 = one site, 2 = more than one site, 3 = diffuse involvement (>50% of the segment). This method is considered “semi-subjective.” The Curie score is the sum of all 10 segments. Serial patient Curie scores are compared to assess treatment response.

Matthay et al reported significantly worse outcomes for patients with total Curie scores >2 following induction chemotherapy compared with those with scores of ≤2.6,12 Subsequent analyses determined that Curie scores >2 after induction but not at diagnosis in patients enrolled on a high-risk COG study were associated with significantly worse event-free survival (EFS).7 There was no correlation between Curie score at diagnosis and survival. More recently, in an International Society of Pediatric Oncology European Network (SIOPEN) high-risk study, researchers were able to validate that patients with a Curie score of ≤2 postinduction have significantly better EFS.13 Although the prognostic value of Curie scores has not yet been validated in high-risk patients receiving current COG standard treatment including tandem stem cell transplants and immunotherapy following induction, MIBG relative scores on bone sectors have been integrated into the recent revision of the International Neuroblastoma Response Criteria (INRC).14 The relative bone score is the ratio of the Curie scores at response assessment to diagnosis (without the soft-tissue component). Resolution of MIBG activity defines a complete response. A partial response is defined as a reduction of 50% or greater in MIBG bone score. Reduction by less than 50% is stable disease. Any new lesion represents progressive disease.

Although the prognostic significance of Curie scores has been established, the manual methods currently used to calculate total scores do not provide mechanisms to longitudinally track specific lesions or easily compare regions of diffuse involvement over time. Often, only the total Curie score is provided in the MIBG report by the imaging physician, without documentation of the specific lesions and sites of disease. We hypothesized that by integrating a computerized user interface and automation, Curie scores would be more accurately quantified and documented for longitudinal review. To test this hypothesis, we designed three semiautomated methods to calculate Curie scores with increasing degrees of computer assistance. A survey was administered to evaluate the imaging physicians’ opinions regarding the feasibility and utility of each method in clinical practice. The aim of the study was to compare the three methods to determine: (1) the efficiency of each method for determining Curie scores; (2) the vari ance of Curie scores between readers; and (3) the feasibility of tracking specific MIBG-positive lesions over time.

2 |. METHODS

2.1 |. Patient cohort

Patients with neuroblastoma and available MIBG scans were identified through The University of Chicago and COG. Institutional review board approval was obtained to collect imaging and clinical information from patients at the University of Chicago. Consents were obtained from patients available to consent, and a waiver was granted for patients unavailable to consent. Deidentified scans were also obtained from the Quality Assurance Review Center (QARC) through COG under a data use agreement with the University of Chicago. The use of these scans was additionally given a waiver of consent. Participating imaging physicians also signed research consents to collect data surrounding their use of the methods and their survey information.

2.2 |. MIBG semiquantitative Curie scoring

Three semiautomated mechanisms to calculate Curie scores were designed to allow imaging physicians to view and score planar anterior and posterior 24-hour MIBG scans. In each method, a computer interface (developed in the University of Chicago’s Abras system15) allowed the imaging physician to view the images, perform windowing and zooming of images, and determine Curie scores. To evaluate the three methods, MIBG scans were reviewed without any accompanying clinical information by five nuclear imaging physicians experienced in interpreting MIBG scans, from four academic institutions: University of Chicago, St. Jude Children’s Research Hospital, University of British Columbia, and Children’s Hospital of Los Angeles. Only planar images were reviewed. Accompanying single-photon emission computed tomography (SPECT) or SPECT/CT images, if performed, were not made available.

2.3 |. Method A: Manual Curie score within a computer interface

Similar to traditional Curie scoring, this method simply facilitated a sum of scores from 10 different anatomic sites, including skeletal (craniofacial, cervical and thoracic spine, chest (ribs/sternum/clavicles/scapula), lumbar and sacral spine, pelvis, humeri, lower arms, femurs, and lower legs) and soft tissue. As in traditional Curie scoring, skeletal sites were individually scored from 0 to 3 as described above. A score of 3 was assigned for the soft-tissue region if disease occupied >50% of the chest or abdomen. A patient’s Curie score at each time point was calculated as the sum of scores over all individual sites, with a maximum score of 30.

For this method, imaging physicians reviewed each set of images and clicked on buttons to directly indicate a 0 to 3 score for each of the nine anatomic segments (“regions”) and the soft-tissue segment (Fig. 1; Supporting Information Figure S2). This method most closely approximates the current scoring method of subjectively evaluating each segment and adding the scores. The interface eases this process by allowing the imaging physician to quickly click the score for each segment, while the overall tally is maintained. Furthermore, in contrast to traditional methods, the 10 individual region scores are preserved, allowing the clinician to return later to review the contribution from each segment.

FIGURE 1.

FIGURE 1

In the interface for method A, the radiologist indicated a score of 0–3 for each of the 10 Curie regions by using the buttons in the lower right part of the screen. See Supporting Information Figure S2 for text in the figure

2.4 |. Method B: Computer-assisted Curie scores

In method B, the imaging physicians marked the lesion locations on the images and then indicated the corresponding Curie segment by clicking a schematic figure. The score of each segment (0, 1, 2, or 3) was automatically computed (Fig. 2; Supporting Information Figure S2). The imaging physician could indicate a lesion in one of three ways: simple point, line for a linear bone lesion, or a loop to show the area of a lesion. When a line or loop was drawn, the interface collected user input on whether the lesions belonged to a segment with >50% tumor involvement. If so, the segment was automatically scored a 3 by the system. Otherwise, one lesion marked in a segment resulted in a 1 score for the segment, and two or more marks resulted in a 2 score. To indicate a soft-tissue lesion, the imaging physician held down a modifier key on the keyboard while drawing the lesion instead of clicking a skeletal segment on the schematic.

FIGURE 2.

FIGURE 2

In method B, the radiologist clicks on lesions in the anterior or posterior images and then indicates their corresponding anatomic segment by clicking on the schematic in the upper right portion of the screen. See Supporting Information Figure S2 for text in the figure

After drawing one or more lesions, the imaging physician indicates the corresponding anatomic segment. The scores were then automatically updated. The schematic was color coded as a visual reference for the imaging physician to reflect the current score for each anatomic segment (red: 3, orange: 2, yellow: 1). The system tracked the individ ual segment scores as well as the current total Curie score. The lesions and contributing segments are then preserved for later review.

2.5 |. Method C: Computer-assisted Curie scores

In method C, the imaging physician defined a Curie anatomic segment region map or “mesh” on the patient images by specifying key anatomic points that corresponded to points shown on a skeleton schematic (Fig. 3A and 3B; Supporting Information Figure S2), making adjustments to the region map as necessary by dragging the intersecting handles between the segments (Fig. 3C). An anterior image mesh defined seven of the nine skeletal regions, and a smaller posterior image mesh defined the other two skeletal regions corresponding to the spine. The ability to manipulate the mesh is especially useful in children, as there may not be uniformity in anatomic landmarks identified by the mesh generation software. The imaging physician could then mark lesions in either image (as points, lines, or loops), except that spine lesions were required to be marked in the posterior image, while medial rib, sternum, and pelvic lesions were required to be marked in the anterior image, in order to be assigned to the proper anatomic segment. Based on location, each lesion mark was automatically assigned to the corresponding Curie anatomic region, and scores were adjusted by the system without requiring additional effort from the imaging physician. A key difference between the methods B and C is that in the latter, the imaging physician need not click on the schematic to assign an anatomic segment to a lesion. Rather, the lesion’s position within the region map automatically determined its anatomic segment. The lesions could be specified at any time during a case, before or after the region map itself was specified, and the region map could likewise be modified as needed.

FIGURE 3.

FIGURE 3

A, In method C, the radiologist first indicates key points that will define the region map (“mesh”). The radiologist clicks on 9–11 points on the anterior image (in blue) that correspond to the points shown on the skeleton image in the upper right corner of the screen, with the left elbow and left fingertip points being optional. B, After the radiologist specifies the key points, the region map (“mesh”) is created, which outlines the Curie anatomic segments for this case, and the radiologist can then adjust the region map if necessary by moving the circular handles between each anatomic segment. C, After the mesh has been created, the radiologist can draw lesions (as points, lines, or loops), and the system automatically assigns anatomic segments and scores based on drawn lesion locations. See Supporting Information Figure S2 for text in the figure

2.6 |. Comparison of the Curie scoring methods A, B, and C

Imaging physicians participated in three sessions, each separated by at least two weeks to reduce recall bias. The first session involved learning method A and then scoring all 38 scans. The second session involved learning methods B and C and then scoring 19 scans using method B and 19 scans using method C, reversing this for the third session with randomization of scan order. The 19 scans were randomized for each imaging physician, so they all read each scan with each method, but in different sessions. Data collected included scores for each region and the time it took to complete scoring of each scan.

A University of Chicago analyst was physically or virtually present (through Skype or WebEx) with the imaging physicians at each session and provided technical guidance to assist the imaging physicians when needed, being careful not to influence any of the clinical decisions. This guidance included noting and helping users correct technical mistakes in using the interface, occasionally noting anatomic segment omissions, and indicating features of the interface (such as using the color-coded schematic to see which anatomic segments have no lesions assigned) that could help the imaging physicians use the interfaces to their full extent. The guidance was most often needed near the beginning of a session, and observers generally became very proficient and confident in using the interfaces as they progressed in experience. The software is designed to be used unaided.

2.7 |. Survey

An 18-question survey was administered to characterize ease-of-use, clinical utility, and potential adoption of each of the methods (Supporting Information Figure S1). The secure REDCap survey was administered via e-mail to all participating imaging physicians and was completed by all five imaging physicians.

2.8 |. Statistical analysis

ANOVA testing was completed by a senior statistician in University of Chicago’s Center for Research Informatics to analyze variance between imaging physicians and methods. Cohen kappa statistics and weighted Fleiss’ kappa statistics were then calculated to evaluate intraobserver and interobserver reliability. The survey data were analyzed descriptively.

4 |. RESULTS

4.1 |. Patient cohort

Thirty-four patients from the University of Chicago with available MIBG scans were enrolled on this study. Additionally, MIBG scans from 4 patients were obtained from QARC for analysis. All patients had a diagnosis of neuroblastoma except for one with metastatic paraganglioma (Supporting Information Table S1). Twenty-seven University of Chicago patients had metastatic disease. The patients ranged in age from 6 weeks to 22 years (median 3 years). Eleven scans were obtained at initial diagnosis, ten during or after induction chemotherapy, three during or prior to immunotherapy, one at end of therapy, and nine during therapy for relapsed disease. The four patients with scans obtained through QARC have unknown clinical information.

4.2 |. Curie scoring

4.2.1 |. Reliability between methods

We first wanted to study whether the same observer obtained similar results using methods A, B, and C. All MIBG scans were scored by each imaging physician using all three methods (Supporting Information Table S2). Reliability between methods was assessed with Cohen kappa coefficient, a statistic that measures interrater agreement for categorical items. A Cohen kappa greater than 0.6 denotes good agreement and greater than 0.8 suggests very good agreement between methods. Scores of 0, 1, 2, and 3 in each region were considered categorical variables as a score of 3 represents the degree to which a region is involved with disease, not the number of lesions. First analysis was performed assuming that all Curie regions are rated with similar reliability. The kappa statistic between methods A and B was 0.869, indicating very good agreement (Table 1). Similarly, the kappa statistic between methods A and C was 0.847 and between methods B and C was 0.861. To compare all three methods, Fleiss’ kappa was used, which is a statistical measure for assessing the reliability of agreement between a fixed number of raters when assigning categorical ratings. Fleiss’ kappa for methods A, B, and C was 0.702, suggesting substantial agreement among the methods. Similar analyses were performed by Curie region, and kappa statistics were calculated between methods A and B (kappa range, 0.678–0.951), methods A and C (kappa range, 0.618–0.917) and methods B and C (kappa range, 0.689–0.935). These data suggest excellent reliability across methods. The poorest reliability existed in regions seven (the lower arms) and region ten (soft tissue) but even these showed substantial reliability across methods.

TABLE 1.

Comparison of scores between methods

Method A vs B Method A vs C Method B vs C
Total 0.869 0.847 0.861
Region 1 Craniofacial 0.887 0.877 0.909
Region 2 Cervical and thoracic spine 0.867 0.842 0.866
Region 3 Ribs/sternum/clavicles/scapula 0.776 0.774 0.775
Region 4 Lumbar and sacral spine 0.867 0.865 0.832
Region 5 Pelvis 0.887 0.873 0.896
Region 6 Upper arms 0.865 0.858 0.803
Region 7 Lower arms and hands 0.678 0.618 0.737
Region 8 Femurs 0.951 0.917 0.935
Region 9 Lower legs and feet 0.908 0.888 0.932
Region 10 Soft tissue 0.721 0.623 0.689

Cohen kappa statistics are shown comparing total Curie scores and Curie scores for each region comparing each pair of methods. All kappa scores are higher than 0.6, indicating very good across methods.

4.3 |. Interobserver reliability

We then calculated the reliability among observers for each method. Interobserver reliability was calculated using weighted Fleiss’ kappa statistics to show consistency between the imaging physicians (Table 2). The kappa statistic was first calculated assuming that all Curie regions were of similar reliability and was 0.840 for method A, 0.811 for method B, and 0.804 for method C, demonstrating excellent reliability between observers. We then evaluated by Curie region and found kappa statistics ranging from 0.743 to 0.933 for method A, from 0.699 to 0.918 for method B, and from 0.728 to 0.901 for method C. Overall interobserver reliability is excellent, similar to intraobserver reliability.

TABLE 2.

Comparison of scores between observers

Method A Method B Method C
Total 0.840 0.811 0.804
Region 1 Craniofacial 0.792 0.811 0.773
Region 2 Cervical and thoracic spine 0.851 0.834 0.838
Region 3 Ribs/sternum/clavicles/scapula 0.827 0.762 0.756
Region 4 Lumbar and sacral spine 0.773 0.789 0.787
Region 5 Pelvis 0.743 0.724 0.728
Region 6 Upper arms 0.893 0.811 0.768
Region 7 Lower arms and hands 0.851 0.918 0.867
Region 8 Femurs 0.905 0.865 0.834
Region 9 Lower legs and feet 0.933 0.901 0.901
Region 10 Soft tissue 0.834 0.699 0.783

Weighted Fleiss’ kappa statistics are shown comparing total Curie scores and Curie scores for each region for each method. All kappa scores are higher than 0.6, indicating very good interobserver reliability.

4.4 |. Time analysis

On average, reading scans with method B took 72% longer than with method A. Method C took 141% longer than method A and 68% longer than method B. For method A, the time for scoring each scan ranged from 15 seconds to 437 seconds, for method B, the time to analyze each scan ranged from 12 to 737 seconds, and for method C, the time to analyze each scan ranged from 41 to 693 seconds (Supporting Information Table S3). On average, method B took 57 seconds longer per scan than method A, and method C took 53 seconds longer than method B, and 110 seconds longer than method A.

4.5 |. Physician assessment of the Curie scoring methods

All five participating imaging physicians completed the survey (Supporting Information Table S4). The imaging physicians noted how likely they would be to utilize each method (method A: 3 very likely, 1 somewhat likely, 1 unlikely; method B: 4 very likely, 1 somewhat unlikely; method C: 2 very likely, 2 somewhat likely, 1 unlikely). The imaging physicians indicated that their preferred method would be somewhat or very useful for routine MIBG scan reading. The majority specified that the data provided by the semiautomated methods would be very useful for central reviewers evaluating MIBG scans as part of a clinical study. All of the imaging physicians agreed that it would be valuable for the treating oncologists to have information about the response of individual MIBG lesions. Comments provided as part of the survey revealed concern that method C took longer than the other methods, and the added time needed to determine the Curie score limited its utility as a clinical tool. However, one imaging physician commented that a semiautomated method of any kind would be better than current practice. Another reported that these strategies could be incorporated into practice. Four thought it was somewhat or very important to keep a record of the lesions that contributed to the score in each region. Three indicated that the information provided by methods B and C regarding each of the component scores would be useful to oncologists for response assessment and treatment decisions.

5 |. DISCUSSION

In this study, we designed and tested three semiautomated methods for evaluating Curie scores with varying degrees of computer assistance. Method A most closely mimics current practice with assignment of Curie scores to each region by visual inspection, with the system providing a running tally. Method B involves marking MIBG-avid lesions on the scan and assigning them to Curie regions on a schematic figure. Method C involves creation of a mesh to define Curie regions on the image and then the marking of MIBG-avid lesions. We showed it is feasible to utilize a user interface and semiautomated method to apply Curie scores. Furthermore, there was consistency across providers and methods. The intermethod and interobserver reliability is very good when evaluating both total scores and each individual region. The imaging physicians easily learned to use each method, indicating these methods could be broadly employed to aid in assigning Curie scores. Currently, many imaging physicians evaluating MIBG scans provide only a total score. Several studies have demonstrated the scores assigned after induction therapy are prognostic of outcome of high-risk patients treated with prior treatment regimens. Curie scores are included in the recently published INRC response criteria and provide important information for treatment decisions.14 Tracking specific MIBG-positive lesions over time is likely to enhance accurate response assessment, provide additional prognostic information, and may ultimately lead to more informed treatment decision-making. These semiautomated methods have the potential to standardize Curie scoring in clinical practice. Methods B and C took 57 seconds (72%) and 110 seconds (141%), respectively, longer to use than method A. However, our survey results indicated that imaging physicians preferred methods B and C to method A, because these methods enabled longitudinal tracking of lesions. The imaging physicians also noted that a disadvantage of method C was the increased time required when compared with methods A and B.

Analysis of patients enrolled on a previous COG clinical trial conducted from 2001 to 2006 demonstrated that Curie scores following induction therapy were prognostic of outcome in patients with stage 4 high-risk neuroblastoma.7 Extremely poor outcomes were observed for patients with MYCN nonamplified tumors with Curie scores >2 and for patients with MYCN-amplified disease with Curie scores >0. Decarolis and colleagues confirmed the prognostic value of Curie scores >2 following induction and showed that a SIOPEN MIBG score >4 following induction was also associated with inferior outcome.8 These studies highlight the prognostic importance of MIBG scoring. The current manual method of determining Curie scores limits the ability to longitudinally monitor specific lesions or regional disease in a standardized manner. We hypothesized that by integrating computational techniques, Curie scores would be more reliably quantified and specific sites of disease could be accurately assessed for response.

These semiautomated methods represent the first step toward making Curie scoring more consistent. Further automation may be achieved via the process previously reported at the University of Chicago with Tc-99m bone scans.16,17 A computer-aided diagnostic approach was designed to identify differences in scans from multi ple time points using a nonlinear image warping technique. Shiraishi et al created a computational algorithm involving image density normalization and downstream processing to successfully identify new and resolved lesions over time. This method was subsequently found to be beneficial 84.6% of the time and has the potential to significantly aid in the evaluation of Tc-99 bone scans. Although bone scans are no longer used in patients with neuroblastoma, a similar approach could be applied to MIBG scans to help identify very subtle changes in metastatic disease patterns, thus making this modality more quantitative and leading to a more precise prognostication method for children with neuroblastoma. Future versions of the computer-assisted methods will need to adjust for improvements in technology, including SPECT imaging.

To bring these methods to clinical practice, the method must be validated in a larger study by comparing scores obtained using semiautomated scoring to scores given by consensus review of expert readers. A prospective study can be used to determine the feasibility of using a semiautomated method in regular clinical practice. Ultimately, broad utilization of these methods could help to standardize the application of Curie scores and aid in monitoring the response of MIBG-avid neuroblastoma over time.

Supplementary Material

Supplemental Figure S1
Supplemental Figure S2
Supplemental Table S1
Supplemental Table S2
Supplemental Table S3
Supplemental Table S4

ACKNOWLEDGMENTS

This work was supported by NIH CTSA UL1 TR000430 and NICHD-NIGMS T32GM007019. Supported, in part, by NSF REU Award No.1062909. The authors would like to thank Ling Teng, Jason Tam, and Sabina Nilakhe for her assistance in early development of the user interface.

Funding information

National Center for Advancing Translational Sciences, Grant/Award Number: UL1 TR000430; National Institute of Child Health and Human Development, Grant/Award Number: T32GM007019; Division of Undergraduate Education, Grant/Award Number: 1062909

This work was presented previously at Advances of Neuroblastoma Research in 2016, American Society of Clinical Oncology in 2016, and American Society of Clinical Pharmacology and Therapeutics in 2018.

Abbreviations:

COG

Children’s Oncology Group

EFS

event-free survival

INRC

International Neuroblastoma Response Criteria

INRG

International Neuroblastoma Risk Group

MIBG

metaiodobenzylguanidine

QARC

Quality Assurance Review Center

SIOPEN

International Society of Pediatric Oncology European Network

SPECT

single-photon emission computed tomography

Footnotes

CONFLICTS OF INTEREST

Samuel G. Armato III and Roger Engelmann receive royalties and licensing fees for CAD technology through the University of Chicago and Samuel G. Armato III is a consultant for Aduro Biotech, Inc., and Boehringer-Ingelheim.

REFERENCES

  • 1.Park JR, Bagatell R, London WB, et al. Children’s Oncology Group’s 2013 blueprint for research: neuroblastoma. Pediatr Blood Cancer. 2013;60(6):985–993. [DOI] [PubMed] [Google Scholar]
  • 2.Matthay KK, Maris JM, Schleiermacher G, et al. Neuroblastoma. Nat Rev Dis Primers. 2016;2:16078. [DOI] [PubMed] [Google Scholar]
  • 3.Grupp SA, Stern JW, Bunin N, et al. Tandem high-dose therapy in rapid sequence for children with high-risk neuroblastoma. J Clin Oncol. 2000;18(13):2567–2575. [DOI] [PubMed] [Google Scholar]
  • 4.Matthay KK, Reynolds CP, Seeger RC, et al. Long-term results for children with high-risk neuroblastoma treated on a randomized trial of myeloablative therapy followed by 13-cis-retinoic acid: a children’s oncology group study. J Clin Oncol. 2009;27(7):1007–1013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Pinto NR, Applebaum MA, Volchenboum SL, et al. Advances in risk classification and treatment strategies for neuroblastoma. J Clin Oncol. 2015;33(27):3008–3017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Matthay KK, Edeline V, Lumbroso J, et al. Correlation of early metastatic response by 123I-metaiodobenzylguanidine scintigraphy with overall response and event-free survival in stage IV neuroblas toma. J Clin Oncol. 2003;21(13):2486–2491. [DOI] [PubMed] [Google Scholar]
  • 7.Yanik GA, Parisi MT, Shulkin BL, et al. Semiquantitative mIBG scor ing as a prognostic indicator in patients with stage 4 neuroblas toma: a report from the children’s oncology group. J Nucl Med. 2013;54(4):541–548. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Decarolis B, Schneider C, Hero B, et al. Iodine-123 metaiodoben zylguanidine scintigraphy scoring allows prediction of outcome in patients with stage 4 neuroblastoma: results of the Cologne interscore comparison study. J Clin Oncol. 2013;31(7):944–951. [DOI] [PubMed] [Google Scholar]
  • 9.Yanik GA, Parisi MT, Naranjo A, et al. Validation of postinduc tion curie scores in high-risk neuroblastoma: a Children’s Oncology Group and SIOPEN group report on SIOPEN/HR-NBL1. J Nucl Med. 2018;59(3):502–508. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Ady N, Zucker JM, Asselain B, et al. A new 123I-MIBG whole body scan scoring method–application to the prediction of the response of metastases to induction chemotherapy in stage IV neuroblastoma. Eur J Cancer. 1995;31a(2):256–261. [DOI] [PubMed] [Google Scholar]
  • 11.Matthay KK, Shulkin B, Ladenstein R, et al. Criteria for evaluation of disease extent by (123)I-metaiodobenzylguanidine scans in neuroblas toma: a report for the International Neuroblastoma Risk Group (INRG) task force. Br J Cancer. 2010;102(9):1319–1326. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Katzenstein HM, Cohn SL, Shore RM, et al. Scintigraphic response by 123I-metaiodobenzylguanidine scan correlates with event-free survival in high-risk neuroblastoma. J Clin Oncol. 2004;22(19):3909–3915. [DOI] [PubMed] [Google Scholar]
  • 13.Yanik GA, Parisi MT, Naranjo A, et al. Validation of post-induction Curie scores in high risk neuroblastoma. A Children’s Oncology Group (COG) and SIOPEN group report on SIOPEN/HR-NBL1. J Nucl Med. 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Park JR, Bagatell R, Cohn SL, et al. Revisions to the international neuroblastoma response criteria: a consensus statement from the National Cancer Institute clinical trials planning meeting. J Clin Oncol. 2017;35(22):2580–2587. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Starkey ASW, Armato SG III. Abras: a portable application for observer studies and visualization. Int J Comp Assist Radiol Surg. 2011;6(suppl. 1): S193–S195. [Google Scholar]
  • 16.Shiraishi J, Li Q, Appelbaum D, Pu Y, Doi K, Development of a computer-aided diagnostic scheme for detection of interval changes in successive whole-body bone scans. Med Phys. 2007;34(1):25–36. [DOI] [PubMed] [Google Scholar]
  • 17.Shiraishi J, Appelbaum D, Pu Y, Engelmann R, Li Q, Doi K, Clinical utility of temporal subtraction images in successive whole-body bone scans: evaluation in a prospective clinical study. J Digit Imaging. 2011;24(4):680–687. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Figure S1
Supplemental Figure S2
Supplemental Table S1
Supplemental Table S2
Supplemental Table S3
Supplemental Table S4

RESOURCES