Skip to main content
PLOS One logoLink to PLOS One
. 2025 May 15;20(5):e0323571. doi: 10.1371/journal.pone.0323571

Assessing newborn scoring with each resuscitation (ANSWER): Protocol for identifying and testing an Apgar score for the 21st century

Henry J Rozycki 1,*, Elizabeth E Foglia 2, Miheret S Yitayew 1, Heidi M Herrick 2
Editor: Abhishek Makkar3
PMCID: PMC12080839  PMID: 40373048

Abstract

Background

The status of infants in the minutes after birth has been summarized by the Apgar score for the last 70 years and is applied at most medically attended deliveries around the world. It has not, however, adapted to changes in neonatal resuscitation over the decades. There are issues with application to premature newborns, how to account for the interventions outlined in the Newborn Resuscitation Program (NRP), inter-rater reliability, and local custom. Developing a modern newborn assessment score will require a series of steps, the first of which is to identify which observations or data can best differentiate a normal newborn transition from an abnormal one.

Methods

Video recordings of at least 35 random normal (meeting NRP goals for heart rate and saturation without intervention, normal physical exam and requiring only normal postnatal care) and 35 random abnormal (not meeting at least 2 normal criteria) from 7 centers participating in the VERIFI study. The presence or absence of observable elements, including vital signs, appearance, and interventions will be recorded every 30 seconds for the first 5 minutes of life. Observations, as well as any changes over time or after intervention will be compared between normal and abnormal, and those that are significant and independent based on logistic regression will become candidate newborn assessment score components.

Expected results

There are likely to be ten or more observations/elements from videos of the first five minutes of life that will differ between normal and abnormal newborns, and these will be tested in all combinations to identify the 1–3 score sets that will be applied to a new set of VERIFI videos to identify which has the best sensitivity/specificity.

Conclusions

Bringing newborn assessment into modern practice will build on the legacy of Dr. Apgar. A series of steps, beginning with identifying the observations/elements that best identify newborns who may need further care can lead to a universal validated tool for the 21st century.

Introduction

The status of almost every baby born in the US, Europe and most other countries is assessed over the first five minutes of life using the system that Virginia Apgar published almost 70 years ago [1]. Along with the date of birth, it is the first number assigned to millions each year. The initial purpose, according to Dr. Apgar, was to identify newborns with difficult transitions to extrauterine life who might therefore require further support, but it quickly evolved to become a risk factor/predictor for mortality [2] and for near [3]- and long-term [4] morbidities for the baby. Over the next decades, as it became almost universal, it also became an outcome marker for problems during pregnancy and delivery. In a review of 500 papers published in 2018–19, the Apgar score was used to assess issues and interventions in pregnancy (35%), mode of delivery and maternal anesthesia (30%), as a risk factor for short and long-term problems for the newborn (19%), and as a descriptor of the newborn’s condition (16%) [5].

The Collaborative Perinatal Project included the Apgar score among the prospective collected data on over 50,000 pregnancies from 1959–1965 and demonstrated an association between low scores and both mortality [2] and one-year neurodevelopmental problems [4]. Obstetrical care, newborn care, and newborn resuscitation have markedly changed since then, but no contemporary studies have examined the impact of these changes on the utility of the Apgar score. There are also major concerns regarding its prognostic accuracy [6], application to premature newborns [7], inter-rater reliability [8,9], how to account for ongoing resuscitative interventions [10], and applicability across different countries and regions [11]. Accordingly, there are increasing calls to overhaul newborn scoring to make it more relevant or even to abolish it [1214].

Given how a simple observational scoring system for all newborns is almost universally applied at delivery and used by so many medical (and non-medical) specialties, a newborn assessment methodology for the 21st century must be 1) simple to apply, 2) informed by easily obtained data, primarily from observation at the assessment time, 3) adaptable to any circumstance at delivery, and 4) consistent across observers and local practices.

Revising the approach to newborn assessment will require a series of steps, including reassessment of relevant inputs, feasibility testing, and real-world comparisons with the current gold standard, the Apgar score. The first step is to determine which potential components of an assessment tool best differentiate a normal and successful transition to extrauterine life from one that is not normal and/or identifies the newborns who require increased medical attention at delivery or after postnatal transition. The Assessing Newborns With Each Resuscitation (ANSWER) study is designed to create a hierarchy of candidate newborn assessment components using video recordings of newborn resuscitation.

Design

The data for this study will be collected from analysis of video recordings as part of the VERIFI (Video Evaluation for Resuscitation of Infants International) project, an international registry of data captured from video recorded delivery room resuscitations to characterize practice, process and outcomes of delivery room resuscitation. The goal is to collect 300–450 videos per year and follow the subjects until hospital discharge or transfer. Currently, seven centers are preparing to participate: the Hospital of the University of Pennsylvania, Oklahoma Children’s Hospital OU Health, Stavanger University Hospital, Universitätskliniken der MedUni Wien, St Louis University/Cardinal Glennon Children’s Hospital, Technisches University Dresden, and Universiteit Leiden. At each center, video- and audio-recordings of the newborn are captured in the first minutes of life. The videos are stored within each site per their individual policies and only anonymized data derived from analysis of the videos done at each site are uploaded to the registry. The ANSWER study will leverage the VERIFI infrastructure to assess the feasibility and inter-rater reliability of candidate measures of newborn health that can be ascertained from a convenience sample of these video recordings. Two raters at each site will independently review and score each video for each of the candidate elements listed in Table 1 and to classify each delivery outcome as Normal, Abnormal, or Indeterminate. Time to first cry and breath will be recorded at the exact time. The other elements will be recorded at every 30 seconds after birth for the first 5 minutes. Observations of the newborn such as heart rate, saturation, and elements derived from the Apgar score will be recorded as close to but preceding each 30-second mark. Intervention elements will be scored if they occurred in the thirty second time preceding the assessment rime. For example, if CPAP is applied at 75 seconds of life, and removed at 185 seconds of life, it would recorded as Yes/Present only at 90, 120, 150 and 180 seconds. If an element cannot be accurately assessed at any of the successive 30 second marks, it will be recorded as Unobservable.

Table 1. Candidate scoring elements from resuscitation videos.

Observation/Element Definition
Time to first breath Record exact time (e.g., 17 seconds)
Time to first cry Record exact time (e.g., 17 seconds)
Elements assessed every 30 seconds of life to 5 minutes
MEASURES
O2 saturation Record actual number at or closest to but prior to each 30 second epoch.
FiO2 Maximum during the preceding 30 second epoch.
Heart Rate Record actual number at or closest to but prior to each 30 second epoch. Record method heart rate was ascertained (e.g., auscultation, pulse oximeter, ECG)
NEWBORN STATUS
Color Assessed at each 30 second mark:
Blue or Pale – face/body
Acrocyanotic – hands and/or feet
Pink – face/body/extremities
Respirations Apnea/No chest movement
Gasping/severe retractions
Inadequate/small/irregular
Adequate/Normal/Crying
Tone Limp/Flaccid
Some flexion/reduced for gestational age
Active motion/appropriate for gestational age
Neuro No reflex response/staring
Incomplete or reduced for gestational age response
Normal/appropriate for gestational age response
Activity [21] Absent
Flexed arms/legs
Active
Grimace [21] No response
Minimal response
Prompt response
INTERVENTIONS
CPAP Y/N administered during the preceding 30 second epoch
Oxygen Delivery Y/N FiO2 > 21% delivered during preceding 30 second epoch.
Non-invasive Positive Pressure Ventilation Y/N administered during the preceding 30 second epoch.
Endotracheal tube Y/N in place during the preceding 30 second epoch
Surfactant Y/N administered in preceding 30 second epoch
LMA Y/N in place during the preceding 30 second epoch
Catecholamines administered Y/N administered in preceding 30 second epoch
Chest compressions Y/N administered in preceding 30 second epoch

Components will be dichotomized (yes/no). In the case of continuous variables, this will be done for a series of cut-offs (e.g., for heart rate, 0, < 50 bpm, < 60 bpm, < 80 bpm, < 100 bpm). For graded variables, each grade will be scored as yes/no (e.g., for Apgar score tone, which has 3 grades, there will be 3 potential yes/no components). An additional set of components will be derived from changes in a particular measure or observation and the time that the change took (e.g., a subject with heart rate <60 at 90 secs, < 90 at 120 secs and > 100 at 150 secs will be coded as rise/inadequate 120, rise/adequate 150) to see if recovery can help differentiate normal from abnormal transition. These improvements will also be separately coded as spontaneous or after intervention.

Clinical data accompanying each video will include the estimated gestational age at delivery, sex, mode of delivery, and birthweight rounded to 250g increments. Videos will be categorized into Normal (all 5 criteria met), Abnormal (≤ 2 criteria met) or Indeterminant (3–4 criteria met) using the checklist below. Only normal and abnormal videos will be used for this study:

  • a) preductal oxygen saturation measurements meeting NRP goals without supplementation, if measured

  • b) Heart rate > 100 throughout postnatal transition.

  • c) no apnea or respiratory distress.

  • d) no resuscitation interventions.

  • e) normal postnatal care. Note that for preterm infants, this will be defined as discharge at ≤ 40 weeks estimated gestational age, with a normal neurological exam, in room air, and normal oral breast or formula feedings, and without diagnoses of necrotizing enterocolitis, bronchopulmonary dysplasia, bacterial infection, or intraventricular hemorrhage > Grade 1.

Exclusion criteria will include major congenital anomalies. Prematurity will not exclude participation and will analyzed as a covariate.

Goal recruitment is at least 5 normal and 5 abnormal video analyses from each site.

The overall study was reviewed by the Virginia Commonwealth University Institutional Review Board and consenting waived because of the complete anonymization of the data, such that only any consent documentation would potentially identify a subject. Each participating site will obtain approval to participate and contribute anonymous data from their individual institutional review boards.

Analysis

The dichotomized variables at each of the ten time points between 30 seconds and 5 minutes of life, will be counted from the normal and the abnormal videos, which should result in several hundred potential assessment tool components for evaluation. In addition, variables that permit calculation of change over time (e.g., heart rate at 150 seconds vs 30 seconds) or the effect of an earlier intervention on a component will be calculated and included in the set of potential score elements. The significance and power of each element will be calculated by Fisher’s Exact Test from the proportions of each groups. Any element with a power < 0.8 will be discarded from further analysis. The individual elements will then be arranged into a hierarchy based on discriminant difference, i.e., inverse p value. Because some elements are likely to be closely related, multiple logistic regression analysis will be performed for each time point using the elements defined as significant from the proportional analysis. Those that remain both significant and independent at each time point will then be tested individually and in their various combinations for their significance between the normal and abnormal deliveries. For example, if four components make the grade for inclusion at 1 minute, analysis will be performed on each individual component, on all possible combinations of two, all combinations of three and all combinations of all four components. The ten combinations with the most significant difference will be designated as candidate newborn assessment tools to compare against the Apgar in future studies.

Timeline

We plan to begin data abstraction from newborn resuscitation videos in the first quarter of 2025 and anticipate completing the initial 70 from the 7 sites within 6 months. The analysis and determination of the best candidate newborn assessment elements, the first step in the overall ANSWER program, should be completed by December 2025. The sites will continue to record videos for the VERIFI study, and these will be used during the first half of 2026 to test and validate a new assessment tool based on the best elements from the first step. In particular, a larger video data set from the VERIFI study will permit a further analysis of any issue with prematurity if the first phase does not have a sufficient number and range of premature deliveries.

Discussion

For any attempt to develop or validate a newborn assessment tool or method, a short-term outcome will need to be identified. Dr. Apgar, in her original paper [1], showed that her score was lower in those babies who died in the first 28 days, but in many countries, this is an uncommon outcome and would necessitate a very large sample size. All recent attempts to modify or replace the Apgar have used different short-term outcomes [1517]. We suggest using the assessment score to identify newborns who could benefit from enhanced observation or care after delivery, as opposed to routine care. This would include virtually all those who have conditions and consequences related to perinatal events without so many false positives as to overwhelm any systems. In the case of preterm infants, who generally go from delivery to a special or intensive care nursery, the definition would be expanded to include care not directly related to prematurity.

The second step will be to apply the Apgar score and the new assessment tools identified in the ANSWER protocol outlined above to the videos deemed Indeterminate and to the other 150 + resuscitation videos in the VERIFI registry to define their sensitivities and specificities across a broader range of deliveries. Next, the best tool sets (up to 3) will be tested for feasibility/ease of use by applying them at delivery along with the Apgar score in a single center over 6–12 months, with a goal of > 1000 deliveries. The data from this clinical trial will also be used to select the best assessment tool to test against the Apgar in large multicenter trials.

Any newborn assessment system will need to demonstrate its superiority to the Apgar score for identifying newborns with difficult or abnormal transitions to life and/or needing postnatal attention, as well as meet the basic principles outlined above. Appropriate clinical studies will include online training and testing for scorers, multiple centers to make sure that local practices do not influence outcomes, and multiple countries and regions to guard against the development of regional differences in the application of a newborn assessment program. The approach would be similar to those employed during the clinical studies to determine the best oxygen saturation goals for preterm infants, SUPPORT [18], COT [19], and BOOST II [20], where national and regional multicenter networks organized and funded the work.

This stepwise approach has several advantages. Components of a newborn assessment will reflect current practice, including fundamental interventions and will utilize video recording that can permit a reviewable analysis of a potential component. Candidate scores will be evaluated in single center tests of feasibility and utility, increasing the likelihood that the final selection for the larger international trials will be successful. Finally, by using the same definitions and protocols, data from several real-world international trials can be combined to produce sufficient power to validate a newborn scoring system for rare but critical outcomes such as long-term neurological injury.

Data Availability

No datasets were generated or analysed during the current study. All relevant data from this study will be made available upon study completion.

Funding Statement

The project described was supported by the National Center for Advancing Translational Sciences (NCATS), National Institutes of Health, through Grant Award Number 5R21TR003994 (HJR and EEF). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Apgar V. A proposal for a new method of evaluation of the newborn infant. Curr Res Anesth Analg. 1953;32(4):260–7. [PubMed] [Google Scholar]
  • 2.Drage JS, Kennedy C, Schwarz BK. The APgar score as an index of neonatal mortality. a report from the collaborative study of cerebral palsy. Obstet Gynecol. 1964;24:222–30. [PubMed] [Google Scholar]
  • 3.Lester BM, Emory EK, Hoffman SL. A multivariate study of the effects of high-risk factors on performance on the Brazelton Neonatal Assessment Scale. Child Dev. 1976;47(2):515–7. doi: 10.2307/1128811 [DOI] [PubMed] [Google Scholar]
  • 4.Drage JS, Kennedy C, Berendes H, Schwarz BK, Weiss W. The Apgar score as an index of infant morbidity. A report from the collaborative study of cerebral palsy. Dev Med Child Neurol. 1966;8(2):141–8. doi: 10.1111/j.1469-8749.1966.tb01719.x [DOI] [PubMed] [Google Scholar]
  • 5.Rozycki HJ, Yitayew M. The Apgar score in clinical research: for what, how and by whom it is used. J Perinat Med. 2022;51(4):580–5. doi: 10.1515/jpm-2022-0340 [DOI] [PubMed] [Google Scholar]
  • 6.Bovbjerg ML, Dissanayake MV, Cheyney M, Brown J, Snowden JM. Utility of the 5-Minute Apgar Score as a Research Endpoint. Am J Epidemiol. 2019;188(9):1695–704. doi: 10.1093/aje/kwz132 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Bashambu MT, Whitehead H, Hibbs AM, Martin RJ, Bhola M. Evaluation of interobserver agreement of apgar scoring in preterm infants. Pediatrics. 2012;130(4):e982-7. doi: 10.1542/peds.2012-0368 [DOI] [PubMed] [Google Scholar]
  • 8.O’Donnell CPF, Kamlin COF, Davis PG, Carlin JB, Morley CJ. Interobserver variability of the 5-minute Apgar score. J Pediatr. 2006;149(4):486–9. doi: 10.1016/j.jpeds.2006.05.040 [DOI] [PubMed] [Google Scholar]
  • 9.Jurdi SR, Jayaram A, Sima AP, Hendricks Muñoz KD. Evaluation of a Comprehensive Delivery Room Neonatal Resuscitation and Adaptation Score (NRAS) Compared to the Apgar Score: A Pilot Study. Glob Pediatr Health. 2015;2:2333794X15598293. doi: 10.1177/2333794X15598293 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Lopriore E, van Burk GF, Walther FJ, de Beaufort AJ. Correct use of the Apgar score for resuscitated and intubated newborn babies: questionnaire study. BMJ. 2004;329(7458):143–4. doi: 10.1136/bmj.38117.665197.F7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Siddiqui A, Cuttini M, Wood R, Velebil P, Delnord M, Zile I, et al. Can the Apgar Score be Used for International Comparisons of Newborn Health? Paediatr Perinat Epidemiol. 2017;31(4):338–45. doi: 10.1111/ppe.12368 [DOI] [PubMed] [Google Scholar]
  • 12.Marlow N. Do we need an Apgar score? Arch Dis Child. 1992;67(7 Spec No):765–7. doi: 10.1136/adc.67.7_spec_no.765 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Rüdiger M, Rozycki HJ. It’s Time to Reevaluate the Apgar Score. JAMA Pediatr. 2020;174(4):321–2. doi: 10.1001/jamapediatrics.2019.6016 [DOI] [PubMed] [Google Scholar]
  • 14.Niemuth M, Küster H, Simma B, Rozycki H, Rüdiger M, European Society for Paediatric Research (ESPR) Neonatal Resuscitation Section Writing Group, et al. A critical appraisal of tools for delivery room assessment of the newborn infant. Pediatr Res. 2024;96(3):625–31. doi: 10.1038/s41390-021-01896-7 [DOI] [PubMed] [Google Scholar]
  • 15.Rüdiger M, Braun N, Aranda J, Aguar M, Bergert R, Bystricka A, et al. Neonatal assessment in the delivery room--Trial to Evaluate a Specified Type of Apgar (TEST-Apgar). BMC Pediatr. 2015;15:18. doi: 10.1186/s12887-015-0334-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Dalili H, Nili F, Sheikh M, Hardani AK, Shariat M, Nayeri F. Comparison of the four proposed Apgar scoring systems in the assessment of birth asphyxia and adverse early neurologic outcomes. PLoS One. 2015;10(3):e0122116. doi: 10.1371/journal.pone.0122116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Witcher TJ, Jurdi S, Kumar V, Gupta A, Moores RR Jr, Khoury J, et al. Neonatal Resuscitation and Adaptation Score vs Apgar: newborn assessment and predictive ability. J Perinatol. 2018;38(11):1476–82. doi: 10.1038/s41372-018-0189-5 [DOI] [PubMed] [Google Scholar]
  • 18.SUPPORT Study Group of the Eunice Kennedy Shriver NICHD Neonatal Research Network, Carlo WA, Finer NN, Walsh MC, Rich W, Gantz MG, et al. Target ranges of oxygen saturation in extremely preterm infants. N Engl J Med. 2010;362(21):1959–69. doi: 10.1056/NEJMoa0911781 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Schmidt B, Whyte RK, Asztalos EV, Moddemann D, Poets C, Rabi Y, et al. Effects of targeting higher vs lower arterial oxygen saturations on death or disability in extremely preterm infants: a randomized clinical trial. JAMA. 2013;309(20):2111–20. doi: 10.1001/jama.2013.5555 [DOI] [PubMed] [Google Scholar]
  • 20.BOOST II United Kingdom Collaborative Group, BOOST II Australia Collaborative Group, BOOST II New Zealand Collaborative Group, Stenson BJ, Tarnow-Mordi WO, Darlow BA, et al. Oxygen saturation and outcomes in preterm infants. N Engl J Med. 2013;368(22):2094–104. doi: 10.1056/NEJMoa1302298 [DOI] [PubMed] [Google Scholar]
  • 21.Tuttle Z, Roberts C, Davis P, Malhotra A, Tan K, Bhatia R, et al. Combining activity and grimace scores reflects perinatal stability in infants <32 weeks gestational age. Pediatr Res. 2024;96(4):983–9. doi: 10.1038/s41390-024-03130-6 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Abhishek Makkar

26 Feb 2025

Dear Dr. Rozycki,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Please address the comments raised by both reviewers.

Please submit your revised manuscript by Apr 12 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Abhishek Makkar, M.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following financial disclosure: 

 [copy in funding statement].  

Please state what role the funders took in the study.  If the funders had no role, please state: ""The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."" 

If this statement is not correct you must amend it as needed. 

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

3. The project described was supported by the National Center for Advancing Translational Sciences (NCATS), National Institutes of Health, through Grant Award Number 5R21TR003994 (HJR and EEF). 

4. Please provide a complete Data Availability Statement in the submission form, ensuring you include all necessary access information or a reason for why you are unable to make your data freely accessible. If your research concerns only data provided within your submission, please write "All data are in the manuscript and/or supporting information files" as your Data Availability Statement.

5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well. 

6. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Does the manuscript provide a valid rationale for the proposed study, with clearly identified and justified research questions?

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Is the protocol technically sound and planned in a manner that will lead to a meaningful outcome and allow testing the stated hypotheses??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Is the methodology feasible and described in sufficient detail to allow the work to be replicable??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors described where all data underlying the findings will be made available when the study is complete??>

The PLOS Data policy

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: No

**********

Please use the space provided to explain your answers to the questions above and, if applicable, provide comments about issues authors must address before this protocol can be accepted for publication. You may also include additional comments for the author, including concerns about research or publication ethics.

You may also provide optional suggestions and comments to authors that they might find helpful in planning their study.

Reviewer #1: The abstract states video recordings for 25 normal and 25 abnormal resuscitations. However, in the Methods, the authors list seven participating sites with five normal and five abnormal recordings from each site, which would be a total of 35 videos each. Please clarify.

When describing the categorization of videos, it might be clearer to state normal is 4 of more criteria met, abnormal

is 2 or few criteria met, indeterminant is 3 criteria met. Also, must criteria (e) normal postnatal care be met for that delivery to be labeled "normal"? It seems criteria a-d could be met, but if the infant develops respiratory distress after 5 minutes and goes to the NICU, that might not be considered "normal".

Are there exclusion criteria for the candidate resuscitation videos that will be reviewed for the study, e.g., major

congenital anomalies? If so, please include.

The authors mention that applying the Apgar score to preterm deliveries can be challenging. Is there a plan to

ensure infants with various degrees of prematurity (eg extreme, very, moderate, late) are included in both the normal and abnormal groups?

Reference #3, the year looks incorrect.

Table 1, Newborn Status, Color is listed twice.

Reviewer #2: Congrats to the authors for working on a new tool for newborn assessment.

Few suggestions are

1. some punctuation and typographical errors in the manuscript that needs attention. for example (line 30 inter-rater reliability should be followed by a comma before the and) example line 95 has two full stops)

Also the authors also have not mentioned about how they would account for the practice difference across centers. For instance in some centers the first step in evaluating would be by auscultation where as some centers prefer EKG leads/ Dry electrodes. Are these going to be standardized across centers.

How would the 25 normal and abnormal ones be selected from the cohort is it a random selection of infants and then categorize them to normal and abnormal.

Are there any measures to control for inter-observer variability how many people would review those videos to determine normal and abnormal.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org

PLoS One. 2025 May 15;20(5):e0323571. doi: 10.1371/journal.pone.0323571.r003

Author response to Decision Letter 1


7 Mar 2025

These are the specific responses/changes to the review and comments on our above-referenced manuscript.

1. The document has been formatted to meet the referenced specifications, including author names/affiliations

2. Funding Statement has been amended in the revised cover letter to include the wording from the review.

3. See #2

4. Data availability statement now indicates “All data are in the manuscript and/or supporting information files.” as requested.

5. A paragraph regarding IRB approval and consent waiver is now included (Lines 133-137)

6. References are checked and corrected

Reviewer #1

Reviewer #1: The abstract states video recordings for 25 normal and 25 abnormal resuscitations. However, in the Methods, the authors list seven participating sites with five normal and five abnormal recordings from each site, which would be a total of 35 videos each. Please clarify.

Abstract corrected to match 35 videos in each group.

When describing the categorization of videos, it might be clearer to state normal is 4 of more criteria met, abnormal is 2 or few criteria met, indeterminant is 3 criteria met. Also, must criteria (e) normal postnatal care be met for that delivery to be labeled "normal"? It seems criteria a-d could be met, but if the infant develops respiratory distress after 5 minutes and goes to the NICU, that might not be considered "normal".

We are using all 5 criteria, including post-natal course in this phase of the score development for 2 reasons. First, one cannot know, based on the video of the first minutes of life, why a baby was admitted to a higher level of care, and there is a possibility that it was due to something that might have affected an observation in the delivery room. Including post-natal course aims keep the ‘normal population’ as clean as feasible. Second, one of the potential uses for a newborn assessment score is to identify newborns who would benefit from a higher level of post-natal care or observation, so the ‘normal’ group should not include any who do receive it.

The divisions are now defined as Normal – all 5, Abnormal - 2 or less and Indeterminant - 3 – 4, and that only Normal and Abnormal will be used. (Lines 117-119)

Are there exclusion criteria for the candidate resuscitation videos that will be reviewed for the study, e.g., major congenital anomalies? If so, please include

Sentence regarding exclusion for major congenital anomalies now added (Line 130-31)

The authors mention that applying the Apgar score to preterm deliveries can be challenging. Is there a plan to ensure infants with various degrees of prematurity (eg extreme, very, moderate, late) are included in both the normal and abnormal groups?

Sentence added to Timeline to address this: “In particular, a larger video data set from the VERIFI study will permit a further analysis of any issue with prematurity if the first phase does not have a sufficient number and range of premature deliveries.” (Lines 162-164)

Reference #3, the year looks incorrect.

Corrected

Table 1, Newborn Status, Color is listed twice.

Corrected

Reviewer #2: Congrats to the authors for working on a new tool for newborn assessment.

Few suggestions are

1. some punctuation and typographical errors in the manuscript that needs attention. for example (line 30 inter-rater reliability should be followed by a comma before the and) example line 95 has two full stops)

Corrected

Also the authors also have not mentioned about how they would account for the practice difference across centers. For instance in some centers the first step in evaluating would be by auscultation where as some centers prefer EKG leads/ Dry electrodes. Are these going to be standardized across centers.

One of the requirements for a universal newborn assessment score is that it be as universally applicable as possible so we would like to not specify how to perform in the delivery room. Observations will mainly be of the baby, and practice differences may be accounted for by the elements that measure the babies responses to interventions.

How would the 25 normal and abnormal ones be selected from the cohort is it a random selection of infants and then categorize them to normal and abnormal.

As stated, at each site, 2 people will review each video used for the VERIFI study to a) define them as normal/abnormal/indeterminate, and b) abstract the data for VRIFI and ANSWER. For this part of developing potential scoring systems, what is important is that the delivery are either normal or abnormal. Subsequent testing and validation of any score system will need to include randomization, blinding, etc. Thus, for the study protocol defined in this paper, a convenience sample of videos will be used. This now specified in Line 98

Are there any measures to control for inter-observer variability how many people would review those videos to determine normal and abnormal.

Each site has 2 reviewers for each video. Prior to starting the study, a sample video was sent to all reviewers at all sites to verify that they scored/assessed criteria equally. Furthermore, for purposed of this ANSWER PROTOCOL, reviewers are instructed to accept criteria for defining normal when they are completely sure. Combined with the objectivity of most of the criteria, (a) preductal oxygen saturation measurements meeting NRP goals without supplementation, if measured (meets objective goal or not) b) Heart rate > 100 throughout postnatal transition (meets objective goal or not) c) no apnea or respiratory distress (instructions to only accept if no question of symptoms). d) no resuscitation interventions. (direct observation of video) e) normal postnatal care (medical history) we feel this will make inter-rater reliability for defining normalcy excellent. The issue will be much more important when it comes to the individual elements in the next phases

Attachment

Submitted filename: Rebuttal Letter.docx

pone.0323571.s002.docx (19.2KB, docx)

Decision Letter 1

Abhishek Makkar

10 Apr 2025

Assessing newborn scoring with each resuscitation (ANSWER): Protocol for identifying and testing an Apgar score for the 21st century

PONE-D-24-50763R1

Dear Dr. Rozycki,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Abhishek Makkar, M.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Abhishek Makkar

PONE-D-24-50763R1

PLOS ONE

Dear Dr. Rozycki,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Abhishek Makkar

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Rebuttal Letter.docx

    pone.0323571.s002.docx (19.2KB, docx)

    Data Availability Statement

    No datasets were generated or analysed during the current study. All relevant data from this study will be made available upon study completion.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES