Skip to main content
Frontiers in Public Health logoLink to Frontiers in Public Health
. 2026 Feb 2;14:1725392. doi: 10.3389/fpubh.2026.1725392

Exploring long-term follow-up of effective implementation trials in schools: a secondary review

Carly Gardner 1,2,3,4,*, Alix Hall 1,2,3,4, Cassandra Lane 1,2,3,4, Alison Zucca 1,2,3,4, Sam McCrabb 1,2,3,4, Edward Riley-Gibson 1,2,3,4, Xiao Tian Loh 1,5, Katherine Farragher 4, Rachel Sutherland 1,2,3,4, Nicole Nathan 1,2,3,4
PMCID: PMC12908588  PMID: 41705046

Abstract

Background

To reduce chronic diseases, evidence-based health promotion interventions (EBIs) must be effectively implemented and sustained in settings such as schools. This review assessed the extent to which EBIs sustained their effects following the completion of an effective implementation trial. It also explored the use of recommended sustainment practices in follow-up studies.

Materials and methods

A Cochrane systematic review served as the basis for identifying school health promotion EBIs with demonstrated implementation effectiveness. Eligible studies were controlled trials in elementary or secondary schools that evaluated an implementation intervention and reported a statistically significant implementation effect. Forward citation searches were conducted across three electronic databases and two trial registration databases to identify relevant follow-up studies. To be included, follow-up studies needed to report comparable quantitative data (i.e., similar samples and outcome measures) to the original trial and to have been collected at least 6 months after the intervention ended. Two independent reviewers screened studies and extracted data using a predetermined template, with input from a third reviewer as needed. The included papers were assessed against sustainment-promoting practices, including whether they planned for, defined, or used a sustainment framework. The percentage of the original implementation effect sustained at follow-up was calculated.

Results

Of the 23 EBIs with a significant implementation effect in the original review, 26% (n = 6) were found to have eligible follow-up studies. These targeted physical activity (n = 3), diet (n = 2), and tobacco prevention (n = 1). Four studies reported implementation outcomes, and two reassessed behavioral outcomes. The percentage of the implementation effect sustained ranged from 51 to 122%, with a median of 76% (IQ1 = 56%, IQ3 = 107%).

Conclusion

To the best of our knowledge, this is the first review to quantify the sustainment of EBIs in schools following an effective implementation trial and to assess the extent to which best-practice sustainment principles were applied. Few studies described sustainment planning or used comparable follow-up measures. Improving long-term outcome measurement and integrating sustainment frameworks into planning and implementation could help sustain effective school health promotion EBIs.

Keywords: chronic disease prevention, EBI sustainment, effective implementation interventions, health promotion, implementation science, schools

Introduction

Globally, chronic diseases account for approximately 75% of deaths each year (1). Many of the behavioral risk factors that contribute to these diseases, such as poor diet and physical inactivity, are established in childhood and track into adulthood (2). Schools are an important and convenient setting for promoting healthy behaviors in children because they offer consistent access to the majority of the population and reach children from diverse backgrounds who may face barriers to other health services. As such, schools are frequently tasked with delivering health promotion and evidence-based interventions (EBIs). Despite their potential, schools face several challenges when implementing health promotion EBIs, including competing curriculum demands, limited resources, and staff turnover (3, 4), which reduce the likelihood that programs will be implemented as intended, thereby undermining their effectiveness and long-term impact (5).

Over the past few decades, the field of implementation science has developed and tested strategies to address barriers that limit implementation (6, 7), including school-based health promotion EBIs (6–12). A 2022 Cochrane Systematic Review identified 38 studies evaluating strategies to improve the school-based implementation of EBIs targeting healthy eating, physical activity, weight gain, tobacco use, or alcohol use. Of these, 22 were randomized controlled trials (RCTs) that demonstrated significant positive effects on implementation outcomes [standardized mean difference (SMD): 1.04, 95% CI: 0.74–1.34; 1,917 participants]. While these findings highlight the potential for implementation support to improve adoption, their value depends on whether schools continue to deliver the EBI once active implementation support ends, a process often referred to as sustainment (13). Without sustained delivery, initial implementation gains are unlikely to translate into enduring public health outcomes.

To increase the likelihood that EBIs are sustained after implementation support ends, researchers have identified several practices to consider (14–18). These include (i) planning for sustainment during intervention and implementation design phases (14, 15, 17), (ii) explicitly defining the outcomes or domain(s) expected to continue, that is, delivery of the EBI, behavior change or use of implementation supports (14, 17, 18), (iii) applying sustainment specific-frameworks to guide both design and evaluation (13, 17), and (iv) reporting the type, duration, and intensity of any ongoing support provided after active implementation ends (19). Despite their potential to inform sustainability, these practices are rarely applied. A 2019 review of 26 community-based public health EBIs reporting sustainment outcomes found only 19% used a sustainment framework, and 27% provided an explicit definition of sustainment (18). Without consistent application of these practices, it is difficult to identify what supports EBI sustainment or to enable comparison across settings.

Despite growing recognition of its importance, sustainment is rarely assessed in school-based implementation trials. This limits our understanding of the long-term delivery or ongoing health impact of effectively implemented EBIs in school settings. Sustainment research across public health programs indicates it is important to consider both program continuation (sustainment) and the maintenance of behavioral outcomes when exploring a program’s continued benefit (18).

Where follow-up in the school setting has occurred, findings suggest that both EBI delivery and behavioral outcomes often decline once implementation support ends. For example, a 2020 review by Herlitz et al. examined the extent to which school-based EBIs were sustained after initial delivery (20). The review found none of the 18 included EBIs sustained all intervention components; however, 17 EBIs continued at least one component in some schools. While valuable, this review included EBIs regardless of whether they were implemented effectively (demonstrated a positive effect on an implementation measure). Evaluating the sustainment of EBIs without first confirming implementation success risks drawing misleading conclusions about what supports long-term delivery, as sustainment cannot occur without initial implementation. To best inform future efforts, evidence on sustainment should focus on EBIs that have demonstrated successful implementation in real-world conditions (15).

To date, no review has explicitly examined whether EBIs that have been effectively implemented in school settings have also been assessed for sustainment or whether recognized sustainment practices were applied. The 2022 Cochrane review by Wolfenden et al. (21) provides an ideal sampling frame for such an investigation, offering a curated set of EBIs that have been shown to be successfully implemented following an effective implementation intervention. This secondary review builds on that work to:

  1. Determine the proportion of school-based health promotion EBIs that assess sustainment (either program sustainment or sustainability of behavioral outcomes) after the conclusion of an effective implementation trial.

  2. Identify if recommended practices of sustainment are present, including sustainment planning, defining sustainment, use of sustainment frameworks, use of sustainment strategies and level of support, and a threshold for program sustainment (e.g., 80% of sites continued to implement the EBI).

  3. Assess the extent to which implementation effects are sustained following the end of an effective implementation trial.

Materials and methods

This systematic review is reported in accordance with the Preferred Reporting Items for Systematic Review and Meta-Analysis Checklist (PRISMA) (see Supplementary file 1 for the completed checklist). Methods for conducting this review follow Cochrane best-practice guidance on systematic reviews (22).

Eligibility criteria

All included studies from the Wolfenden review (21) were used as a sampling base. We acknowledge the sample review was updated in parallel to undertaking this review (23); however, the previous 2022 version was maintained as the sampling frame for the following reasons: (1) The newer edition restricts study eligibility to RCTs; however, RCT designs are uncommonly used for sustainment research, and thus we deemed the breadth in design eligibility in the 2022 version more appropriate for assessing sustainment; and (2) that additional EBIs identified (in the 2024 version) would unlikely have had time to publish subsequent follow-up studies aligned to our definition of sustainment (i.e., >6 months post implementation end). Studies were conducted in the school setting and targeted health promotion areas, including physical activity, obesity prevention, healthy eating, smoking or vaping cessation, and harmful alcohol use. Participant populations included staff, students, and parents involved in the schools in which studies were conducted.

To be eligible for inclusion, studies had to employ a parallel control group design that included RCTs or cluster RCTs (cRCTs); non-RCTs or cluster non-RCTs with a parallel control group; and controlled before-and-after studies (CBAs) or cluster CBAs.

Eligible trials were checked for a statistically significant implementation effect at the time point immediately following the implementation phase. When the original review reported outcomes at a different time point, the SMD was recalculated using data from the end-of-implementation time point. Statistical significance was taken from the paper (i.e., extracted from the paper) or calculated by the research team for the post-implementation effect (21). In cases where this could not be calculated (e.g., results not reported by the control and intervention groups), an EBI was included if it self-reported a statistically significant effect on the primary implementation measure at the end of implementation.

Those studies with a significant implementation effect were searched and screened for relevant long-term follow-up studies (defined as data accessible to us at a timepoint meeting the eligibility criteria described below, which could include the original trial paper, an additional publication, or data provided upon request). A follow-up study was included if it met the following criteria:

  • Timepoint: Follow-up occurred ≥ 6 months after the end of implementation.

  • Methods and study design: All quantitative study designs were considered, as previous research indicated that follow-up time points are infrequently conducted using controlled designs (24).

  • Measure: Sustainment measured using a comparable implementation or behavioral outcome as reported at the end of implementation in the original study.

  • Sample: Conducted with a comparable sample. This could be the same sample from the original implementation evaluation study or a new group that had received the effective implementation intervention using the same implementation strategies.

A follow-up study was excluded if:

  • Timepoint: Follow-up occurred <6 months after the end of implementation

  • Methods: Qualitative methods only

  • Measure: Not comparable to the implementation effect study.

  • Sample: Conducted with a new sample (e.g., at scale) but used different implementation strategies, not comparable to the implementation effect study, and thus the implementation effect is unknown.

Sustainment was reported based on the longest included time point post-implementation. It was classified using recommended categories from the literature (14, 17, 25, 26), that is, maintenance (6–12 months post-implementation), early sustainment (12–24 months post-implementation), and sustainment (>24 months post-implementation). ‘Post implementation’ was classified in the following ways (19).

  • No ongoing support: The study outlined a timeline with a clear end date and described final data collection or implementation support strategies as ending.

  • Lower than implementation support: The study described an intensive implementation support period as ending but outlined low-intensity support, such as access to a support officer (fewer counts than implementation), return to usual practice, or a periodic newsletter.

There were no restrictions on country of origin, language, or publication date. To assess sustained implementation, the unit of analysis was the EBI. This approach allowed us to (a) incorporate previously reported trial information (e.g., pilot studies) for the use of sustainment frameworks, and so on and (b) allow for substitution: replacing the implementation effect study with a subsequent (eligible) trial that included a follow-up timepoint. As a result, multiple individual studies could be included for a single EBI. Figure 1 displays the unit of analysis and the corresponding studies (implementation effect/follow-up) that meet the inclusion criteria.

Figure 1.

Flowchart depicting the inclusion criteria of implementation effect studies. It includes three main columns: Unit of analysis, Implementation effect study, and Follow-up study. The Unit of analysis involves an EBI program, policy, or practice that is implemented and sustained. The Implementation effect study confirms a positive effect based on statistical significance at the end of implementation. The Follow-up study includes data reported using quantitative methods, with the follow-up timepoint being six months or more after implementation ends. The outcome measured in the follow-up is comparable to the implementation effect study.

Explanation of unit of analysis and the inclusion summary.

Search strategy

Information sources

All citations relating to effective implementation trials (those with a statistically significant effect measured directly at the end of implementation) included in Wolfenden (2022) were used in the forward citation search.

Search process

We developed the following search strategy in consultation with an information specialist. Follow-up studies were sourced by the following:

  1. Conducting a forward citation search in PubMed, Scopus, and Google Scholar for each effective EBI. Each citation associated with the EBI listed in the original review was searched (e.g., effect paper, pilot paper, process evaluation, and so on). Individual citations were searched in each search engine, and following ‘cited by’ or forward citation links, all listed studies were retrieved.

  2. Searching the EBI registration listed on either the International Clinical Trials Registry Platform (ICTRP) or the relevant platform in the EBI country (e.g., Australian EBIs in the Australian New Zealand Clinical Trials Registry) sites

  3. Searching the EBI name in Google for associated EBI websites, if available

  4. Contacting the corresponding author of the original implementation at EBI via email if additional data is required

Screening began manually in EndNote, where each EBI was assigned to a group, and all citations identified during the search were collated and deduplicated. Citations were uploaded to Covidence, and title and abstract screening was conducted independently by two researchers (XL, CG). Two reviewers independently conducted full-text screening and recorded inclusion or exclusion in the online data capture tool REDCap (27) (CG, KF). Findings were cross-checked, and disagreements were resolved by consensus or by assessment by a third member of the research team (AH, AZ, NN). Search results are reported in the PRISMA flow chart (Figure 2).

Figure 2.

PRISMA Flowchart illustrating the process of selecting effective implementation trials: PRISMA flowchart illustrating the process of selecting effective implementation trials. Flowchart illustrating the process of selecting effective implementation trials. It begins with 38 included trials from Wolfenden et al 2022, identifying 23 as effective. A forward citation search retrieves 11,117 papers, which are then de-duplicated to 5,483 records. From these, 5,392 are excluded, and 91 full-text articles are assessed for eligibility, excluding 85 based on criteria like unrelated study, no additional timepoints, and non-comparable design. This results in five eligible follow-up papers, and one additional follow-up from author contact, ultimately identifying six effective evidence-based interventions with eligible follow-up.

PRISMA flow diagram for included studies.

Data collection process

Data extraction was conducted independently by two researchers using a piloted extraction form in REDCap (CG, KF). Any conflicting data was resolved by consensus or by a third member of the research team (AH).

Data extraction

Data on EBI characteristics, implementation strategies, implementation behavioral outcomes, and sustainment characteristics were extracted at the EBI unit of analysis, using data from the Wolfenden review (21) and all eligible papers. The sustainment items selected for extraction are based on those identified by existing reviews and frameworks as pertinent to sustained EBI delivery (14, 15, 17, 28). A full-text version of the extraction template is available in Supplementary file 2.

  • EBI characteristics: Author, year, EBI name (if one), country, setting, health behavior(s), trial design.

  • Implementation strategies: extracted from Wolfenden (using EPOC framework) (29)

  • End of implementation outcomes extracted from the Wolfenden, selected primary outcome reported in the review (21)

  • Sustainment characteristics: Extracted items relating to sustainment were classified ‘yes/no’ and ‘occasionally’ with a section to describe sustainment planned for in design, definition of sustainment used, use of ‘sustainment’ or related term, use of sustainment-specific framework (author, date of framework), sustainment support (level of support and description), and threshold considered sustained (description).

Effect measures

Where possible, the follow-up effect on implementation and behavioral outcomes was calculated using the same method as in the implementation effect studies (21). Specifically, for studies reporting a difference in outcomes between the control and intervention groups at long-term follow-up, we calculated SMDs and 95% confidence intervals to enable comparison across studies to assess the magnitude of the implementation intervention effect. This applied to papers reporting both continuous and dichotomous outcomes, using a standardized formula to convert to an SMD. For follow-up studies that reported implementation or behavioral outcomes for the intervention group only (i.e., no intervention vs. control comparison), we calculated the percentage of the outcome sustained at long-term follow-up. This indicated the percentage of the implementation effect that was sustained after an effective implementation trial ended. This follows methods used in other sustainability reviews that report sustainability as a percentage (14, 30) and in reviews assessing the effect of implementation at scale-up (31).

This was calculated as follows:

Level ofoutcomeatlongterm followupLevel ofoutcome after initial implementation×100

Synthesis methods

For Aim 1, the percentage of effective implementation trials that assessed long-term sustainment was calculated. For Aim 2, sustainment characteristics were reported descriptively using a combination of yes/no indicators and text responses.

For Aim 3, a random-effects meta-analysis was initially planned to assess sustained implementation effects and sustained behavioral outcomes. However, this was not feasible due to the small number of studies reporting a difference in sustained implementation and/or effect on behavioral outcomes between the intervention and control groups at follow-up. Therefore, the median and interquartile range of the percentage of sustained effect across all EBIs were reported separately for implementation and behavioral outcomes to provide a broad indication of the extent to which school-based EBIs are being sustained following a successful implementation trial.

Reporting bias assessment and critical appraisal

Quality assessments were conducted independently by two review authors (CG, ER-G) using Cochrane’s Risk of Bias tool version 1 (consistent with the Wolfenden review), and a third review author (AH) was consulted to resolve discrepancies. Risk of bias was assessed for all follow-up studies. The risk of bias for implementation effect studies was previously assessed in the original review (21). Judgments made in the original review were extracted and included in this review, where follow-up studies used the same design and sample as the implementation trial (domains include selection bias, performance bias, detection bias, recruitment to cluster, and baseline imbalance). Domains relevant to the follow-up timepoint were reassessed (attrition bias, reporting bias, loss of clusters, incorrect analysis, and potential confounding). Where follow-up studies used a different sample than the one used during EBI implementation, all domains were reassessed. “Overall assessment of bias” was reassessed for all EBIs. Risk of bias was reported as “high,” “low,” or “unclear” as per the Cochrane Handbook of Systematic Review of Interventions (32). Where follow-up studies used non-controlled designs, the Newcastle-Ottawa Quality Assessment Scale (NOS) was used, specifically the adapted version for cross-sectional studies, which has been applied in other reviews (33, 34). The NOS rates the quality of studies with a star system, assigning a maximum of 10 stars for the cross-sectional study tool. Three domains were assessed: selection of study groups, comparability of groups, and assessment of outcomes.

Results

Search

Of the 38 included implementation trials in the Wolfenden (2022) review, 23 reported a significant implementation effect and were eligible for this forward citation search (21). After deduplication, a total of 5,483 individual studies were generated through the forward citation search (Figure 2). A total of 91 studies were screened for full text, leading to the identification of six EBIs with associated follow-up papers for inclusion. In most cases, excluded papers (n = 85) did not report data from a follow-up time point (n = 51); instead, they typically reported data at implementation time points (e.g., process evaluations or subgroup analyses). The search identified a follow-up time point for one of the eligible studies, collected during a later cRCT implementation. This updated data replaced the implementation effect trial included in the Wolfenden (2022) review, as the newer trial used comparable implementation strategies and outcomes (8, 35). The included follow-up papers were grouped with their corresponding implementation-effect studies and reported in Table 1.

Table 1.

Characteristics table of included implementation-effect studies and corresponding follow-up articles.

Primary author, year, EBI name (if one), country Setting and targeted health behavior(s) Implementation trial design (sample size) Implementation outcome Implementation strategies (EPOC) (29) Implementation effect
SMD [CI]
Follow-up paper, author (year) Follow-up study design (sample size) Time since implementation ended (months)
EBIs with eligible follow-up timepoint n = 6
Implementation outcomes reported at follow-up
Nathan et al., 2022, Physically Active Children in Education (PACE), Australia (8, 35) Elementary schools (kindergarten – 6th grade), Physical Activity cRCT (61 schools) Teacher implementation of physical activity policy across a school week (minutes) Education materials
Educational outreach visits or academic detailing
Local opinion leaders
Tailored interventions
Other: mandate change
0.65 [95% CI 0.21 to 1.10] Nathan et al. (2022) (8) cRCT (61 schools) 6 months
Wolfenden et al., 2017, Canteens: (no formal name), Australia (53) Elementary schools, nutrition cRCT (70 schools) Healthy items represented >50% of the canteen menu (more is better) Audit and feedback
Continuous quality improvement
External funding
Educational materials
Educational meeting
Educational outreach visits
Local consensus process
Local opinion leader
Tailored intervention
Other: newsletter, acknowledgment of compliant schools, kitchen equipment supplied
1.37 [95% CI 0.68 to 2.07] Wolfenden et al. (2019) (37) cRCT (56 schools) 12 months
Sallis et al. 1997, Sports, Play, and Active Recreation for Kids (SPARK), United States (42) Elementary schools (grades 4–5), PA Non RCT (seven schools) – two intervention arms: teacher trained; two
PE specialist
Amount of physical education per week (minutes) (more is better) All schools (including control):
Other: equipment supplied
All intervention schools:
Educational materials
Length of consultation
Teacher trained arm only:
Educational meetings
Educational outreach visits
1.10 [95% CI 0.55 to 1.64] McKenzie (1997) (39) Experimental, (n = 14 teachers) 18 months
Nathan et al. (2012). Good for Kids. Good for Life, Australia (40) Elementary schools (kindergarten – 6th grade), nutrition Non-RCT (828 schools) Prevalence of vegetable and fruit breaks Local consensus process
Local opinion leaders
Educational meetings
Educational materials
Tailored intervention
Monitoring the performance of the delivery of health care
Other: Incentives
0.59 [95% CI 0.32 to 0.86] The author supplied data Cross-sectional survey (15 schools) 120 months
Behavioral outcomes at follow-up
Mathur et al. (2016), Bihar School Teachers Study, India (41) Government high schools (grades 8–10), Tobacco cessation and prevention cRCT (72 schools) Cessation of tobacco products among teachers and implementation of school tobacco policies. Local opinion leaders
Continuous quality improvement
Educational meetings
Educational materials
Local consensus process
1.11 [95% CI 0.88 to 1.34] Sorensen (2013) (36) cRCT, follow-up paper (72 schools, 677 employees) 9 months
Farmer et al., 2017, PLAY, New Zealand (38) Government elementary schools, Physical Activity cRCT (16 schools) Play space evaluation score Audit and feedback
Local consensus process
Tailored interventions
Other: external funding
1.2 [95% CI 0.13 to 2.26] Farmer et al. (2017) (38) cRCT (16 schools, 669 students) 12 months

EBI, evidence-based intervention; cRCT, cluster randomised controlled trial; EPOC, Cochrane Effective Practice and Organisation of Care (EPOC) taxonomy; SMD, standardized mean difference.

Characteristics of included studies

As shown in Table 1, two follow-up studies collected data 6–12 months after implementation ended (8, 36) (classified as maintenance), and three between 12 and 24 months (37–39) (classified as early sustainment). One was conducted longer than 24 months (personal communication, September 9, 2024) (classified as sustainment). Most included EBIs had been delivered to students and staff in elementary schools (8, 37–40), with one in the high school setting (36). Three interventions aimed to increase student physical activity (8, 38, 39), two targeted dietary behaviors (37, 40), and one aimed to prevent and cease tobacco use (36). While all implementation effect studies included a control comparison, only two follow-up studies reported implementation outcomes with a comparison group (8, 37), and a further two reported behavioral outcomes by comparison group (36, 38), but not implementation outcomes. Two reported follow-up data for intervention groups only (39, 40). The EBIs, including studies and data collection time points, are summarized in Figure 3.

Figure 3.

Summary figure for included Evidence Based Interventions (EBIs). Columns include the unit of analysis, implementation effect study, months post-implementation, follow-up study, and outcome measured. The studies cover topics like physical activity, teacher tobacco quit rates, and canteen menu contents. Specific studies include PACE, Bihar School Teachers Study, PLAY, Canteens, SPARK, and Good for Kids. Each entry details the year of the study, the duration post-implementation, and the specific outcomes related to implementation or behavioral measures.

Summary of included EBIs, including papers used, timepoints, and outcomes measured at follow-up.

Quality assessment

Risk of bias was used to assess the quality of five included follow-up studies shown in Figure 4, and one study was assessed using the NOS shown in Figure 5. Three studies were deemed low risk, two were deemed high risk or low quality, and one risk was unclear. The highest-risk-of-bias domains relate to blinding of participants and personnel during outcome implementation and to blinding of outcome assessment, reflecting the nature of the trials, where blinding is difficult to achieve.

Figure 4.

A table assesses risk of bias for different follow-up studies. It includes five studies: Farmer 2017, Sorensen 2013, Nathan 2021, McKenzie 1997, and Wolfenden 2019. Criteria columns D1-D7 and C1-C5 are marked with symbols: plus (low risk), minus (high risk), and question mark (unclear risk), with colors green, red, and yellow respectively. Results show overall risk as low, unclear, or high for each study. Specific symbols and colors indicate the level of bias for each criterion, with a summary risk rating on the right.

Risk of bias.

Figure 5.

Table evaluating a follow-up study named Nathan F&V. It assesses three domains: Selection (maximum 5 stars), Comparability (maximum 2 stars), and Outcome (maximum 3 stars). In Selection, the study receives one star for "Ascertainment of the exposure (risk factor)." In Outcome, it receives one star for "Assessment of the outcome." No stars are awarded in Comparability. The overall quality score is labeled as "Poor."

Newcastle Ottawa Scale quality assessment.

Aim 1: percentage of EBIs that assess sustainment after the conclusion of an effective implementation trial

Six of the 23 (26%) implementation effect studies collected comparable follow-up data. Sustained implementation was assessed by four (17%), and maintenance of the EBI’s behavioral outcomes was assessed by two (9%) at a long-term follow-up time point.

Aim 2: presence of recommended sustainment practices reported by EBIs

As shown in Table 2, four of the six included EBIs were planned for long-term follow-up, as reported in the implementation effect study (8, 38, 41), or through program monitoring with repeated cross-sectional surveys (40). Two used “sustainment” or related terms in the original EBI papers (8, 41), and one of these defined sustainability as “the extent to which a newly implemented treatment is maintained or institutionalized within a service setting’s ongoing, stable operations” (8). No studies followed a sustainment-focused framework or considered a threshold for sustainment. While no studies reported providing ongoing implementation support, one EBI (8) stated that all participant schools were able to access resources and phone support with a health department officer during the maintenance period; hence, the classification as providing ongoing support at a level ‘lower than implementation support’.

Table 2.

Presence of sustainment practices (definition and so on) in follow-up studies, reported by the sustainment period.

EBI (follow-up paper author, year) Long-term follow-up planned Y/N (description) Sustainment defined Y/N (reference) Sustainment term used in original papers Y/N, term used (paper type, author, year) Sustainment framework or theory used Y/N Level of external support at the sustainment follow-up time point Strategies reported at follow-up
Studies reporting at maintenance timepoint (6–12 months)
PACE (Nathan, 2022) (8) Y (planned 6-month follow-up in this paper) Y (Proctor et al.) (16) Y, sustainability (Protocol, Nathan, 2019) N Lower than implementation Online resources, telephone support
Bihar School
Teachers Study (Sorensen 2013) (36)
Y (planned 6-month follow-up in implementation effect study) N Y, maintenance, (outcomes, Mathur, 2016) (41) N No ongoing support None reported
Studies reporting at early sustainment timepoint (12–24 months)
PLAY (Farmer 2017) (38) Y (planned 12-month follow-up in this paper) N N N No ongoing support None reported
Wolfenden (Wolfenden 2019) (37) N (long-term follow-up not mentioned in papers) N N N No ongoing support None reported
SPARK (McKenzie 1997) (39) N (long-term follow-up not mentioned in papers) N N N No ongoing support None reported
Studies reporting at the sustainment timepoint (24 months or greater)
Nathan V&F (author supplied data 2024) Y, repeated cross-sectional surveys (correspondence) N N N No ongoing support None reported

Aim 3: level of effect sustained in relation to the implementation outcome/s following an effective implementation trial

Of the six studies that conducted a longer-term follow-up, four were RCTs (8, 37, 38, 41) and two were non-RCTs (42). Regarding the sustainability outcomes assessed, four assessed the long-term implementation effect (8, 37–39), and two assessed the long-term behavioral effect only (36, 38). Table 3 reports the sustained effect of the measured outcome.

Table 3.

Percentage of outcome effect sustained at the follow-up time point following the end of an effective implementation intervention.

Timepoint Both End of implementation Follow-up Follow-up End of implementation Follow-up Follow-up
EBI (follow-up author year) (RoB) Outcome measure used Outcome effect (M) SMD [CI] Length of follow-up (months) Follow-up outcome effect (M) Outcome in the intervention group Outcome in the intervention group % outcome sustained
Implementation outcome reported
PACE (Nathan, 2021)
Green oval with the word "low" in black text centered inside. The image is bordered by two horizontal black lines.
Weekly minutes of classroom physical activity 0.76 (0.56, 0.96) 6 0.46 (0.26, 0.66) M = 148.5 (SD = 44.1) M = 146.7 (SD = 40.3) 61%b
Wolfenden (37)
Green oval with the word "low" in black text centered inside. The image is bordered by two horizontal black lines.
Healthy items represented >50% of the canteen menu (more is better) 1.38 [0.67, 2.09] 12 0.71
(−0.07, 1.49)
N = 22 (81.5%) N = 24 (89%) 51%b
SPARK (McKenzie 1997) (39)
The word "high" is displayed in black lowercase letters inside a red oval, with a horizontal black line above it.
Amount of PE per week (minutes) (more is better) 1.10 [0.55, 1.64] 18 NR (no baseline data)a M = 64.6
(CI 59, 70.2)
59.2 (SD = 27.5) 92%a
Nathan 2012 (40) (author supplied data)
**/10
Prevalence of vegetable and fruit breaks at schools 0.59 [0.32, 0.86] 120 NRa N = 318 (82%) 15 (100%) 122%a
Behavioral outcome reported
Bihar Teacher Study (Sorensen 2013) (36)
The word "unclear" in bold black font inside a yellow oval background.
Tobacco quit rate −0.25 [−47, −0.03] 9 −0.20 (−0.42, 0.0.02) 49.5% (with outcome) 17.4% (with outcome) 80%b
PLAY (Farmer, 2017) (38)
Green oval with the word "low" in black text centered inside. The image is bordered by two horizontal black lines.
Student activity for the whole day counts per minute −0.03 (−0.19, 0.13) 12 −0.12 (−0.28, 0.037) M = 634 (SD = 164) M = 566 (SD = 161) −400%b

EBI, evidence-based intervention; RoB, overall risk of bias (reported in full detail in Figures 4, 5); PE, physical education; NR, not reported; M, mean; SMD, standard mean difference; CI, confidence interval.

a

No reported difference between intervention and control at long-term follow-up; the effect at follow-up could not be calculated. Instead, the percentage sustained was calculated as the level of outcome achieved in the intervention group at long-term follow-up divided by the level of outcome achieved in the intervention group following initial implementation.

b

The reported difference between intervention and control at both timepoints was calculated as the percentage of effect, defined as the SMD at long-term follow-up divided by the SMD following initial implementation.

Implementation outcome

Two of the follow-up studies reporting the long-term implementation effect employed an RCT design (8, 37) and compared the implementation effect between the control and intervention groups at follow-up. The SMDs for these two EBIs were calculated at the follow-up time point and are reported in Table 3. Both EBIs showed a reduction in implementation effect at long-term follow-up, as indicated by the between-group comparison. Across the four studies reporting sustained implementation effects, the percentage of the sustained effect ranged from 51 to 122%, with a median of 76% (IQ1 = 56, IQ3 = 107%).

Behavioral outcome

Two included EBIs reported behavioral outcomes at a follow-up time point after implementation ended, as shown in Table 3. The first (36) achieved 80% of the behavioral outcome effect 9 months after the implementation support ended. The other (38) reported a negative effect on the behavioral outcome at the end of implementation and found a further reduction in effect, four times the original effect, at the 12-month follow-up timepoint.

Discussion

This review explored whether implementation trials that successfully improved the delivery of school-based health promotion EBIs also assessed longer-term sustainment and, if so, what level of sustainment was achieved. We leveraged existing data to explore the sustainment of school-based health promotion EBIs, an approach suggested by leaders in public health to maximize the value of prior investments and overcome challenges in sustainability research, such as resource constraints and the long timeframes typically required for follow-up studies (16). Of 23 effective implementation studies, we found only a small proportion (26%) conducted a comparable follow-up study, and just 4% assessed the delivery of the EBI beyond 24 months (16). Furthermore, data quality and comparability across follow-up studies were low, limiting the ability to draw firm conclusions on EBI sustainment. For example, several studies were excluded because they differed from the original implementation trial and the follow-up study, even though they appeared to assess sustainment (43, 44). Reasons for exclusion included different data types (e.g., quantitative vs. qualitative follow-up analyses) and the assessment of non-comparable outcomes or implementation strategies to those used in the original implementation trial. The lack of follow-up assessments, combined with limited data comparability, represents a missed opportunity to understand the full, long-term impact of implementation investments (17), an issue recognized in previous reviews on sustainment (24). To assess the impact of effective implementation on ongoing EBI delivery, the field needs to prioritize collecting and reporting comparable data (i.e., reassessing outcome measures for implementation and behavioral outcomes), ideally using controlled designs and at multiple time points, including at least 12 months post an effectively implemented EBI (16, 17, 24, 43–45).

We also examined whether the included EBIs applied practices that researchers have recommended for sustained EBI delivery. Consistent with previous reviews in public health settings, few studies included sustainment practices (14, 46) such as planning (n = 4), definitions (n = 1), or sustainment-specific frameworks (n = 0). This may be because sustaining EBI delivery was not the focus of these studies. These gaps may reflect when the studies were conducted, as the field has advanced in recent years. To improve future evidence, quality sustainment research should plan long-term assessments from the outset, draw on established sustainment definitions (18) and frameworks (13, 17), and use consistent outcome measurement to capture sustained implementation and behavioral outcomes (47). It is recommended that future research be undertaken to develop reporting guidelines for sustainment research that standardize the inclusion of these practices. This should help to improve the quality of future long-term follow-up studies and, in time, the ability to undertake rigorous evidence syntheses that improve our understanding of how to sustain effectively implemented programs (5, 17).

Positive health impact requires both sustained delivery of an EBI as well as maintenance of the positive impact the EBI has on behavioral outcomes (48). None of our included studies assessed both of these outcomes, with most (four out of six) focusing on implementation outcomes. Of the four that reported sustained implementation effects, two continued to show 90% or more implementation effects. In contrast, the other two EBIs reported results comparing intervention to control, with sustained effects of 51 and 61%. This could indicate program diffusion across the entire sample or other confounding factors, such as time or context. Across the four studies reporting implementation outcomes, 50% sustained effects at or above 80%, a rate more than twice that reported in previous public health sustainment reviews (14, 30), including in the school setting (20). Our observations are exploratory and limited by the small sample size and heterogeneity in outcomes; however, they suggest the potential for sustainment of effectively implemented EBIs (14, 20, 30). Theoretically, this makes sense, as EBI delivery cannot be sustained unless it is first successfully implemented, a hypothesis that aligns with implementation frameworks (48).

While continued delivery of EBIs is important for achieving population benefit, it is not sufficient on its own. The EBI itself must continue to exert a positive behavioral effect. In this review, we identified one study that found worsening behavioral outcomes at follow-up, with a follow-up effect four times lower than observed at the end of implementation, compared to control (38), despite a significant impact at the end of implementation. This emphasizes the importance of monitoring both continued program delivery and behavioral outcomes to assess ongoing impact at sustainment timepoints. This aligns with Moore et al.’s sustainability definition that includes constructs for continued program delivery, behavior change maintenance, and continued behavioral benefits (2017) (18). When program monitoring indicates that benefits are not continuing or have become harmful, guidance may be needed on how to improve the EBI or de-implement it (49). Importantly, to conduct reliable long-term EBI monitoring, we need increased research into developing and validating rigorous measurement tools that can address complexity and capture causal mechanisms in real-world program sustainment (47, 50).

Strengths and limitations

We employed an efficient search strategy using a robust Cochrane systematic review as the sampling base. This approach leveraged EBIs already classified as implementation trials, enabling extensive, systematic searches of citations to identify related long-term follow-up papers. However, some eligible EBIs not included in the original review may have been overlooked. The majority of the included studies were from English-speaking, high-income countries and, as such, may have limited generalizability outside these contexts. The small sample is heterogeneous across target behaviors, design, outcome measures, and contexts, making interpretation challenging. However, it reflects the realities and challenges commonly faced in undertaking reviews in the area of implementation (14, 21). Furthermore, focusing on schools as the implementation setting provides consistency for this exploratory study, an approach used by other implementation-focused reviews (10, 21, 51).

Ideally, we would have calculated the sustained effect for all studies using data comparing control and intervention groups at long-term follow-up, but few reported adequate data. Instead, we relied on descriptive data from intervention groups in two studies and reported results for both behavioral and implementation outcomes. As a result, there is limited capacity to show whether between-group differences persist during the sustainment period. Despite this, we provide insights into sustained effects over time and highlight specific directions for improving reporting in the field.

An important strength of this study is its reporting of sustained EBIs following receipt of effective implementation interventions alone; however, this introduces potential selection bias, limiting the capacity to compare sustained effects between effective and ineffective trials. Similarly, publication bias may have affected follow-up data, with unfavorable results more likely to remain unpublished. To best understand sustained effects over time, it is important to publish long-term follow-up data regardless of outcomes.

Conclusion

This review advances a priority in implementation science – understanding the long-term delivery of EBIs (52) – by examining whether effectively implemented health promotion EBIs in schools were sustained over time. Achieving EBI sustainment can yield sustained benefits for participants, enhance return on investment, and contribute to population-level health gains. However, if EBIs or their implementation are not effective, efforts to sustain them may be premature and potentially wasted. In such cases, resources should first focus on improving implementation fidelity and impact. Generating meaningful sustainment evidence requires long-term reporting of implementation outcomes from trials that demonstrate initial success. To achieve this, the field must prioritize appropriate resourcing and adequate timelines that support sustained follow-up (7, 16, 17). Failure to turn successful EBI implementation into sustained delivery represents a missed opportunity to yield maximum returns on implementation investments and will continue to fall short of population health impact.

Acknowledgments

We would like to acknowledge the guidance of the librarians at the University of Newcastle in the development of the search strategy. Additionally, we thank Luke Wolfenden for advice on the conceptual direction of this review.

Funding Statement

The author(s) declared that financial support was received for this work and/or its publication. The research has been funded as part of a Medical Research Future Fund (MRFF) Investigator Grant (MRF1194785) held by NN and the MRFF Early to Mid-Career Researchers scheme (MRF2030911). CG is funded by a PhD scholarship through the University of Newcastle.

Edited by: Mohan Bairwa, All India Institute of Medical Sciences, India

Reviewed by: Sachin Shinde, Harvard University, United States

Lawrence StLeger, Deakin University, Australia

Abbreviations: EBI, Evidence-based intervention; RCT, Randomized controlled trial; SMD, Standardized mean difference; cRCT, Cluster randomized controlled trial; CBAs, Controlled before and after studies; cCBAs, Cluster controlled before and after studies; PA, Physical activity; EPOC, Cochrane Effective Practice and Organisation of Care (EPOC) taxonomy.

Data availability statement

The data analyzed in this study are subject to the following licenses/restrictions: Data were extracted from published studies and summarized in this review; the extraction file is available from the authors upon reasonable request. Requests to access these datasets should be directed to carly.gardner@health.nsw.gov.au.

Author contributions

CG: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing. AH: Conceptualization, Formal analysis, Methodology, Resources, Supervision, Writing – original draft, Writing – review & editing. CL: Conceptualization, Methodology, Supervision, Writing – review & editing. AZ: Data curation, Investigation, Writing – review & editing. SM: Data curation, Formal analysis, Writing – review & editing. ER-G: Data curation, Investigation, Writing – review & editing. XL: Data curation, Investigation, Writing – review & editing. KF: Data curation, Investigation, Writing – review & editing. RS: Supervision, Writing – review & editing. NN: Conceptualization, Funding acquisition, Methodology, Resources, Supervision, Writing – review & editing.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpubh.2026.1725392/full#supplementary-material

Table_1.docx (41.2KB, docx)

References

  • 1.World Health Organization. Noncommunicable diseases. Available online at: https://www.who.int/news-room/fact-sheets/detail/noncommunicable-diseases#:~:text=Noncommunicable%20diseases%20%28NCDs%29%20kill%2041%20million%20people%20each,%22premature%22%20deaths%20occur%20in%20low-%20and%20middle-income%20countries
  • 2.Akseer N, Mehta S, Wigle J, Chera R, Brickman ZJ, Al-Gashm S, et al. Non-communicable diseases among adolescents: current status, determinants, interventions, and policies. BMC Public Health. (2020) 20:1908. doi: 10.1186/s12889-020-09988-5, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Nathan N, Elton B, Babic M, McCarthy N, Sutherland R, Presseau J, et al. Barriers and facilitators to the implementation of physical activity policies in schools: a systematic review. Prev Med. (2018) 107:45–53. doi: 10.1016/j.ypmed.2017.11.012, [DOI] [PubMed] [Google Scholar]
  • 4.Wendt JSD, Fletchtner-Mors M, Meshkovska B, Luszczynska A, Lien N, Forberger S, et al. Barriers and facilitators to the adoption of physical activity policies in elementary schools from the perspective of principals: an application of the consolidated framework for implementation research–a cross-sectional study. Front Public Health. (2023) 11:935292. doi: 10.3389/fpubh.2023.935292 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Shoesmith A, Hall A, Wolfenden L, Shelton RC, Powell BJ, Brown H, et al. Barriers and facilitators influencing the sustainment of health behavior interventions in schools and childcare services: a systematic review. Implement Sci. (2021) 16:62. doi: 10.1186/s13012-021-01134-y, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Ashcraft LE, Goodrich DE, Hero J, Phares A, Bachrach RL, Quinn DA, et al. A systematic review of experimentally tested implementation strategies across health and human service settings: evidence from 2010-2022. Implement Sci. (2024) 19:43. doi: 10.1186/s13012-024-01369-5, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Damschroder LJ, Reardon CM, Widerquist MAO, Lowery J. The updated consolidated framework for implementation research based on user feedback. Implement Sci. (2022) 17:75. doi: 10.1186/s13012-022-01245-0, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Nathan N, Hall A, McCarthy N, Sutherland R, Wiggers J, Bauman AE, et al. Multi-strategy intervention increases school implementation and maintenance of a mandatory physical activity policy: outcomes of a cluster randomised controlled trial. Br J Sports Med. (2022) 56:385–93. doi: 10.1136/bjsports-2020-103764 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Lane C, Naylor PJ, Shoesmith A, Wolfenden L, Hall A, Sutherland R, et al. Identifying essential implementation strategies: a mixed methods process evaluation of a multi-strategy policy implementation intervention for schools. Int J Behav Nutr Phys Act. (2022) 19:44. doi: 10.1186/s12966-022-01281-5, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Barnes C, McCrabb S, Stacey F, Nathan N, Yoong SL, Grady A, et al. Improving implementation of school-based healthy eating and physical activity policies, practices, and programs: a systematic review. Transl Behav Med. (2021) 11:1365–410. doi: 10.1093/tbm/ibab037, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Barnes C, Sutherland R, Janssen L, Jones J, Robertson K, Gowland-Ella J, et al. Improving the adoption of a school-based nutrition program: findings from a collaborative network of randomised trials. Implement Sci. (2025) 20:5. doi: 10.1186/s13012-025-01417-8, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Sutherland R, Campbell E, McLaughlin M, Nathan N, Wolfenden L, Lubans DR, et al. Scale-up of the physical activity 4 everyone (PA4E1) intervention in secondary schools: 24-month implementation and cost outcomes from a cluster randomised controlled trial. Int J Behav Nutr Phys Act. (2021) 18:137. doi: 10.1186/s12966-021-01206-8, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implmentation. Science. (2013) 8. doi: 10.1186/1748-5908-8-117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Wiltsey Stirman S, Kimberly J, Cook N, Calloway A, Castro F, Charns M. The sustainability of new programs and innovations: a review of the empirical literature and recommendations for future research. Implement Sci. (2012) 7:17. doi: 10.1186/1748-5908-7-17 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Scheirer MA, Dearing JW. An agenda for research on the sustainability of public health programs. Am J Public Health. (2011) 101:2059–67. doi: 10.2105/AJPH.2011.300193, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Proctor E, Luke D, Calhoun A, McMillen C, Brownson R, McCrary S, et al. Sustainability of evidence-based healthcare: research agenda, methodological advances, and infrastructure support. Implement Sci. (2015) 10:88. doi: 10.1186/s13012-015-0274-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Shelton RC, Cooper BR, Stirman SW. The sustainability of evidence-based interventions and practices in public health and health care. Annu Rev Public Health. (2018) 39:55–76. doi: 10.1146/annurev-publhealth-040617-014731 [DOI] [PubMed] [Google Scholar]
  • 18.Moore JE, Mascarenhas A, Bain J, Straus SE. Developing a comprehensive definition of sustainability. Implement Sci. (2017) 12:110. doi: 10.1186/s13012-017-0637-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Wolfenden L, Shoesmith A, Hall A, Bauman A, Nathan N. An initial typology of approaches used by policy and practice agencies to achieve sustained implementation of interventions to improve health. Implement Sci Commun. (2024) 5:21. doi: 10.1186/s43058-024-00555-2, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Herlitz L, MacIntyre H, Osborn T, Bonell C. The sustainability of public health interventions in schools: a systematic review. Implement Sci. (2020) 15:4. doi: 10.1186/s13012-019-0961-8, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Wolfenden L, McCrabb S, Barnes C, O'Brien KM, Ng KW, Nathan NK, et al. Strategies for enhancing the implementation of school-based policies or practices targeting diet, physical activity, obesity, tobacco or alcohol use. Cochrane Database Syst Rev. (2022) 8:CD011677. doi: 10.1002/14651858.CD011677.pub3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Higgins JP, Altman DG, Gotzsche PC, Juni P, Moher D, Oxman AD, et al. The Cochrane collaboration's tool for assessing risk of bias in randomised trials. BMJ. (2011) 343:d5928. doi: 10.1136/bmj.d5928, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Lee DC, O'Brien KM, McCrabb S, Wolfenden L, Tzelepis F, Barnes C, et al. Strategies for enhancing the implementation of school-based policies or practices targeting diet, physical activity, obesity, tobacco or alcohol use. Cochrane Database Syst Rev. (2024) 12:CD011677. doi: 10.1002/14651858.CD011677.pub4, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Riley-Gibson E, Hall A, Shoesmith A, Wolfenden L, Shelton RC, Pascoe W, et al. A systematic review to determine the effect of strategies to sustain chronic disease prevention interventions in clinical and community settings. Transl Behav Med. (2025) 15:ibae070. doi: 10.1093/tbm/ibae070 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Ament SM, de Groot JJ, Maessen JM, Dirksen CD, van der Weijden T, Kleijnen J. Sustainability of professionals' adherence to clinical practice guidelines in medical care: a systematic review. BMJ Open. (2015) 5:e008073. doi: 10.1136/bmjopen-2015-008073, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Glasgow RE, Harden SM, Gaglio B, Rabin B, Smith ML, Porter GC, et al. RE-AIM planning and evaluation framework: adapting to new science and practice with a 20-year review. Front Public Health. (2019) 7:64. doi: 10.3389/fpubh.2019.00064, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. (2009) 42:377–81. doi: 10.1016/j.jbi.2008.08.010, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Moullin JC, Sklar M, Green A, Dickson KS, Stadnick NA, Reeder K, et al. Advancing the pragmatic measurement of sustainment: a narrative review of measures. Implement Sci Commun. (2020) 1. doi: 10.1186/s43058-020-00068-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Cochrane Effective Practice and Organisation of Care (EPOC) (2015). EPOC Taxonomy. Available online at: www.epoc.cochrane.org/epoc-taxonomy
  • 30.Scheirer MA. Is sustainability possible? A review and commentary on empirical studies of program sustainability. Am J Eval. (2005) 26:320–47. doi: 10.1177/1098214005278752 [DOI] [Google Scholar]
  • 31.Lane C, McCrabb S, Nathan N, Naylor PJ, Bauman A, Milat A, et al. How effective are physical activity interventions when they are scaled-up: a systematic review. Int J Behav Nutr Phys Act. (2021) 18:16. doi: 10.1186/s12966-021-01080-4, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Higgins JP, Savović J, Page MJ, Elbers RG, Sterne JAC. Chapter 8: Assessing risk of bias in a randomized trial [last updated October 2019]. In: eds. Higgins JP, Thomas J, Chandler J, Cumpston M, Li T, Page MJ, et al., Cochrane Handbook for Systematic Reviews of Interventions version 6.5. Cochrane, (2024) (Accessed June 2023). [Google Scholar]
  • 33.Modesti PA, Reboldi G, Cappuccio FP, Agyemang C, Remuzzi G, Rapi S, et al. Panethnic differences in blood pressure in Europe: a systematic review and Meta-analysis. PLoS One. (2016) 11:e0147601. doi: 10.1371/journal.pone.0147601, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.McLaughlin M, Delaney T, Hall A, Byaruhanga J, Mackie P, Grady A, et al. Associations between digital health intervention engagement, physical activity, and sedentary behavior: systematic review and meta-analysis. J Med Internet Res. (2021) 23:e23180. doi: 10.2196/23180, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Nathan NK, Sutherland RL, Hope K, McCarthy NJ, Pettett M, Elton B, et al. Implementation of a school physical activity policy improves student physical activity levels: outcomes of a cluster-randomized controlled trial. J Phys Act Health. (2020) 17:1009–18. doi: 10.1123/jpah.2019-0595, [DOI] [PubMed] [Google Scholar]
  • 36.Sorensen G, Pednekar MS, Sinha DN, Stoddard AM, Nagler E, Aghi MB, et al. Effects of a tobacco control intervention for teachers in India: results of the Bihar school teachers study. Am J Public Health. (2013) 103:2035–40. doi: 10.2105/ajph.2013.301303, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Wolfenden L, Nathan N, Reilly K, Delaney T, Janssen LM, Reynolds R, et al. Two-year follow-up of a randomised controlled trial to assess the sustainability of a school intervention to improve the implementation of a school-based nutrition policy. Health Promot J Austr. (2019) 30:26–33. doi: 10.1002/hpja.238 [DOI] [PubMed] [Google Scholar]
  • 38.Farmer VL, Williams SM, Mann JI, Schofield G, McPhee JC, Taylor RW. The effect of increasing risk and challenge in the school playground on physical activity and weight in children: a cluster randomised controlled trial (PLAY). Int J Obes. (2017) 41:793–800. doi: 10.1038/ijo.2017.41, [DOI] [PubMed] [Google Scholar]
  • 39.McKenzie, Sallis JF, Kolody SJ, Faucette FN. Long-term effects of a physical education curriculum and staff development program: SPARK. Res Q Exerc Sport. (1997) 68:280–91. doi: 10.1080/02701367.1997.10608009 [DOI] [PubMed] [Google Scholar]
  • 40.Nathan N, Wolfenden L, Bell AC, Wyse R, Morgan PJ, Butler M, et al. Effectiveness of a multi-strategy intervention in increasing the implementation of vegetable and fruit breaks by Australian primary schools: a non-randomized controlled trial. BMC Public Health. (2012) 12:651. doi: 10.1186/1471-2458-12-651, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Mathur N, Pednekar M, Sorensen GS, Nagler EM, Stoddard AM, Lando HA, et al. Adoption and implementation of tobacco control policies in schools in India: results of the Bihar school teachers study. Asian Pac J Cancer Prev. (2016) 17:2821–6. [PubMed] [Google Scholar]
  • 42.Sallis JF, McKenzie TL, Alcaraz JF, Kolody B, Faucette N, Hovell MF. The effects of a 2-year physical education program (SPARK) on physical activity and fitness in elementary school students. Am J Public Health. (1997) 87:1328–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Masse LC, McKay H, Valente M, Brant R, Naylor PJ. Physical activity implementation in schools: a 4-year follow-up. Am J Prev Med. (2012) 43:369–77. [DOI] [PubMed] [Google Scholar]
  • 44.Pate RR, Saunders R, Dishman RK, Addy C, Dowda M, Ward DS. Long-term effects of a physical activity intervention in high school girls. Am J Prev Med. (2007) 33:276–80. doi: 10.1016/j.amepre.2007.06.005, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Naylor PJ, Macdonald HM, Zebedee JA, Reed KE, McKay HA. Lessons learned from action schools! BC--an 'active school' model to promote physical activity in elementary schools. J Sci Med Sport. (2006) 9:413–23. doi: 10.1016/j.jsams.2006.06.013, [DOI] [PubMed] [Google Scholar]
  • 46.Hailemariam M, Bustos T, Montgomery B, Barajas R, Evans LB, Drahota A. Evidence-based intervention sustainability strategies: a systematic review. Implement Sci. (2019) 14:57. doi: 10.1186/s13012-019-0910-6, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Hall A, Shoesmith A, Doherty E, McEvoy B, Mettert K, Lewis CC, et al. Evaluation of measures of sustainability and sustainability determinants for use in community, public health, and clinical settings: a systematic review. Implement Sci. (2022) 17:81. doi: 10.1186/s13012-022-01252-1, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Moullin JC, Dickson KS, Stadnick NA, Rabin B, Aarons GA. Systematic review of the exploration, preparation, implementation, sustainment (EPIS) framework. Implement Sci. (2019) 14:1. doi: 10.1186/s13012-018-0842-6, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Shelton RC, Chambers DA, Glasgow RE. An extension of RE-AIM to enhance sustainability: addressing dynamic context and promoting health equity over time. Front Public Health. (2020) 8:134. doi: 10.3389/fpubh.2020.00134 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Lewis CC, Frank HE, Cruden G, Kim B, Stahmer AC, Lyon AR, et al. A research agenda to advance the study of implementation mechanisms. Implement Sci Commun. (2024) 5:98. doi: 10.1186/s43058-024-00633-5, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Langford R, Bonell C, Jones H, Campbell R. Obesity prevention and the health promoting schools framework: essential components and barriers to success. Int J Behav Nutr Phys Act. (2015) 12. doi: 10.1186/s12966-015-0167-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Admin Pol Ment Health. (2011) 38:4–23. doi: 10.1007/s10488-010-0327-7, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Wolfenden L, Nathan N, Janssen LM, Wiggers J, Reilly K, Delaney T, et al. Multi-strategic intervention to enhance implementation of healthy canteen policy: a randomised controlled trial. Implement Sci. (2017) 12:6. doi: 10.1186/s13012-016-0537-9, [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Table_1.docx (41.2KB, docx)

Data Availability Statement

The data analyzed in this study are subject to the following licenses/restrictions: Data were extracted from published studies and summarized in this review; the extraction file is available from the authors upon reasonable request. Requests to access these datasets should be directed to carly.gardner@health.nsw.gov.au.


Articles from Frontiers in Public Health are provided here courtesy of Frontiers Media SA

RESOURCES