Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Mar 27.
Published in final edited form as: J Appl Sch Psychol. 2019 Mar 27;35(3):257–289. doi: 10.1080/15377903.2019.1577780

Languishing Students: Linking Complete Mental Health Screening in Schools to Tier II Intervention

Stephanie A Moore 1, Ashley M Mayworm 2, Rachel Stein 3, Jill D Sharkey 4, Erin Dowdy 5
PMCID: PMC6588197  NIHMSID: NIHMS1028039  PMID: 31231173

Abstract

Despite innovations in the screening and early identification of students who may benefit from school mental health services, many schools struggle to link screening to intervention decisions, particularly at the Tier II level. Universal complete mental health screening, which measures strengths along with risk factors, is a strength-based approach that enables identification of students who do not report active mental health risk yet have limited psychosocial strengths. These languishing students are ideal candidates for Tier II interventions. Using a case study to link screening to intervention, this paper describes a contemporary approach to complete mental health screening, identify candidates for Tier II intervention, select appropriate interventions, and monitor student outcomes. Implications and challenges for school psychologists are discussed.

Keywords: complete mental health, screening, multi-tiered systems of support, Tier II intervention

Introduction

Over the past decade, the field of school psychology has made great strides in understanding and improving methods for the screening and early identification of students who may benefit from school-based mental health services (Kamphaus, Reynolds, & Dever, 2014). However, there is still a need for additional clarification regarding early identification of student needs (Albers, Glover, & Kratochwill, 2007), including refining our understanding of critical constructs to screen for (Dowdy, Furlong, Eklund, Saeki, & Ritchey, 2010) and how to best link screening data to intervention programming (Volpe, Briesch, & Chafouleas, 2010). With the goal of incrementally understanding how to link screening to intervention, this paper will describe a contemporary approach to mental health screening, with a specific focus on identifying candidates for Tier II intervention. This paper aims to highlight the benefit of linking screening data to Tier II interventions for students who are at-risk for poor physical health, mental health, and social outcomes (e.g. Westerhof & Keyes, 2010), but who often go unidentified when traditional screening approaches are used (Antaramian, Huebner, Hills, & Valois, 2010; Moore et al., 2015; Suldo & Shaffer, 2008). These students, described as languishing, do not report experiencing significant mental health challenges but concurrently report limited psychosocial strengths and resources to support adaptive coping (e.g. Greenspoon & Saklofske, 2001; Keyes, 2002).

Screening for Complete Mental Health

Historically, mental health screening has focused on assessing for problem behaviors and emotions – that is, symptoms of mental distress, psychopathology, or the risk of developing mental health disorders. This deficit-based approach, however, fails to attend to the positive contributions to mental health including subjective well-being and social-emotional strengths (Moore et al., 2015). A more contemporary, strengths-based approach to early identification includes a balance of both distress and strength indicators and is termed complete mental health screening (Furlong, Dowdy, Carnazzo, Bovery, & Kim, 2014a). This approach is aligned with “dual-factor” (Suldo & Shaffer, 2008) and “two-continua” (Keyes, 2005) models of mental health that conceptualize mental health as encompassing both symptoms of psychological distress and social-emotional strengths, and recognize that the absence of illness does not necessarily indicate wellness. Empirical support for the inclusion of strengths when considering mental health is found in dual-factor studies that identify students with higher levels of strengths to have a variety of more positive life outcomes including higher academic achievement, physical health indicators, student engagement, and social functioning (Antaramian et al., 2010; Suldo & Shaffer, 2008).

Although classification approaches vary widely within the complete mental health screening literature, students are often classified into groups using predetermined values (e.g., raw scores, T scores, standard deviations) on both strengths and distress indicators (Kim, Dowdy, Furlong, & You, 2017). For example, to create a dual-factor classification system, students are placed into a group according to the severity of symptoms of distress, and then also grouped based on their reported levels of strengths. Many have elected to form two distress groups (i.e., high or low) and two strengths groups (i.e., high or low; Antaramian et al., 2010; Greenspoon & Saklofske, 2001; Suldo & Shaffer, 2008). Complete mental health groups are then formed by crossing distress and strengths group membership. This method provides for four unique groups of students including: (a) a group high in strengths and low in distress (i.e., complete mental health; thriving); (b) a group low in strengths and high in distress (i.e., highest risk; troubled); (c) a group high in strengths and high in distress (i.e., counter-intuitive, inconsistent group; symptomatic but content); and (d) a group low in strengths and low in distress (i.e., languishing).

In practice, however, schools’ decisions about the desired number of groups are informed by their interest in gathering nuanced information about the mental health functioning of their students and the ability to provide appropriate follow-up intervention. Decisions about the number of groups should be informed by consultation of the measurement literature to inform selection of cut point(s) used to form groups (e.g., to maximize sensitivity and specificity). However, research supporting the predictive utility of many screening tools (particularly those that measure strengths) is not available, thus, cut points are often determined logically, yet arbitrarily, and are set via sample-specific norms (i.e., standardized scores), population-based standardized scores (e.g., T scores), or raw score criterion for symptom frequency/severity (Kim et al., 2017). The often arbitrary nature by which cut points are selected enables school stakeholders to strategize with their school-based teams to determine their capacity for follow-up, and to subsequently adjust cut scores and the number of groups formed (Moore et al., 2015). As an illustration, Moore et al. (2015) conducted complete mental health screening and formed nine groups using multiple cut points for their distress and strength measures.

A common concern about universal screening is that assessment will identify a larger number of youth in need of intervention than can feasibly be served by a school’s infrastructure (e.g., Dever, Raines, Barclay, 2012). Although more students may be identified through universal screening than are currently being served, when screening is implemented in conjunction with multi-tiered supports, the number of students requiring intervention is theoretically expected to decrease over time (Dever et al., 2012; Moore et al., 2015). Vannest (2012) further reassures that identifying risk via universal screening is not the same as identifying a mental health disorder or disability, and that screening results should be verified for their validity through additional data sources (e.g., teacher report, school records, targeted follow-up assessments). Available research suggests that fewer than 5% of students will be identified as having extremely elevated risk for psychological distress (some of whom already receive services), whereas approximately 20% of students would have elevated or extremely elevated levels of risk (Vannest, 2012). The addition of strength-based measures in complete mental health screening also identifies students who are not experiencing symptoms but who require intervention to support their well-being (Moore et al., 2015).

Although concerns about the number of students falling into the highest risk groups can be addressed, in part, by making informed decisions about the ways in which mental health groups are formed (i.e., adjusting cut points to be congruent with service-delivery capacity), best practice recommendations are for school stakeholders to carefully examine their existing intervention resources and capacity prior to conducting universal screening. Screening should only be conducted when the capacity to intervene is adequate. Smaller-scale screening efforts (e.g., of classrooms or grade levels) are recommended prior to schoolwide screening to ensure staff are equipped to coordinate student follow-up (Moore et al., 2015). When universal screening is conducted with a focus on strengths in addition to distress, schools have the opportunity to not only provide services to students identified as at risk of mental health problems, but also to attend to students who may report low levels of assets despite not currently experiencing symptoms of distress.

Identifying languishing youth.

Universal complete mental health screening also allows for the identification of a unique group of students who report low levels of strengths in addition to low levels of distress. This group of students, often termed languishing, or vulnerable, has been consistently identified in dual-factor studies, and are a critical group for targeted prevention and intervention efforts due to their poorer life outcomes when compared to students with complete mental health (Antaramian et al., 2010; Moffa, Dowdy, & Furlong, 2016; Suldo & Shaffer, 2008). For example, when compared to youth with complete mental health indicators (i.e., low distress, high strengths), languishing youth report lower levels of engagement (Antaramian et al., 2010), reduced academic self-concept, fewer beliefs about school’s importance for reaching long term goals, lower levels of general physical health (Suldo & Shaffer, 2008), and lower levels of school belonging (Moffa et al., 2016).

Individuals with languishing mental health can face similar outcomes to youth with high risk indicators (i.e., “troubled” group; high distress, low strengths). For example, previous research found that languishing and troubled groups did not significantly differ in their self-reported levels of engagement, environmental educational supports (e.g., family support for learning, teacher-student relationships, peer support for learning; Antaramian et al., 2010) or school belonging (Moffa et al., 2016). Other research has further supported that a languishing mental health status is associated with poor emotional health and “with substantial psychosocial impairment at levels comparable to an episode of pure depression” (Keyes, 2002, p. 217). Further, because this group of students can be difficult to identify without systematic screening for complete mental health, they are less likely to receive appropriate intervention support (Antaramian et al., 2010). Incorporating measures of well-being into standard screening practices, that is, screening for complete mental health, promotes identification of languishing youth as well as efforts to incorporate wellness-focused interventions into a school’s service delivery framework.

Screening to Inform Tier II Intervention

Screening for complete mental health is aligned with current school-based service delivery models that emphasize universal services, prevention, early intervention, and data-based decision making. Specifically, multi-tiered systems of support (MTSS) have been recommended as a school-based approach to help all students achieve both academically and behaviorally (Nantais, Martin, & Barnes, 2014). Within MTSS frameworks, universal Tier I supports are provided to all students, and approximately 80% of the school population is expected to benefit from these universal supports alone. However, approximately 20% of the student population is expected to require, and receive, additional targeted (Tier II, 15%) and intensive (Tier III; 5%) interventions (Fletcher & Vaughn, 2009). Effective MTSS frameworks use universal screening data to make decisions regarding what interventions to provide at the universal level, as well as determine which students may benefit from additional Tier II or Tier III services (von der Embse, Iaccarino, Mankin, Kilgus, & Magen, 2016).

Within a complete mental health screening framework, students who are found to have complete mental health (i.e., thriving) are likely to only need universal Tier I support. Whereas students who are identified as having low strengths and high distress (highest risk group), along with students with low strengths and low distress (languishing group) and high strengths and high distress (inconsistent group), may benefit from or require additional services. Conceptually, students in the highest risk group likely need individualized Tier III services, whereas students who are languishing or inconsistent may benefit from Tier II services, due to having a lower level of need. Screening data have found between 5 and 13% of students report both low strengths and low distress (Greenspoon & Saklofske, 2001; Moffa et al., 2016; Venning, Wilson, Kettler, & Eliott, 2013), making this group of languishing students an ideal target for Tier II services (Suldo & Shaffer, 2008). Research suggests that more information is available regarding the process of identifying students for Tier III intensive supports, based on their high level of need, as compared to information and resources for identifying students for Tier II intervention (Newcomer, Freeman, & Barrett, 2013). Thus, this paper focuses on the process of linking universal complete mental health screening to Tier II interventions.

Within an MTSS framework, Tier II interventions are designed to provide secondary supports to students who did not respond to universal Tier I mental and behavioral health approaches, but who are also not currently in need of intensive Tier III individualized services (Hawken, Adolphson, MacLeod, & Schumann, 2009). In addition to preventing more severe emotional and behavioral difficulties for youth, effective Tier I and Tier II interventions also aim to reduce resource (e.g., financial, time) burdens associated with intensive Tier III approaches when they are not needed. Typically, these Tier II interventions are of low-to-moderate intensity and focus on addressing similar needs across groups of students (i.e., students with similar difficulties receive the same intervention and progress monitoring approach; Anderson & Borgmeier, 2010). Tier II interventions commonly include self-monitoring, behavior contracts, mentoring, and small group interventions (i.e., group counseling and social skills groups; Hawken et al., 2009).

Several comprehensive reviews of Tier II mental and behavioral health interventions in schools (e.g., Bruhn, Lane, & Hirsch, 2014; Yong & Cheney, 2013) suggest that there are many available and effective evidence-based Tier II interventions. These include packaged and manualized interventions, such as Check and Connect (Christenson et al., 2008), Check Connect and Expect (Cheney et al., 2009), First Step to Success (Walker et al., 2014); check-in/check-out (Hawken & Horner, 2003); group counseling programs (e.g., Coping Power [Lochman & Wells, 2004], Coping Cat [Kendall & Hedtke, 2006]); as well as non-packaged but evidence-based approaches such as daily behavior report cards (Iznardo, Rogers, Volpe, Labelle, & Robaey, 2017), social skills training, and mentoring programs (Hawken et al., 2009). See Bruhn et al. (2014) and Yong and Cheney (2013) for detailed information about evidence-based Tier II interventions.

Although evidence-based Tier II interventions exist, schools often have difficulty determining which interventions to implement, prioritizing interventions, and deciding which students best fit the aims and goals of selected interventions (Miller, Cook, & Zhang, 2018). Additionally, despite recommendations to use data to inform the selection of intervention (von der Embse et al., 2016), Bruhn et al.’s (2014) review of Tier II interventions found that fewer than half of the 28 included studies used a screening measure (alone or in conjunction with other tools) to identify students for the intervention. Moreover, the most commonly used screening tools were Systematic Screening for Behavioral Disorders (SSBD; Walker & Severson, 1992) and the Student Risk Screening Scale (SRSS; Drummond, 1994), both of which have strong empirical support but are deficit-focused. Thus, none of the studies reviewed by Bruhn et al. (2014) used a universal, complete mental health screening approach that measured both strengths and distress to identify students in need of additional services at the Tier II level. The case study presented in the current paper will demonstrate how screening data can be used to determine which students are languishing and may benefit from Tier II intervention and how to select interventions based on identified student deficits and strengths.

Current Paper

Despite the importance of implementing effective, evidence-based Tier II mental and behavioral health interventions, schools often either are not fully implementing this level of intervention (Lane, Carter, Jenkins, Dwiggins, & Germer, 2015), or have too many different Tier II programs in place (average of 14 reported by Hawken et al., 2009), which burdens schools and reduces fidelity of implementation. The process of selecting appropriate Tier II interventions and identifying appropriate students for these interventions can be complex, unclear, and deficit-focused. Given the utility of complete mental health screening for identifying groups of students who do not report significant distress, but who also report low levels of strengths, the screening to intervention process highlights the benefits of screening to identify and serve youth with languishing mental health. In this paper we provide guidance on four common questions for schools: (a) how do schools implement universal complete mental health screening?; (b) how do schools identify students in need of Tier II services, particularly those students who would not be identified by traditional deficit-focused screening methods (i.e., languishing students)?; (c) how do schools select appropriate Tier II intervention based on the needs of students?; and (d) how do schools evaluate outcomes for students receiving Tier II intervention?

As we will describe, the universal screening process is advantageous in its utility to identify and serve students with varying levels of strengths and distress. As with any screening process, assessment data should also be used to inform schoolwide programming (i.e., Tier 1 interventions) and interventions for youth reporting significant mental health risk or distress (i.e., Tier III interventions), not just Tier II support. Readers are referred to Moore et al. (2015) and Vannest (2012) for additional information about the screening to intervention follow-up at Tier I and Tier III levels of student need, as this paper focuses on Tier II support.

Case Example

Context and Participants

The current project took place within an ongoing partnership between school psychology trainers/researchers at the affiliated University and a local high school. As part of this partnership, school psychology faculty and graduate students worked alongside the partnering high school’s administration, staff, student, and family community to implement a variety of projects focused on supporting the needs of enrolled students. These initiatives grew out of relationships established through fieldwork placements. Over time the University school psychologist trainees recognized a consistent need for Tier II supports. Subsequently, University trainers/researchers recommended moving towards a more formal MTSS process (Nantias et al., 2014). The goal was to use data gathered through universal complete mental health screening to inform referrals to Tier II and Tier III school resources and to guide planning for additional interventions.

The example that follows describes the screening to Tier II intervention processes that occurred at a high school in California during the 2015–2016 school year. At the time of screening, a total of 2,181 students were enrolled across grades 9 through 12. School-level data indicated that enrolled students were 53.8% Hispanic or Latinx; 39.0% non-Hispanic White; 4.6% Asian, Pacific Islander or Filipino; 1.5% Black or African American, and 1.1% American Indian or Alaska Native. In addition, 44.1% were classified as socioeconomically disadvantaged (i.e., eligible to receive free or reduced-price lunch) and 14.2% were classified as English Learners.

In accordance with best-practice recommendations for conducting universal screening (Desrochers & Houck, 2013), the school’s intervention team, including administrators, teachers, and counselors involved with the Student Intervention Committee decided to move forward with a universal screening approach to identify school-wide and individual student needs to inform intervention efforts at each of Tier I, II, and III. The team expressed interest in identifying students at-risk for, or currently experiencing, social-emotional and behavioral problems in addition to supporting the well-being of students who were not reporting distress. Thus, universal complete mental health screening was selected as the primary approach to data collection that would inform intervention referrals.

Universal Complete Mental Health Screening

Key steps for conducting universal complete mental health screening are explicated in Moore et al. (2015) and include determining key participants, selecting screening instruments, seeking parent consent and youth assent, administering the screening instruments, scoring and analyzing the collected data, and following up. In accordance with these recommendations, the members of the school’s intervention team met with their university partners prior to the start of the school year to discuss goals and constructs of interest for the upcoming schoolwide assessment. Together, the team determined that the goals of screening were to (a) better understand students’ well-being and psychological distress and (b) assist with data-based decision making related to prevention, intervention, and promotion activities for students.

Screening tools

Once the screening goals were established, the team set out to identify appropriate screening instruments (Moore et al., 2015). Because complete mental health screening calls for the assessment of both well-being and distress, schools often need to co-administer at least two measures, with one focusing on each dimension of mental health. In the current case example, the research literature was studied to determine what screening measures exist and their pros and cons (e.g., see Moore et al., 2015); instruments were evaluated based on constructs measured, appropriateness for the school population, cost, administration time, and psychometric properties (e.g., reliability and validity). With these considerations in mind, a list of five different screening instruments was developed by the University/school team, which included the distress-focused measures: BASC-2 Behavioral and Emotional Screening System (BASC-2 BESS; Kamphaus & Reynolds, 2007; the BASC-3 BESS [3rd edition] is now available) and Social Emotional Distress Scale – Secondary (SEDS-S; Dowdy, Furlong, Nylund-Gibson, Moore, & Moffa, 2018), and distress/strength or strength-focused measures: Strengths and Difficulties Questionnaire (SDQ; Goodman, 1997), Positive and Negative Affect Schedule (PANAS; Watson, Clark, & Tellegen 1988), and Social Emotional Health Survey-Secondary (SEHS-S; Furlong, You, Renshaw, Smith, & O’Malley, 2014; You et al., 2014). After reviewing these measures in depth, the team eliminated the BESS due to cost, the SDQ due to concerns about its validity for use with Latinx students (Twyford, Buckley, Moffa & Dowdy, 2018), and PANAS due to the fact that it only provides scores for positive and negative affect (which were deemed less useful in determining intervention needs). Ultimately, the SEHS-S was selected as a measure of strengths and paired with the SEDS-S as a measure of distress. Whereas the SEHS-S and SEDS-S were determined to be the best screeners for the school context, student population, and resources and aims of this particular school, all schools and districts are advised to use a similar problem-solving process to determine what screening tools will best meet the needs of their unique setting and goals.

The SEHS-S was selected as a brief, but comprehensive, measure of students’ social-emotional strengths. Additionally, the SEHS-S is free to administer, making it accessible to many schools. The SEHS-S is a 36-item self-report questionnaire measuring 12 positive psychological dispositions that contribute to four second-order positive mental health domains and an overall covitality score. Covitality has been defined as the counterpart to comorbidity and is described as the “synergistic effect of positive mental health resulting from the interplay among multiple positive psychological building blocks” (Furlong et al., 2014b, p. 1013). The positive psychological building blocks captured within covitality include: belief in self (self-efficacy, self-awareness, persistence), belief in others (family coherence, school support, peer support), emotional competence (emotional regulation, empathy, self-control), and engaged living (optimism, gratitude, zest). Responses to the 36 SEHS-S items are reported on either a 4-point (1 = not at all true, 2 = a little true, 3 = pretty much true, and 4 = very much true) or a 5-point scale (1 = not at all, 2 = very little, 3 = somewhat, 4 = quite a lot, 5 = extremely; only for the gratitude and zest subscales); higher scores indicate better social-emotional health in each area. An advantage of the SEHS-S for use in universal screening is the ability to identify overall levels of strengths (i.e., covitality) for individual students as well as individual domains for which intervention may be warranted.

The SEDS-S was selected as a measure of psychological distress, which was co-administered with the SEHS-S. Ten SEDS-S items provide information regarding students’ feelings and behaviors related primarily to internalizing distress (e.g., anxiety, sadness, stress). Externalizing problems were not queried, as research has found that adolescents are less accurate informants about their own problem behaviors than are teachers and parents (e.g., Smith, Pelham, Gnagy, Molina, & Evans, 2000). A self-report measure of well-being and internalizing concerns was selected, over teacher- or parent-report, given that adolescents are ideal informants on these topics (Furlong et al., 2014a; Smith, 2007). Moreover, the existing teacher referral system in place was judged to adequately identify students with externalizing problems, including attendance and problematic behavior (e.g., aggression, conduct, substance use).

A review of all of the available screening instruments is beyond the scope of this article; readers are referred to Jenkins et al. (2014); Levitt, Saka, Romanelli, and Hoagwood (2007); and Severson, Walker, Hope-Doolittle, Kratochwill, and Gresham (2007) for a review of problem-focused instruments and Moore et al. (2015) for a brief review of instruments assessing for well-being. A summary of the specific screeners reviewed in these articles is provided in Table 1.

Table 1.

Summary of Published Articles that Review School Mental Health Screening Measures

Article Measures Reviewed Screening Attributes Discussed
Jenkins et al. (2014) 1. Behavioral and Emotional Screening System (BESS; Kamphaus & Reynolds, 2007) – Age/grade range
2. Behavior Intervention Monitoring Assessment System (BIMAS; McDougal et al., 2011) – Types of forms
3. Strengths and Difficulties Questionnaire (SDQ; Goodman, 1997) – Number of items/administration time
4. Systematic Screening for Behavior Disorders (SSBD; Walker & Severson, 1992) – Areas assessed
5. Social Skills Improvement System Performance Screening Guide (SSIS PSG; Elliott & Gresham, 2008) – Type of score (e.g., T score, percentiles)
– Cost
– Scoring method (i.e., hand, online)
– Standardization sample
– Response scale
– Reliability and validity
Levitt, Saka, Romanelli, and Hoagwood (2007) Broad – Conditions of focus
1. Pediatric Symptom Checklist (PSC; Jellinek et al. 1986) – Informants
2. Strengths and Difficulties Questionnaire (SDQ; Goodman, 1997) Specialized – Age range
Specialized – Administration time
3. Child Behavior Checklist (CBCL; Achenbach, 1991) – Reliability/validity
4. Behavior Assessment System for Children (BASC; Reynolds & Kamphaus, 1998)
5. Diagnostic Predictive Scales (DPS; Lucas et al., 2001)
6. Voice Diagnostic Interview Schedule for Children (DISC; Shaffer et al., 2000)
7. Systematic Screening for Behavior Disorders (SSBD; Walker & Severson, 1990)
Targeted
8. Connors Parent and Teacher Rating Scales (Connors, 1990) or Connors Rating Scales-Revised (Connors et al., 1998a, b)
9. Swanson Nolan and Pelham checklist (SNAP; Swanson, 1992)
10. Vanderbilt ADHD Diagnostic Teacher and Parent Rating Scales (Wolraich et al., 1998)
11. Eyberg Child Behavior Inventory (Eyberg & Pincus, 1999)
12. Multidimensional Anxiety Scale for Children (MASC; March, 1997)
13. State-Trait Anxiety Inventory for Children (Spielberger, 1973)
14. Beck Anxiety Inventory (BAI; Beck & Steer, 1993)
15. Child PTSD Reaction Index (Frederick et al., 1992)
16. Reynolds Adolescent Depression Scale (RADS; Reynolds, 1987, 2002)
17. Children’s Depression Inventory (CDI; Kovacs, 1992)
18. Center for Epidemiologic Studies Depression Scale (CES-D; Radloff, 1977)
19. Columbia DISC Depression Scale (Lucas, 2004; Shaffer et al. 2000)
20. Beck Depression Inventory-II (BDI, BDI-II; Beck et al., 1996)
21. Columbia Health/Suicide Screen (Shaffer et al., 2004)
22. Alcohol Use Disorders Identification Test (AUDIT; Saunders et al., 1993)
23. CRAFFT (Knight et al., 1999)
24. Personal Experience Screening Questionnaire (PES-Q; Winters, 1991)

Severson, Walker, Hope-Doolittle, Kratochwill, and Gresham (2007) 1. Systematic Screening for Behavior Disorders SBD (SSBD; Walker & Severson, 1990) – Purpose/informant
2. School Social Behavior Scale (SSBS; Merrell, 1993) – Sample and psychometrics
3. Revised Behavior Problem Checklist (Quay & Peterson, 1983) – Syndromes
4. Eyberg Child Behavior Inventory (ECBI; Eyberg & Ross, 1978; Eyberg & Pincus, 1999) – Positive findings
– Drawbacks
5. Student Risk Screening Scale (SRSS; Drummond, 1993) – Combinations of tests
6. Conner’s Rating Scales Revised (CRS-R; Conners, 1990) – Special populations

Moore et al. (2015) 1. Multidimensional Students’ Life Satisfaction Scale (MSLSS; Huebner, 1994) – Age/grade
2. Students’ Life Satisfaction Scale (SLSS; Huebner, 1991) – Number of items/administration time
3. Positive and Negative Affect Scale for Children (PANAS-C; Laurent et al., 1999) – Constructs assessed
– Reliability/validity
4. Social Emotional Health Survey-Secondary (SEHS-S; Furlong, You et al., 2014) – Where measure can be located

Consent process

In the summer preceding the academic year in which screening took place, school staff sent home parent consent forms in students’ annual enrollment packets. The empirical literature examining active (i.e., requiring a signed, affirmative “yes” in order for a student to participate) versus passive (i.e., nonresponse indicates consent) parental consent offers advantages and disadvantages for each approach. For example, although active consent attends more closely to the family-school relationship (Levitt et al., 2007) the use of active consent may also lead to bias in the screening sample. That is, research has found that active consent may be biased against students who may benefit most from screening efforts, such that active consent results in disproportionate selection of non-minority culture participants, females, individuals with superior academic achievement, youth from two-parent households, and youth involved in extra-curricular activities (Anderman et al., 1995; Unger et al., 2004). Therefore, in alignment with district protocols and goals to acquire information on the largest percentage of the school population which included approximately 45% of students experiencing socioeconomic disadvantage and 60% from non-White racial backgrounds, the team proceeded with a passive consent process. Enrollment packets included a form describing the scope of the survey and asked parents to return a signed consent form if they did not want their child to participate in the screening. Student assent was also sought prior to administration of the survey, by asking students to mark either, “No, I decline to take the survey” or “Yes, I agree to take the survey.” Of the 2,181 students enrolled, 56 parents declined consent and an additional 55 students declined assent. Accounting for additional students who were chronically absent during the screening period or who dis-enrolled yielded a total of 1,811 students (83% of total school enrollment) who completed the screening survey.

Screening implementation

Consultation with teachers and administrators is critical to develop an efficient and effective protocol for implementing the surveys. First, it is important to minimize the impact of survey completion on instructional time (Dever et al., 2012); this school selected the second period of the school day, within the first three weeks of school, to be the ideal period for survey completion due to the extra time allotted to this class period. The school staff also elected to implement electronic administration of the complete mental health screening survey given previous experiences and difficulties in processing data collected via paper and pencil. Prior experiences surveying the school via paper-and-pencil administration did allow for the entire student body to be surveyed at one time, however the process of entering data was too cumbersome and time consuming for school staff (i.e., data processing took over two months despite hiring additional staff to aid with data entry). A school administrator scheduled online administration to take place over a period of approximately 10 school days, with second period teachers rotating their classes through computer labs or using tablets or laptops brought to their classrooms to complete the survey. Teachers were provided with the master screening schedule and reminded on the day prior to their scheduled screening day about survey procedures and their scheduled screening time. Within the two-hour period, three classrooms were able to complete the survey in each location. Teachers were also provided with screening administration scripts describing the goals of the survey and the importance of the survey in meeting students’ needs and in working to improve their school experience.

The university partners were responsible for processing collected screening data and providing information back to the high school partners regarding intervention planning. Best practice recommendations are to quickly and efficiently organize screening data to inform follow-up efforts with students. School psychologists and school officials are compelled to evaluate their capabilities of following up with students identified as in need of early intervention or treatment prior to conducting universal screening (Cook, Volpe, & Livanis, 2010; Moore et al., 2015).

Identification of Target Groups and Students

In previous dual-factor or dual-continua screening, T scores or logical cut-points have been used when determining criteria for “high” versus “low” scores. For example, Suldo & Shaffer (2008) first classified students according to the presence of internalizing and externalizing distress, with students with T scores above 60 on subscales of either internalizing or externalizing distress classified into a high distress group. This group consisted of approximately 30% of their sample. Given the absence of norming information for the majority of well-being measures, Suldo and Shaffer (2008) classified students falling above the 30th percentile in overall well-being into a high well-being group.

In accordance with previous dual-factor research, in the current case example students were classified into triage groups based upon the intersection of their strength (SEHS-S) and distress (SEDS-S) scores. Although previous dual-factor research has yielded four mental health groups, the participating school was concerned that the number of students identified in need of services might be too broad to appropriately meet student needs. Therefore, in recognition that the number of groups formed is somewhat arbitrary, a total of nine groups were formed to ensure that the number of students in the highest risk groups would be appropriate and manageable for the student support team providing Tier III services (Moore et al., 2015). While it is important for schools to consider their capacity to address the needs of students identified as having the highest level of need via screening, it must be remembered that any student reporting significant symptoms of distress should be followed up with immediately to ensure their safety and well-being. Appropriate follow-up for youth identified as having the highest risk may include referral to Student Support Team, referral to school counseling or mental health staff for additional assessment via specialized or targeted measures (i.e., second-gate assessments; Dowdy, Dever, Raines, & Moffa, 2016; Levitt et al. 2007), referral to community mental health supports, or development of an individualized treatment plan (e.g., referred for a comprehensive evaluation and development of an Individualized Education Program [IEP] with mental health services and goals delineated). Moore et al. (2015) provide an example of follow-up procedures for youth identified as having high risk. Prior to screening, school staff should be prepared to meet this need.

In the current example, the following classification criteria were used to form mental health groups. Research on the SEHS-S and SEDS-S has not yet provided norming information; therefore, a standardized z-score for each student’s overall score on each measure was computed. Previous research has shown that the total covitality score on the SEHS-S is approximately normally distributed (Furlong, Dowdy, & Nylund-Gibson, 2018), therefore standard deviations were used to evenly distribute students with above and below average strengths. Students were categorized as having low strengths (z ≤ −1; ≤ 1 SD), low-average strengths (−1 < z < 0; 1 SD to 0 SD), high-average strengths (0 < z < 1; 0 SD to 1 SD), or high strengths (z ≥ 1; ≥ 1 SD). Drawing upon criteria used when forming distress groups based upon norm-referenced measures (e.g., measures using standardized T scores) in which students are classified as having normal, elevated, or very elevated behavioral and emotional risk, students’ overall SEDS-S z-scores were used to form three distress groups: average distress (z ≤ 1; ≤1 SD), above average distress (1 < z < 2; 1 SD to 2 SD), and high distress (z ≥ 2; ≥2 SD). Using these criteria, students were placed into one of nine groups: Troubled (N = 82; 4.5%), Moderate Risk (N = 51, 2.8%), Lower Risk (N = 77; 4.3%), Languishing (N = 183; 10.1%), Getting By (N = 460; 25.4%), Moderate Thriving (N = 594; 32.7%), Complete Mental Health (N = 282; 15.8%), and two Symptomatic but Content groups (N1 = 60; N2 = 22; 4.5%; see Table 2).

Table 2.

Dual-Factor Mental Health Triage Groups

  Average Distress Above Average Distress High Distress
Low Strengths 4. Languishing
183
2. Moderate Risk
51
1. Troubled
82
Low Average Strengths 5. Getting By 460 3. Lower risk
77
High Average Strengths 6. Moderate Thriving 594 9. Symptomatic but Content
60
8. Symptomatic but Content
22
High Strengths 7. Complete Mental Health
282

Note. Cells are numbered in order of need for follow-up. Shaded cells indicate highest priority for intervention.

Given the administration’s interest in promoting student engagement, the relatively large number of students identified as Languishing (10.1% of those surveyed), and recommendations to be conscientious of the number of students who can feasibly be served with existing school resources (Cook et al., 2010; Moore et al., 2015), school staff elected to take additional steps to inform intervention referrals from survey data. Of most concern, students in the Troubled and Moderate Risk groups were referred to the counseling staff for additional assessment and potential referral to Tier III interventions. This additional in-depth assessment was conducted to make sure screening did not inappropriately screen high-risk students into the Moderate Risk group. Given their above average internalizing distress, students in the Lower Risk and Symptomatic but Content groups were referred to the school counseling team for participation in Tier II group interventions aimed at reducing symptoms of anxiety and depression and supporting continued use of effective coping skills and existing strengths. Students in the Getting By, Moderate Thriving, and Complete Mental Health groups were served through schoolwide programming. Results at the school-level were shared with teachers and school staff and used to guide professional development topics, social-emotional learning curricula for the classroom, and schoolwide assemblies and mental health promotion efforts.

Students classified in the Languishing group were the target of potential strengths-focused Tier II interventions for the 2015–2016 academic year. In the current case study’s sample, students who screened into the Languishing group were represented across grades levels (grade 9 = 29.5%, grade 10 = 32.3%, grade 11 = 22.4%, grade 12 = 15.8%). A majority identified as male (56.3%) and markedly more identified as Latinx (66.1%) than White (19.7%) or another race or ethnicity (13.6%). Examining trends across the nine groups, relatively more females were in groups characteristic of higher risk (e.g., 78.0% of Troubled, 72.5% of Moderate Risk, 66.2% of Lower Risk, compared to 50.7% in Complete Mental Health and 50.2% in Moderate Thriving). The Languishing group (66.1% Latinx, 19.7% White) had a higher proportion of Latinx students than the Complete Mental Health group (33% Latinx, 51.1% White), whereas proportions in the Moderate Thriving group were more equal (40.1% Latinx, 43.6% White).

Selection of Appropriate Tier II Interventions

By definition, languishing students report low levels of psychosocial strengths and/or well-being and average levels of emotional and behavioral risk or distress. Thus, Tier II interventions most appropriate for languishing students include those programs aimed at improving their strengths and well-being (Suldo & Shaffer, 2008). Choosing an appropriate Tier II Intervention is a complicated task that requires educators and administrators to consider a number of factors. First, schools need to determine what type of behavior(s) or problem(s) are the focus of the intervention. Ideally this is guided by screening data as well as information from school-based stakeholders (e.g., students, teachers, parents, administrators) and school data (e.g., attendance, suspensions, grades). The current school used screening data, school data, and stakeholder information to identify the areas of most need for their Languishing students being targeted for Tier II supports. Specifically, the school had an interest in increasing student engagement and school climate through Tier II intervention, with the goal of building on students’ strengths. Therefore, the screening survey included a measure of students’ feelings of connection with their school (e.g., affective engagement). The School Connectedness Scale (SCS) consists of five questions asking students about their feelings of connection toward school (e.g., “I feel like I am a part of this school”; Anderman, 2002; McNeely, Nonnemaker, & Blum, 2002). Analysis of data revealed that about half of the students in the Languishing group reported school connectedness scores below the average for the school (n = 91). Additionally, per counselor feedback, many of these 91 students also had attendance issues and/or were failing one or more classes. Taken together, this information and the school’s broader interests in enhancing student engagement and school climate were used to inform selection of a Tier II intervention for this group of students.

Once schools have determined the focus and goals of their intervention, then they need to determine what evidence-based interventions target these identified goals and whether they are appropriate for the age and characteristics of their student population. There is no standard process for making these decisions, as every school has different student populations and identified needs. Newcomer et al. (2013) suggest that schools create a matrix of existing interventions in their schools and the specific functions of behavior and student needs targeted by each intervention. This matrix can then be used to match identified student needs with specific Tier II supports. If the school does not yet have appropriate interventions in place to meet identified student needs, then officials will need to review the literature on Tier II evidence-based interventions and determine which align with their intervention goals. Table 3 provides a summary of selected evidence-based Tier II interventions for use in high schools, as well as the specific student outcomes that are targeted. For further reading, additional comprehensive reviews and guides that delineate the process of selecting and implementing evidence-based mental and behavioral health interventions in schools and other community-based settings are available through the National Resource Center for Mental Health Promotion and Youth Violence Prevention (n.d.) and Project PRIME (n.d.).

Table 3.

Summary of Selected Evidence-Based Tier II Interventions for High School Students

Intervention Age/Grade Range Format Target Areas Addressed References
Aggression Replacement Training (ART) 12 to 17 years Small group  – Prosocial skills
 – Anger management
 – Moral reasoning


Glick & Gibbs (2011)
Behavior contracts Grades
K-12
Individual  – Behavioral issues
 – Individualized to student needs

Bowman-Perrott, Burke, de Marin, Zhang, & Davis (2015)
Building Decision Skills Middle and high school Small group or classroom  – Ethics
 – Character education
 – Problem-solving

Leming (2001)
Check and Connect Grades
K-12
Individual  – Adult mentorship/relationship
 – Student engagement
 – School dropout/attendance
 – Academic achievement
 – Behavioral issues
 – Persistence

Sinclair et al. (1998)
Check in-Check out (CICO) Grades
K-12
Individual  – Adult mentorship/relationship
 – Behavioral accountability
 – School engagement
 – Academic achievement
 – Individualized to student needs

Cheney et al. (2009); March & Horner (2002); Hawken, MacLeod, & Rawlings (2007)
Cognitive Behavioral Interventions for Trauma in Schools (CBITS) Grades
5–12
Small group, individual, and parent groups  – Trauma symptoms
 – Social problem solving
 – Relaxation and stress management
 – Parent education

Jaycox (2004)
Function-based behavior support plans Grades
K-12
Individual  – Behavioral issues
 – Individualized to student needs

McIntosh, Brown, & Borgmeier (2008); Newcomer & Lewis (2004); Umbreit (1995)
Girls Circle 9 to 18 years Small group  – Resiliency/protective factors
 – Self-esteem
 – Positive interpersonal connection
 – Diversity
 – Mother-daughter relationship

Gies et al. (2015)
Mindfulness-based interventions Grades
K-12
Small group or classroom  – Mindfulness
 – Relaxation
 – Stress management
 – Attention and concentration
 – Emotional coping skills

Zenner et al. (2014)
Self-monitoring Grades
K-12
Individual  – Self-control/self-regulation
 – Self-efficacy
 – Behavioral and academic accountability
 – Individualized to student needs

Pinkelman & Horner (2017)
The C.A.T. Project (modified version of Coping Cat for older children) 14 to 17 years Small group  – Emotional adjustment
 – Anxiety management
 –Cognitive and emotional coping skills
Kendall, Choudhury, Hudson, & Webb (2002)

Finally, schools have to consider implementation feasibility (e.g., cost, personnel) when selecting appropriate interventions. An intervention will not be successful if the school does not have the money to purchase the intervention or have personnel trained, is lacking the needed personnel to implement the intervention, and/or does not have staff buy-in necessary to sustain the program. Therefore, schools should carefully consider the costs, training requirements, fit with existing school structures and programs, and uniqueness of the intervention in addressing something that is not currently being addressed through already existing programs/interventions.

In the case example described here, the school identified mentorship programs broadly, as evidence-based interventions that could address the Languishing students’ needs for enhancing strengths, increasing engagement with school, and improving school connection. The school learned that its University partners had developed Check, Connect, and Respect (CCR) as an adaptation and extension of Check & Connect (Christenson et al., 2008). The U. S. Department of Education’s What Works Clearinghouse (WWC) has found Check & Connect to be efficacious, with positive effects on school dropout (staying in school). Moreover, CCR was developed to not only address risks, but also build the psychological strengths identified students were missing. Thus, mentors also provided lessons on strength building, such as ways to increase hope, gratitude, and connections with peers. The school-University partnership would allow the school to implement the intervention at little to no cost, with volunteers from the University undergraduate and graduate program in psychology serving as the CCR mentors. Ultimately, CCR was selected by school staff for students in the Languishing group who also self-reported below average school connection.

Students identified as potentially benefitting from CCR were contacted by CCR mentors, with the support of the school’s counseling team and students’ teachers. Mentors met individually with identified youth to provide information about the program and to obtain youth assent to participate. When youth were unsure about whether they wanted to participate, mentors answered any remaining questions and scheduled a follow-up meeting with the student to further discuss the program and their interest in participating. Students who did not assent did not participate in CCR and were provided with information about additional school and community resources to support wellness. A parental notification form was sent home with assenting students to inform parents of their child’s participation in the CCR program and included instructions about who to contact if they did not wish for their child to participate in the intervention.

Students in the Languishing group who did not meet criteria to participate in CCR, due to their higher self-reported school connectedness scores, were placed on a waiting list to participate in the CCR intervention during the second half of the school year, given space allotments and their relatively lower need given their self-reported connections to their peers and larger school community. These students were also served by schoolwide programming.

Progress Monitoring and Evaluation

It is critical to evaluate student outcomes related to selected Tier II interventions (Hixson, Christ, & Bruni, 2014). Student outcome data are essential to ensure that follow up is appropriately meeting student needs, at either the group or individual level. For instance, outcome data may demonstrate that the intervention is not targeting the area of foci as anticipated or that an individual student requires additional support. Through evaluating outcome data, schools can make decisions about continuing, modifying, or ending an intervention program, either for individual students or within the school system as a whole.

For the case study being described, in order to evaluate the impact of CCR as delivered in this setting, and to provide services to as many identified students as possible, a descriptive pretest-posttest non-experimental design was implemented. The SEHS-S and SEDS-S screening data were used as pretest data, as students began the CCR intervention shortly after screening took place and were aligned with the schools’ intended goals of monitoring both strength and distress indicators. These measures were then re-administered to individual students as they exited CCR (posttest data). In addition, school-provided attendance data and grades where analyzed to evaluate change for participating students before and after completing CCR. At exit, participating students were also asked to complete the Mentor-Student Relationship Survey, a modified version of the Monitor-Student Relationship Survey developed for Check & Connect (Anderson, Christenson, Sinclair, & Lehr 2004), which asks students to rate how connected and understood they feel with their mentor.

It is also important to monitor the fidelity of implementation of selected interventions, as an evidence-based intervention is only likely to be successful if it is implemented as designed. There are a number of methods for fidelity monitoring, which vary in intensity and rigor. These include, from least to most rigorous, implementer self-report (verbally or in written format to supervisor), self-report fidelity rating scales or checklists completed after each session (or on a specific interval), and observation by a trained observer (audio, video or in-vivo, on a specific interval; Breitenstein et al., 2010). In the current example, fidelity of implementation of the intervention was monitored through mentor notes and supervision. Supervision was provided by University faculty and advanced graduate students trained in CCR implementation. Mentors completed notes, in which they indicated via a checklist which aspects of the intervention were delivered, as well as qualitative information, after each CCR session. They also met weekly in group supervision to discuss their implementation of CCR and receive feedback on next steps and areas for improvement. Individual supervision was also provided, as needed, for mentors requiring additional support or facing particularly challenging student situations. In the current example, fidelity data were not analyzed systematically, but it is a recommended best practice that schools systematically monitor fidelity of implementation of school-based mental and behavioral health interventions and provide support to increase fidelity if poor implementation fidelity is identified (Hixson et al., 2014). Schools may also benefit from using the RE-AIM Checklist, which is a checklist that supports schools in the process of systematically integrating and sustaining Tier 2 social-behavioral interventions (see Cheney & Yong, 2014).

Discussion

As illustrated by the literature and the case example highlighted in this paper, universal complete mental health screening is an important step towards successfully identifying Tier II intervention needs and assigning appropriate students to support. When schools are able to screen students for both strengths and distress indicators they obtain a more nuanced understanding of student needs, which aids in moving to intervention efforts. Although moving from universal complete mental health screening to Tier II intervention requires forethought to implement and evaluate, as this paper shows, the potential of connecting under-identified vulnerable youth, such as those who are languishing, to appropriate Tier II intervention is promising.

Schools are often hesitant to engage in universal screening, due to wanting to ensure that appropriate follow up is available for students in need of support. While schools are often prepared to support students with high levels of need, it can be more challenging to group and provide intervention for students in Tier II. In particular, numerous evidence-based interventions have been developed to address specific mental and behavioral health issues (e.g., anxiety, depression, trauma, conduct problems), which has led to a clearer understanding of how to support students with these symptom presentations. What is less clear is how to support other vulnerable populations who do not yet demonstrate psychological distress but are nonetheless vulnerable and at-risk for experiencing poor outcomes. As this paper demonstrates, schoolwide-complete mental health screening data can be used to identify evidence-based interventions that are aligned with the information gathered from screening data. This allows schools to better meet the needs of students whose lack of well-being may otherwise go un- or under-identified.

The selection of specific Tier II interventions is a challenging part of connecting screening to intervention. When selecting Tier II interventions, school professionals should consider developing a menu of services that address various needs. For example, students who need to work on self-efficacy and self-awareness might be a good fit for CCR. Whereas students who need to work on emotional regulation, empathy, or self-control might benefit from a different Tier II intervention. Although it might be tempting, and overwhelming, to develop Tier II supports to meet the specific needs of every student, it is likely that implementing a limited number of high-quality interventions can address the psychological supports of a student body. For example, Lenzi et al. (2015) provide empirical support for a configuration protective model, which states that an adequate balance of strengths across domains is a protective factor against a variety of behavioral and emotional challenges. Thus, a configuration of supports need not address every domain, but an adequate variety of skills across multiple domains (Lenzi et al., 2015).

Schools’ intervention efforts are often hampered by pragmatic considerations. It is critically important that any intervention be acceptable to the consumers, which in the schools means that it is: acceptable to administrators, teachers, students, and their families; aligned with school schedules; addresses significant concerns; and resource efficient (Sharkey, Dougherty, Felix, & Dowdy, 2019). The most important key to success with this project included administrative support and staff enthusiasm for the project, which is supported in the broader literature (Castillo & Curtis, 2014). When schools do not have strong Tier II supports, there is often pressure on Tier III supports such as special education and individual school-based counseling to take on students who could benefit from less resource intensive programs. In our case, administrators were excited about better meeting the needs of students with less costly interventions; teachers were enthusiastic because they felt this project would meet a need of their students who struggled but did not qualify for more intensive support. In schools where support for complete mental health screening and Tier II support has not been garnered, school psychologists may want to engage in consultation with the school or district and implement needs assessment to gather the input of school stakeholders (Sharkey et al., 2019).

Implementation Considerations

Although the present study was based on a university-school partnership, it is recognized that some might be concerned about implementing Tier II interventions without support from a university. Fortunately, there are a number of examples from the literature that provide viable options, including school mentorship programs implemented by professionals hired by, or already working in the schools (e.g., Goulet, Archambault, Janosz & Christenson, 2018). In addition, the following recommendations are provided to aid in the future implementation of Check, Connect, and Respect and other Tier II interventions. First, practitioners are encouraged to think broadly about who can serve as school-based mentors. This may include using both school-based staff (e.g., school social workers, special education teachers, school psychologists, school counselors) and outreach to local community organizations to identify potential volunteers, paraprofessionals, or other low cost but impactful personnel to support successful implementation. Second, in selecting interventions, schools will maximize success and sustainability by drawing upon existing school resources. Check and Connect, the original model for Check, Connect, and Respect, was designed with existing staff serving as mentors and has been successfully implemented in schools (see Hartwig & Maynard, 2015). Ultimately, when practitioners are considering implementing a Tier II intervention in their own context, it is advisable to start small and then scale up (Goulet et al., 2018), ensuring that the program is viable before trying to increase the size of their reach.

Challenges and Limitations

Overall, there are many challenges and limitations to implementing complete mental health screening to inform Tier II intervention. Obtaining consent from parents or caregivers in a timely manner was difficult and resulted in only 62% of targeted students participating in the evaluation of the implemented intervention. School psychologists should anticipate difficulty obtaining consent with any student, and particularly with the Languishing group targeted for Tier II supports. Thus, protocols for gaining consent should be put in place with creative mechanisms set ahead of time. A protocol might include mailing consents directly to parents with a self-addressed stamped envelope, asking students to hand deliver and return consent forms with an incentive such as a “no homework pass,” inviting parents to meetings at school with a teacher to explain the program and importance of evaluation, or visiting parents at home. The protocol should be laid out ahead of time with a detailed process for the steps to take to obtain consent for all identified students.

Teacher and staff buy-in is another area of difficulty that is often encountered when schools engage in universal screening and corresponding interventions. Although administrators often have good intentions with bringing these efforts to their faculty, individual teachers may not share the same vision. This can pose added difficulties in allowing students to leave class for screening and intervention activities, gathering teacher input, and encouraging teachers to actively support the intervention process. In the case example presented, teacher participation varied, depending on the individual teacher. School psychologists who want to improve upon teacher participation rates can develop protocols for obtaining teacher data such as frequent reminders, multiple modes of survey delivery (e.g., paper and a web-survey link that can be used on any electronic device), and incentives such as recognition at staff meetings or gift cards.

A limitation to the current case study was the limited information available to describe characteristics of languishing students that may be of particular interest to schools. Information regarding the academic characteristics (e.g., academic performance, attendance, discipline referrals) of screened youth were not available for this study. Previous research with middle school students found that languishing youth, when compared to youth with complete mental health, scored lower on standardized reading assessments and had more absences, but did not significantly differ with respect to overall grade point average or standardized math scores (Suldo & Shaffer, 2008). Antaramian et al. (2010) similarly found languishing youth to have significantly lower grade point averages than their complete mental health peers, but found no differences across groups with respect to standardized test scores in reading, math, or science. Although research with middle school students is able to provide some information about trends in academic outcomes for students in Languishing groups, additional research is needed to explore academic profiles of languishing adolescents to further elucidate distinguishing characteristics of languishing students, and to inform selection of interventions that can bolster these students’ educational and social-emotional outcomes.

An additional limitation of this case study was the lack of attention to demographic differences in the identified groups. The mental health group descriptive data highlighted racial ethnic disproportionalities in complete mental health status of the students. Specifically, the Languishing group (66.1% Latinx, 19.7% White) had a much higher proportion of Latinx students than the Complete Mental Health group (33% Latinx, 51.1% White). Interestingly, the Moderate Thriving group included more equal proportions of Latinx (40.1%) and White (43.6%) students. Given that past studies have identified languishing students as experiencing lower levels of engagement (Antaramian et al., 2010), reduced academic self-concept (Suldo & Shaffer, 2008), and lower levels of school belonging (Moffa et al., 2016), schools may need to particularly focus on these factors for historically marginalized populations. Including examination of racial and ethnic disparities within universal screening efforts can help identify specific subgroups within schools that need better services at all levels.

Future Recommendations

Solutions to these challenges will come as more research rigorously tests implementation of universal complete mental health screening to Tier II intervention. Thus, it is important that implementation trials are published and disseminated so applied psychologists can learn from and improve upon existing practice. Future studies should implement quasi-experimental designs to further understand how schools can successfully implement universal screening leading to intervention. Additionally, research should continue to address pragmatic concerns (e.g., teacher buy-in, time, personnel resources, intervention cost) to allow schools to more readily engage in tiered work to support all students. The continued collaboration of schools and researchers to do this important work will ultimately lead to enhanced evidence-based guidelines for effective implementation of complete mental health screening and MTSS for mental and behavioral health.

Funding and Acknowledgements

The authors would like to thank the high school with whom they partnered for their continued support and dedication to student wellness.

The preparation of this article was supported in part by the National Institute of Mental Health (#T32MH109436–02). The opinions expressed are those of the authors and do not represent views of the National Institutes of Health.

Contributor Information

Stephanie A. Moore, Johns Hopkins Bloomberg School of Public Health

Ashley M. Mayworm, Loyola University Chicago

Rachel Stein, University of Colorado Denver.

Jill D. Sharkey, University of California, Santa Barbara

Erin Dowdy, University of California, Santa Barbara.

References

  1. Achenbach TM (1991). Manual for the child behavior checklist/4–18 and the 1991 profile Burlington, VT: University of Vermont Department of Psychiatry. [Google Scholar]
  2. Albers CA, Glover TA, & Kratochwill TR (2007). Introduction to the special issue: How can universal screening enhance educational and mental health outcomes? Journal of School Psychology, 45(2), 113–116. doi: 10.1016/j.jsp.2006.12.002 [DOI] [Google Scholar]
  3. Anderman EM (2002). School effects on psychological outcomes during adolescence. Journal of Educational Psychology, 94, 795–809. doi: 10.1037//0022-0663.94.4.795 [DOI] [Google Scholar]
  4. Anderman C, Cheadle A, Curry S, Diehr P, Shultz L, &Wagner E (1995). Selection bias related to parental consent in school-based survey research. Evaluation Review, 19, 663–674. doi: 10.1177/0193841X9501900604. [DOI] [Google Scholar]
  5. Anderson CM, & Borgmeier C (2010). Tier II interventions within the framework of school-wide positive behavior support: Essential features for design, implementation, and maintenance. Behavior Analysis in Practice, 3, 33–45. doi: 10.1007/BF03391756 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Anderson AR, Christenson SL, Sinclair MF, & Lehr CA (2004). Check & Connect: The importance of relationships for promoting engagement with school. Journal of School Psychology, 42, 95–113. doi: 10.1016/j.jsp.2004.01.002 [DOI] [Google Scholar]
  7. Antaramian SP, Huebner ES, Hills KJ, & Valois RF (2010). A dual‐factor model of mental health: Toward a more comprehensive understanding of youth functioning. American Journal of Orthopsychiatry, 80, 462–472. doi: 10.1111/j.1939-0025.2010.01049.x [DOI] [PubMed] [Google Scholar]
  8. Beck AT, & Steer RA (1993). Manual for the beck anxiety inventory manual San Antonio, TX: The Psychological Corporation. [Google Scholar]
  9. Beck AT, Steer RA, & Brown GK (1996). Manual for the beck depression inventory-II San Antonio, TX: The Psychological Corporation. [Google Scholar]
  10. Breitenstein SM, Gross D, Garvey C, Hill C, Fogg L, & Resnick B (2010). Implementation fidelity in community-based interventions. Research in Nursing & Health, 33(2), 164–173. doi: 10.1002/nur.20373 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bowman-Perrott L, Burke MD, de Marin S, Zhang N, & Davis H (2015). A meta-analysis of single-case research on behavior contracts: Effects on behavioral and academic outcomes among children and youth. Behavior Modification, 39, 247–269. doi:10.1177%2F0145445514551383 [DOI] [PubMed] [Google Scholar]
  12. Bruhn AL, Lane KL, Hirsch SE (2014). A review of tier 2 interventions conducted within multitiered models of behavioral prevention. Journal of Emotional and Behavioral Disorders, 22(3), 171–189. doi: 10.1177/1063426613476092 [DOI] [Google Scholar]
  13. Castillo JM, & Curtis MJ (2014). Best practices in systems-level change. In Harrison PL & Thomas A (Eds.), Best practices in school psychology: Systems-level services (pp. 11–28). Bethesda, MD: National Association of School Psychologists. [Google Scholar]
  14. Cheney DA, Stage SA, Hawken LS, Lynass L, Mielenz C, & Waugh M (2009). A 2-year outcome study of the check, connect, and expect intervention for students at risk for severe behavior problems. Journal of Emotional and Behavioral Disorders, 17, 226–243. doi: 10.1177/1063426609339186 [DOI] [Google Scholar]
  15. Cheney DA, & Yong M (2014). RE-AIM Checklist for integrating and sustaining tier 2 social-behavioral interventions. Behavior Management, 50, 39–44. doi: 10.1177/1053451214532343 [DOI] [Google Scholar]
  16. Christenson SL, Thurlow ML, Sinclair MF, Lehr CA, Kaibel CM, Reschly AL, Mavis A, & Pohl A (2008). Check & Connect: A comprehensive student engagement intervention manual Minneapolis: University of Minnesota, Institute on Community Integration. [Google Scholar]
  17. Conners CK (1990). Conners ratings scales manual North Tonawanda, NY: Multi-health Systems, Inc. [Google Scholar]
  18. Conners CK, Sitarenios G, Parker JD, & Epstein JN (1998a). The Revised Conners’ Parent Rating Scale (CPRS-R): Factor structure, reliability, and criterion validity. Journal of Abnormal Child Psychology, 26, 257–268. [DOI] [PubMed] [Google Scholar]
  19. Conners CK, Sitarenios G, Parker JD, & Epstein JN (1998b). Revision and restandardization of the Conners Teacher Rating Scale (CTRS-R): Factor structure, reliability, and criterion validity. Journal of Abnormal Child Psychology, 26, 279–291. [DOI] [PubMed] [Google Scholar]
  20. Cook CR, Volpe RJ, & Livanis A (2010). Constructing a roadmap for future universal screening beyond academics. Assessment for Effective Intervention, 35, 197–205. doi: 10.1177/1534508410379842. [DOI] [Google Scholar]
  21. Desrochers J, & Houck G (2013). Mental health screening in schools [Handout]. In Desrochers J & Houck G (Eds.), Depression in children and adolescents Washington, DC: National Association of School Psychologists & National Association of School Nurses. [Google Scholar]
  22. Dever BV, Raines TC, & Barclay CM (2012). Chasing the unicorn: practical implementation of universal screening for behavioral and emotional risk. School Psychology Forum: Research in Practice, 6, 108–118. [Google Scholar]
  23. Dowdy E, Furlong M Eklund K, Saeki E, & Ritchey K (2010). Screening for mental health and wellness: Current school based practice and emerging possibilities. In Doll B, Pfohl W, & Yoon J (Eds.), Handbook of youth prevention science (pp. 70–95). New York, NY: Routledge. [Google Scholar]
  24. Dowdy E, Furlong MJ, Nylund-Gibson K, Moore S, & Moffa K (2018). Initial validation of the Social Emotional Distress Survey-Secondary to support complete mental health screening. Assessment for Effective Intervention, 43(4), 241–248. doi: 10.1177/1534508417749871 [DOI] [Google Scholar]
  25. Dowdy E, Dever BV, Raines TC, & Moffa K (2016). A preliminary investigation into the added value of multiple gates and informants in universal screening for behavioral and emotional risk. Journal of Applied School Psychology, 32(2), 178—198. doi: 10.1080/15377903.2016.1165327 [DOI] [Google Scholar]
  26. Drummond T (1993). The Student Risk Screening Scale (SRSS) Grants Pass, OR: Josephine County Mental Health Program. [Google Scholar]
  27. Drummond T (1994). The Student Risk Screening Scale (SRSS) Grants Pass, OR: Josephine County Mental Health Program. [Google Scholar]
  28. Elliott SN, & Gresham FK (2008). SSIS performance screening guide Minneapolis: Pearson Assessments [Google Scholar]
  29. Eyberg SM, & Pincus D (1999). Eyberg Child Behavior Inventory and Sutter-Eyberg Student Behavior Inventory-Revised. Psychological assessment resources Odessa, FL: Psychological Assessment Resources. [Google Scholar]
  30. Eyberg SM, & Ross AW (1978). Assessment of child behavior problems: The validation of a new inventory. Journal of Clinical Child Psychology, 7(2), 113–117. [Google Scholar]
  31. Fletcher JM, & Vaughn S (2009). Response to intervention: Preventing and remediating academic difficulties. Child Development Perspectives, 3(1), 30–37. doi: 10.1111/j.1750-8606.2008.00072.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Frederick C, Pynoos R, & Nader K (1992). Childhood posttraumatic stress reaction index— A copyrighted instrument Los Angeles: University of California. [Google Scholar]
  33. Furlong M, Dowdy E, Carnazzo K, Bovery B, & Kim E (2014a). Covitality: Fostering the building blocks of complete mental health. Communique, 42(8), 28–29. [Google Scholar]
  34. Furlong MJ, Dowdy E, & Nylund-Gibson K (2018). Social Emotional Health Survey-Secondary Manual Santa Barbara, CA: UC Santa Barbara International Center for School-Based Youth Development; Available from, www.project-covitality.info/ [Google Scholar]
  35. Furlong MJ, You S, Renshaw TL, Smith DC, & O’Malley MD (2014b). Preliminary development and validation of the Social and Emotional Health Survey for secondary students. Social Indicators Research, 117, 1011–1032. doi: 10.1007/s11205-013-0373-0 [DOI] [Google Scholar]
  36. Gies SV, Cohen MI, Edberg M, Bobnis A, Spinney E, & Berger E (2015). The Girls Circle: An evaluation of a structured support group program for girls Final report prepared for Office of Juvenile Justice and Delinquency Prevention, Office of Justice Programs; Retrieved from https://onecirclefoundation.org/docs/Research-GC-DSG-2015.pdf [Google Scholar]
  37. Glick B & Gibbs JC (2011). Aggression replacement training: A comprehensive intervention for aggressive youth (3rd ed.). Champaign, IL: Research Press. [Google Scholar]
  38. Goodman R (1997). The Strengths and Difficulties Questionnaire: A research note. Journal of Child Psychology and Psychiatry, 38, 581–586. doi: 10.1111/j.1469-7610.1997.tb01545.x [DOI] [PubMed] [Google Scholar]
  39. Goulet M, Archambault I, Janosz M, & Christenson SL (2018). Evaluating the implementation of Check & Connect in various school settings: Is intervention fidelity necessarily associated with positive outcomes? Evaluation and Program Planning, 68, 34–46. doi: 10.1016/j.evalprogplan.2018.02.004 [DOI] [PubMed] [Google Scholar]
  40. Greenspoon PJ, & Saklofske DH (2001). Toward an integration of subjective well-being and psychopathology. Social Indicators Research, 54, 81–108. doi: 10.1023/A:1007219227883 [DOI] [Google Scholar]
  41. Hartwig EK & Maynard BR (2015). Practitioner perspectives of implementing Check & Connect. Journal of Evidence-Informed Social Work, 12, 438–449. doi: 10.1080/15433714.2013.873752 [DOI] [PubMed] [Google Scholar]
  42. Hawken LS, Adolphson SL, MacLeod KS, & Schumann J (2009). Secondary-tier interventions and supports. In Sailor W, Dunlop G, Sugai G, & Horner R (Eds.), Handbook of positive behavior support (pp. 395–420). New York, NY: Springer Publishing. [Google Scholar]
  43. Hawken LS, & Horner RH (2003). Evaluation of a targeted intervention within a schoolwide system of behavior support. Journal of Behavioral Education, 12, 225–240. doi: 10.1023/A:1025512411930 [DOI] [Google Scholar]
  44. Hawken L, MacLeod K, & Rawlings L (2007). Effects of the Behavior Education Program (BEP) on problem behavior with elementary school students. Journal of Positive Behavior Interventions, 9, 94–101. doi:10.1177%2F10983007070090020601 [Google Scholar]
  45. Hixson MD, Christ TJ, & Bruni T (2014). Best practices in the analysis of progress monitoring data and decision making. In Harrison PL & Thomas A (Eds.), Best practices in school psychology: Foundations (6th ed., pp. 343–354). Bethesda, MD: National Association of School Psychologists. [Google Scholar]
  46. Huebner ES (1991). Initial development of the student’s life satisfaction scale. School Psychology International, 12, 231–240. doi: 10.1177/0143034391123010 [DOI] [Google Scholar]
  47. Huebner ES (1994). Preliminary development and validation of a Multidimensional Life Satisfaction Scale for children. Psychological Assessment, 6, 149–158. doi: 10.1037/1040-3590.6.2.149 [DOI] [Google Scholar]
  48. Iznardo M, Rogers MA, Volpe RJ, Labelle PR, & Robaey P (2017). The effectiveness of daily behavior report cards for children with ADHD: A meta-analysis. Journal of Attention Disorders, 1–14. doi: 10.1177/1087054717734646 [DOI] [PubMed] [Google Scholar]
  49. Jaycox L (2004). Cognitive Behavioral Intervention for Trauma in Schools Longmont, CO: Sopris West Educational Services. [Google Scholar]
  50. Jenkins LN, Demaray MK, Wren NS, Secord SM, Lyell KM, Magers AM, … & Tennant J (2014). A critical review of five commonly used social-emotional and behavioral screeners for elementary or secondary schools. Contemporary School Psychology, 18, 241–254. doi: 10.1007/s40688-014-0026-6. [DOI] [Google Scholar]
  51. Jellinek MS, Murphy JM, & Burns BJ (1986). Brief psychosocial screening in outpatient pediatric practice. Journal of Pediatrics, 109, 371–378. [DOI] [PubMed] [Google Scholar]
  52. Kamphaus RW, & Reynolds CR (2007). Behavior Assessment System for Children—Second Edition (BASC–2): Behavioral and Emotional Screening System (BESS) Bloomington, MN: Pearson. [Google Scholar]
  53. Kamphaus RW, Reynolds CR, & Dever BV (2014). Behavioral and mental health screening. In Kettler RJ, Glover TA, Albers CA, Feeney-Kettler KA (Eds.), Universal screening in educational settings: Evidence-based decision making for schools (pp. 249–273). Washington, DC, US. [Google Scholar]
  54. Kendall PC, Choudhury M, Hudson J, & Webb A (2002). The C.A.T. Project Manual for the Cognitive-Behavioral Treatment of Anxious Adolescents (ages 14–17) Ardmore, PA: Workbook Publishing. [Google Scholar]
  55. Kendall PC, & Hedtke K (2006). Coping Cat workbook (2nd ed). Ardmore, PA: Workbook Publishing. [Google Scholar]
  56. Keyes CLM (2002). The mental health continuum: From languishing to flourishing in life. Journal of Health and Social Research, 43, 207–222. doi: 10.2307/3090197 [DOI] [PubMed] [Google Scholar]
  57. Keyes CLM (2005). Mental illness and/or mental health? Investigating axioms of the Complete state model of health. Journal of Consulting and Clinical Psychology, 73, 539–548. doi: 10.1037/0022-006X.73.3.539 [DOI] [PubMed] [Google Scholar]
  58. Kim EK, Dowdy E, Furlong MJ, & You S (2017). Mental health profiles and quality of life among Korean adolescents. School Psychology International, 38, 98–116. doi: 10.1177/0143034316682296 [DOI] [Google Scholar]
  59. Knight JR, Shrier L, Bravender T, Farrell M, Vanderbilt J, & Shaffer H (1999). A new brief screen for adolescent substance abuse. Archives of Pediatrics and Adolescent Medicine, 153, 591–596. [DOI] [PubMed] [Google Scholar]
  60. Kovacs M (1992). Children’s Depression Inventory (CDI) Manual North Tonawanda, NY: Multi-Health Systems, Inc. [Google Scholar]
  61. Lane KL, Carter EW, Jenkins A, Dwiggins L, & Germer K (2015). Supporting comprehensive, integrated, three-tiered models of prevention in schools: Administrators’ perspectives. Journal of Positive Behavior Interventions, 17, 209–222. doi: 10.1177/1098300715578916 [DOI] [Google Scholar]
  62. Laurent J, Catanzaro SJ, Joiner TE Jr., Rudolph KD, Potter KI, Lambert S, Osborne L, & Gathright T (1999). A measure of positive and negative affect for children: scale development and preliminary validation. Psychological Assessment, 11, 326–338. doi: 10.1037/1040-3590.11.3.3 [DOI] [Google Scholar]
  63. Leming JS (2001). Integrating a structured ethical reflection curriculum into high school community service experiences: Impact on students’ socio-moral development. Adolescence, 36(141), 33–45. [PubMed] [Google Scholar]
  64. Lenzi M, Furlong MJ, Dowdy E, Sharkey JD, Gini G, & Altoè G (2015). The quantity and variety across domains of psychological and social assets associated with school victimization. Psychology of Violence, 5, 411–421. doi: 10.1037/a0039696 [DOI] [Google Scholar]
  65. Levitt JM, Saka N, Romanelli LH, & Hoagwood K (2007). Early identification of mental health problems in schools: the status of instrumentation. Journal of School Psychology, 45, 163–191. doi: 10.1016/j.jsp.2006.11.005 [DOI] [Google Scholar]
  66. Lochman JE, & Wells KC (2004). The Coping Power Program for preadolescent boys and their parents: Outcome effects at the 1-year follow-up. Journal of Consulting and Clinical Psychology, 72(4), 571–578. doi: 10.1037/0022-006X.72.4.571 [DOI] [PubMed] [Google Scholar]
  67. Lucas CP (2004). Measures and techniques for screening for suicide risk and undetected psychopathology. Paper presented at the Annual Meeting of the Society for Prevention Research Quebec City, Canada. [Google Scholar]
  68. Lucas CP, Zhang H, Fisher PW, Shaffer D, Regier DA, Narrow WE, et al. (2001). The DISC Predictive Scales (DPS): Efficiently screening for diagnoses. Journal of the American Academy of Child and Adolescent Psychiatry, 40, 443–449. [DOI] [PubMed] [Google Scholar]
  69. March JS (1997). Multidimensional anxiety scale for children: Technical manual North Tonawanda, NY: Multi-Health Systems. [Google Scholar]
  70. March RE, & Horner RH (2002). Feasibility and contributions of functional behavioral assessment in schools. Journal of Emotional and Behavioral Disorders, 10, 158–170. [Google Scholar]
  71. McDougal JL, Bardos AN, & Meier ST (2011). Behavior Intervention Monitoring Assessment System technical manual Toronto: Multi-Health Systems [Google Scholar]
  72. McIntosh K, Brown JA, & Borgmeier CJ (2008). Validity of functional behavior assessment within a response to intervention framework. Assessment for Effective Intervention, 34, 6–14. doi: 10.1177/1534508408314096 [DOI] [Google Scholar]
  73. McNeely C, Nonnemaker J, & Blum R (2002). Promoting school connectedness: Evidence from the National Longitudinal Study of Adolescent Health. Journal of School Health, 72, 138–146. doi: 10.1111/j.1746-1561.2002.tb06533.x [DOI] [PubMed] [Google Scholar]
  74. Merrell KW (1993). Using behavioral rating scales to assess social skills and antisocial behavior in school settings: Development of the School Social Behavior Scales. School Psychology Review, 22(1), 115–133. [Google Scholar]
  75. Miller FG, Cook CR, & Zhang Y (2018). Initial development and evaluation of the student intervention matching (SIM) form. Journal of School Psychology, 66, 11–24. doi: 10.1016/j.jsp.2017.10.005 [DOI] [PubMed] [Google Scholar]
  76. Moffa K, Dowdy E, & Furlong M (2016). Exploring the contributions of school belonging to complete mental health screening. The Educational and Developmental Psychologist doi: 10.1017/edp.2016.8 [DOI] [Google Scholar]
  77. Moore SA, Widales-Benitez O, & Carnazzo KW, Kim EK, Moffa K, & Dowdy E (2015). Conducting universal complete mental health screening via student self-report. Contemporary School Psychology, 19, 253–267. doi: 10.1007/s40688-015-0062-x [DOI] [Google Scholar]
  78. Nantais M St. Martin K,A, Barnes AC (2014). Best practices in facilitating professional development of school personnel in delivering multitiered services. In Harrison PL & Thomas A (Eds.), Best Practices in School Psychology: Systems–Level Services (pp. 71–82). Bethesda, MD: National Association of School Psychologists [Google Scholar]
  79. National Resource Center for Mental Health Promotion and Youth Violence Prevention. (n.d.). Selecting evidence-based programs Retrieved from the National Center for Healthy Safe Children website: https://healthysafechildren.org//sites/default/files/Selecting_EBPs_Website_508.pdf [Google Scholar]
  80. Newcomer LL, Freeman R, & Barrett S (2013). Essential systems for sustainable implementation of tier 2 supports. Journal of Applied School Psychology, 29, 126–147. doi: 10.1080/15377903.2013.778770 [DOI] [Google Scholar]
  81. Newcomer L, & Lewis TJ (2004) Functional behavioral assessment: An investigation of assessment reliability and effectiveness of function-based interventions. Journal of Emotional and Behavioral Disorders, 12, 168–181. doi:10.1177%2F10634266040120030401 [Google Scholar]
  82. Pinkelman SE, & Horner RH (2017). Improving implementation of function-based interventions: Self-monitoring, data collection, and data review. Journal of Positive Behavior Interventions, 19, 228–238. doi:10.1177%2F1098300716683634 [Google Scholar]
  83. PRIME Project. (n.d.). How to select an evidence-based intervention: A guide Retrieved from the University of Connecticut’s Implementation Science website: https://implementationscience.uconn.edu/wp-content/uploads/sites/1115/2014/12/PRIME_quickguide_edvidence-based_intervention.pdf [Google Scholar]
  84. Quay H & Peterson R (1983). Manual for revised Behavior Checklist Privately released (P.O. Box Coral Cables, 33124) [Google Scholar]
  85. Radloff LS (1977). The CES-D scale: A self-report depression scale for research in the general population. Applied Psychological Measurement, 1, 385–401. [Google Scholar]
  86. Reynolds CR, & Kamphaus RW (1998). Behavior assessment system for children: Manual, Second Edition Circle Pines, MN: American Guidance Service, Inc. [Google Scholar]
  87. Reynolds WM (1987). Reynolds adolescent depression scale: Professional manual Odessa, FL: Psychological Assessment Resources, Inc. [Google Scholar]
  88. Saunders JD, Aasland OG, Babor TF, de la Fuente JR, & Grant M (1993). Development of the Alcohol Use Disorders Identification Test (AUDIT): WHO collaborative project on early detection of persons with harmful alcohol consumption-II. Addiction, 88, 791–804. [DOI] [PubMed] [Google Scholar]
  89. Severson HH, Walker HM, Hope-Doolittle J, Kratochwill TR, & Gresham FM (2007). Proactive, early screening to detect behaviorally at-risk students: issues, approaches, emerging innovations, and professional practices. Journal of School Psychology, 45, 193–223. doi: 10.1016/j.jsp.2006.11.003. [DOI] [Google Scholar]
  90. Shaffer D, Fisher P, Lucas CP, Dulcan MK, & Schwab-Stone ME (2000). NIMH Diagnostic Interview Schedule for Children Version IV (NIMH DISC-IV): Description, differences from previous versions, and reliability of some common diagnoses. Journal of the American Academy of Child and Adolescent Psychiatry, 39, 28–38. [DOI] [PubMed] [Google Scholar]
  91. Shaffer D, Scott M, Wilcox H, Maslow C, Hicks R, Lucas CP, … Greenwald S (2004).The Columbia Suicide Screen: Validity and reliability of a screen for youth suicide and depression. Journal of the American Academy of Child and Adolescent Psychiatry, 43, 71–79. doi: 10.1097/00004583-200401000-00016 [DOI] [PubMed] [Google Scholar]
  92. Sharkey JD, Dougherty D, Felix ED, & Dowdy E (2019). Designing Local Programming: Tools and Models. In Mayer MJ, Jimerson SR (Eds.). School Safety and Violence Prevention: Science, Practice, Policy (199–222). Washington, DC: American Psychological Association. [Google Scholar]
  93. Sinclair MF, Christenson SL, Evelo DL, & Hurley CM (1998). Dropout prevention for youth with disabilities: Efficacy of a sustained school engagement procedure. Exceptional Children, 65, 7–21. doi: 10.1177/001440299806500101 [DOI] [Google Scholar]
  94. Smith SR (2007). Making sense of multiple informants in child and adolescent psychopathology: A guide for clinicians. Journal of Psychoeducational Assessment, 25, 139–149. doi: 10.1177/0734282906296233 [DOI] [Google Scholar]
  95. Smith BH, Pelham WE Jr., Gnagy E, Molina B, & Evans S (2000). The reliability, validity, and unique contributions of self-report by adolescents receiving treatment for attention-deficit/hyperactivity disorder. Journal of Consulting and Clinical Psychology, 68(3), 489–499. doi: 10.1037//0022-006X.68.3.489 [DOI] [PubMed] [Google Scholar]
  96. Spielberger CD (1973). Manual for the state-trait anxiety inventory for children Palo Alto, CA: Consulting Psychologists Press. [Google Scholar]
  97. Suldo SM, & Shaffer EJ (2008). Looking beyond psychopathology: The dual-factor model of mental health in youth. SchoolPsychology Review, 37, 52–68. [Google Scholar]
  98. Swanson JM (1992). School based assessment and interventions for ADD students Irvine, CA: KC Publications. [Google Scholar]
  99. Twyford JM, Buckley L, Moffa K, & Dowdy E (2018). The Strengths and Difficulties Questionnaire (SDQ) self-report: Factor structure of the self-report from Latinx youth. International Journal of School and Educational Psychology doi: 10.1080/21683603.2017.1414005 [DOI] [Google Scholar]
  100. Umbreit J (1995). Functional assessment and intervention in a regular classroom setting for the disruptive behavior of a student with attention deficit hyperactivity disorder. Behavioral Disorders, 20, 267–278. doi:10.1177%2F019874299502000407 [Google Scholar]
  101. Unger JB, Gallaher PG, Palmer PH, Baezconde-Garbanati L, Trinidad DR, Cen S, & Johnson CA (2004). No news is bad news: characteristics of adolescents who provide neither parental consent nor refusal for participation in school-based survey research. Evaluation Review, 28, 52–63. doi: 10.1177/0193841X03254421. [DOI] [PubMed] [Google Scholar]
  102. Vannest KJ (2012). Implementing interventions and progress monitoring subsequent to universal screening. School Psychology Forum: Research in Practice, 6(4), 119—136. [Google Scholar]
  103. Venning A, Wilson A, Kettler L, & Eliott J (2013). Mental health among youth in South Australia: A survey of flourishing, languishing, struggling, and floundering. Australian Psychologist, 48, 299–310. doi: 10.1111/j.1742-9544.2012.00068.x [DOI] [Google Scholar]
  104. Volpe RJ, Briesch AM, & Chafouleas SM (2010). Linking screening for emotional and behavioral problems to problem-solving efforts: An adaptive model of behavioral assessment. Assessment for Effective Intervention, 35(4), 240–244. doi: 10.1177/1534508410377194 [DOI] [Google Scholar]
  105. von der Embse NP, Iaccarino S, Mankin A, Kilgus SP & Magen E (2016). Development and validation of the Social, Academic, and Emotional Behavior Risk Screener – Student Rating Scale. Assessment for Effective Intervention, 42, 186–192. doi: 10.1177/1534508416679410 [DOI] [Google Scholar]
  106. Walker HM, & Severson HH (1990). Systematic screening for behavior disorders (SSBD): Users guide and technical manual Longmont, CO: Sopris West. [Google Scholar]
  107. Walker HM, Severson H, Seeley J, Feil E, Small J, Golly A, … & Forness S (2014). The evidence base of the First Step to Success early intervention for preventing emerging antisocial behavior patterns. In Walker HM & Gresham FM (Eds.), Handbook of evidence-based practices for emotional and behavioral disorders: Applications in schools (pp. 518–537). New York, NY: Guilford. [Google Scholar]
  108. Walker HM, & Severson HH (1992). Systematic Screening for Behavior Disorders (2nd ed.). Longmont, CO: Sopris West. [Google Scholar]
  109. Watson D, Clark LA, & Tellegen A (1988). Development and validation of brief measures of positive and negative affect: The PANAS scales. Journal of Personality and Social Psychology, 54(6), 1063–1070. doi 10.1037/0022-3514.54.6.1063 [DOI] [PubMed] [Google Scholar]
  110. Westerhof GJ, & Keyes CLM (2010). Mental illness and mental health: The two continua model across the lifespan. Journal of Adult Development, 17, 110–119. doi: 10.1007/s10804-009-9082-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Wolraich ML, Feurer ID, Hannah JN, Baumgaertel A, & Pinnock TY (1998). Obtaining systematic teacher reports of disruptive behavior disorders utilizing DSM-IV. Journal of Abnormal Child Psychology, 26, 141–152. [DOI] [PubMed] [Google Scholar]
  112. Yong M, & Cheney DA (2013). Essential features of tier 2 social-behavioral interventions. Psychology in the Schools, 50(8), 844–861. doi: 10.1002/pits.21710 [DOI] [Google Scholar]
  113. You S, Furlong MJ, Dowdy E, Renshaw T, Smith DC, & O’Malley MD (2014). Further validation of the Social and Emotional Health Survey for high school students. Applied Quality of Life Research, 9, 997–1015. doi: 10.1007/s11482-013-9282-2 [DOI] [Google Scholar]
  114. Zenner C, Herrnleben-Kurz S, & Walach H (2014). Mindfulness-based interventions in schools – a systematic review and meta-analysis. Frontiers in Psychology, 5 (603), 1–20. doi: 10.3389/fpsyg.2014.00603 [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES