Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

The social behavior challenges educators face on a daily basis across American schools has long been established. Overall levels of minor but chronic disrespect, noncompliance, and aggression continue to rise and are linked to later more intensive behavioral challenges (Benedict et al. 2007; Conroy et al. 2004; Duda et al. 2004; Muscott et al. 2009; Heaviside et al. 1998). Further, for those students who are identified as having an emotional/behavioral disorder (EBD) and receiving specialized instruction, intervention impact has been limited (Bradley et al. 2004; Wagner et al. 2005). Recent data indicate that over half of students receiving special education services under the EBD category fail to graduate with a high school diploma (Wagner et al. 2005), 20% have been arrested at least once while a student (VanAcker 2004), and the majority will require ongoing mental health and social assistance across their lifetimes (Walker et al. 2004).

To date, as described above, the overall impact of specialized supports for students with EBD have not been universally effective in altering significant behavioral challenges. However, services provided under the Individuals with Disabilities Education Act (IDEA) have lead to modest improvements (Wagner et al. 2005). A further challenge for educators tasked with providing specialized and intensive individualized supports is the overall gross under-identification within the IDEA category of “seriously emotionally disturbed” (EBD). Since passage of the law, fewer than 1 % of students have been identified as having EBD (U.S. Department of Education 2005). Experts within the EBD field estimate that 5–7 % of students manifest emotional and behavioral concerns significant enough to warrant special education services (Kauffman 2005; Walker et al. 2004). Simple mathematics indicates that using 5 % as the expected prevalence across the school-age population equals 2,810,149 children not receiving services under the Serious Emotional Disturbance (SED) label who might otherwise be eligible (U.S. Department of Education 2005). And while outcomes to date indicate improvements are needed in serving this population, the complete absence of specialized supports will surely exacerbate the social, emotional, and behavioral challenges among students at high risk.

The combination of increased overall levels of problem behavior and the large numbers of students at risk for significant emotional and behavioral challenges who currently may not be receiving specialized services has been a driving factor in establishing school-wide positive behavior support (SW-PBS) as both a preventative measure and universal intervention to support high-risk students (Sugai et al. 2000a). SW-PBS provides educators with a problem-solving framework that matches evidence-based practices to presenting problems unique to each school setting (see Chap. ## for a more in depth overview of SW-PBS). An additional feature of SW-PBS is the attention to creating school-wide systems that provide the necessary training and technical assistance to all faculty and staff within the school and ideally across the school district. School teams apply the problem-solving logic of SW-PBS: (a) data-based decisions to identify behavioral concerns and to progress monitor, (b) implementation of evidence-based practices matched to data, and (c) implementing systems of support to increase implementation fidelity across a continuum of available interventions creating a parallel to the academic response-to-intervention (RTI) multi-tiered system framework. Universal supports target common social behavioral challenges by explicitly teaching pro-social alternative behaviors, providing multiple opportunities for student practice, and providing high rates of positive specific feedback on skill use. In addition, universal expectations, instructional strategies, and environmental supports are adapted across structured (e.g., classroom) and unstructured (e.g., playground, cafeteria) school environments. Implementation of universal SW-PBS is monitored through student outcomes as well as annual fidelity checks (e.g., School-wide Evaluation Tool, Horner et al. 2004). For those students who are not successful with universal supports alone, a continuum of social, emotional, cognitive, and behavioral interventions are put in place through tier 2, or small-group interventions, and tier 3 individualized strategies.

The remainder of this chapter focuses on building a complete continuum of SW-PBS with emphasis on tiers 2 and 3. It is important to note that practices and systems of support overviewed within this chapter are intended to be part of an interconnected system of behavioral supports to create an integrated multi-tiered system of support (MTSS). Schools should be implementing universal supports with fidelity as a foundation to establishing a complete continuum of supports prior to systemic development of additional connected tiers. Ongoing attention to maintaining universal supports with high fidelity is also a critical prerequisite for successful installation of tier 2 and tier 3 interventions. An additional consideration in the establishment of tiers 2 and 3 is attention to the phases of implementation (see Chap. ##) and matching training and coaching for schools based on their readiness. Finally, within the continuum of SW-PBS, educators should view additional tiers beyond universal not as discrete and separate but rather as intensifying instruction that reinforces the universal support strategies (i.e., teaching, practicing, providing feedback, and modifying environments) provided at tier 1 to match the intensity of student need. Likewise, educators should not view tiers and students as synonymous. In other words, there are no “tier 2 students” but there will be students who will require additional levels of support across the school day to increase the likelihood they are successful.

Essential Features of Tier 2 and 3 SW-PBS Systems

Tier 2 and 3 supports within a multi-tiered system are designed to provide additional instruction and supports for students who are struggling to meet the goals and objectives taught in the core academic and behavioral curriculum (Chard 2013). Within SW-PBS, school teams are encouraged to implement a variety of interventions and supports based on student need. Regardless of the specific tier 2 and 3 supports used, as these interventions likely vary by school, there are several fundamental components associated with the identification of students requiring these levels of support and subsequent documentation of effectiveness. Essential features of tier 2 and 3 systems include: (a) insuring universal strategies are in place with high fidelity, (b) commitment from school administrators and a specialized behavior support team (tier 2/3 SW-PBS team) to lead intervention efforts, (c) a proactive method to identify in a timely manner the students who require additional supports, (d) a process for matching intervention with student need, (e) selection, adoption, and use of research-based interventions, (f) regular monitoring of student progress to assess performance and rate of improvement, and (g) coordinated decision-making to determine student movement (e.g., more or less intensive supports) within the multi-tiered system (Horner et al. 2010; Crone et al. 2010). Along with these key features, effective tier 2/3 systems also include plans for monitoring fidelity of implementation across interventions and of the overall system, as well as periodic evaluation of impact and effects relative to the goals of improving social and academic outcomes (see Fig. 1 for additional features).

Fig. 1
figure 1

Key features of effective Tier 2 systems of support within a continuum of SW-PBS. SW-PBS school-wide positive behavior support (Adapted from Crone et al. 2010)

Universal SW-PBS Implemented with Fidelity

Following the logic of RTI research within schools implementing SW-PBS, the process begins with all students having access to a rigorous and relevant academic and social–behavioral curriculum in an organized, effective, and positive classroom environment (Chard 2013; Lewis and Sugai 1999). In terms of behavior and social skills, all staff implement universal strategies that include teaching clearly defined expectations, rules, and procedures; explicit instruction for meeting school-wide behavioral goals; high rates of recognition for social and behavioral success; and consistent response for incorrect behavior that includes reteaching and practice opportunities for students who need these. These universal-level supports are implemented with fidelity by all staff, for all students, across all school settings to ensure each child participates in high-quality instruction before determining that he or she requires additional intervention. In particular, all classrooms consistently implement essential instructional management strategies, so that any additional supports or recommended environmental modifications as part of the tier 2/3 process will be positioned to have maximal impact (Simonsen et al. 2008). Specifically, in the same manner of having expectations and rules that are communicated consistently and taught school-wide across settings (e.g., cafeteria, hallway, restroom, playground), individual teachers also use an instructional approach for promoting desired behaviors and social skills within their classrooms. Productive classroom environments are established by using strategies that structure the learning environment so that a majority of problems are prevented from occurring. Strategies include using behavior-specific praise to recognize desired behaviors, providing frequent opportunities to respond during instruction, using material that is correctly matched with student’s instructional level, providing choices to students (e.g., order of task completion), alternating easier tasks with those that are more challenging, pacing instruction adequately, implementing clear routines and procedures, and designing physical arrangements of the room that permit active supervision by adults (Kern and Clemens 2007; Simonsen et al. 2008).

Following adequate academic and behavioral instruction, assessment data are gathered and reviewed on a regular basis to evaluate each student’s success in core instruction. Typically students are not considered for additional intervention (i.e., tier 2) until they have had sufficient time to respond to the universal strategies (e.g., approximately six to 8 weeks) implemented during core instruction. During this time, it is critical to confirm that universal interventions or core instructional strategies are implemented with fidelity.

Tier 2/3 SW-PBS Team

Initial development of SW-PBS begins with the formation of a team to focus on core instruction for social–behavioral skills. As the need for tier 2/3 intervention emerges, it is recommended that schools maintain the universal support-focused SW-PBS team but form a second group to address tiers 2 and 3. The tier 2/3 team should maintain linkages with the universal SW-PBS team by insuring all tier 2/3 strategies are couched within the set of school-wide expectations and that all staff are aware of their roles in supporting tier 2/3 interventions by including common members such as the building administrator, but also capitalize on existing behavioral expertise in the school, district, or regional SW-PBS initiative. Tier 2/3 teams often include school psychologists, special educators, behavior specialists, counselors, and one or more content-area specialists (e.g., language arts or mathematics).

Personnel who serve on the tier 2/3 teams establish systems and practices for students requiring more intensive social, emotional, and/or behavioral support. Members of this group ensure timely access to interventions, oversee implementation of selected interventions, regularly use data to monitor student progress during intervention, and evaluate overall program outcomes (Crone and Horner 2003). Membership of the team is determined according to ability and interest in fulfilling key roles and completing specific responsibilities. At a minimum, a tier 2/3 team typically includes one or more school administrators who are able to make decisions about staff, student, and school-wide schedules, reallocate resources, and set an overall tone of commitment and importance about the development, implementation, and ongoing monitoring of behavioral interventions. Also, it is recommended that the tier 2/3 team include staff with specific expertise in academic and behavioral assessment and intervention. While a primary focus of the team is development and implementation of social, emotional, and behavioral supports, many of the students identified for additional interventions will be at risk also for or already experiencing poor academic outcomes. The specialized behavior support team needs at least one member who can identify academic skill deficits and match students with appropriate academic intervention in addition to behavioral supports that may be needed. Finally, the tier 2/3 team should be structured to include general education teacher representation. Teachers are responsible often for making initial referrals for assistance, selecting appropriate strategies to meet student needs, and for implementing and monitoring effects of interventions (Debnam et al. 2013). Regardless of educational role, the more critical aspects when deciding on membership are structuring the team in a way that allows collaborative and data-based decision-making, division of a shared work load, and commitment to an instructional approach for behavior management and discipline (McIntosh et al. 2013)

Student Identification for Tier 2/3 Supports

Once SW-PBS practices, data-decision making and systems of support are established, implemented with fidelity, and have a measurable impact on student outcomes, tier 2/3 teams should develop a proactive process to actively seek out students who either are not responding to the core instruction or are displaying patterns of behavior that warrant additional intervention to lessen risk (e.g., Glover and Albers 2007; Kratochwill et al. 2004; Simmons et al. 2000; Walker and Shinn 2002). To identify students in need of tier 2/3 behavioral interventions, tier 2/3 teams develop a comprehensive system so that all students in a classroom, school, or district have an equal opportunity to be considered against characteristics of risk (e.g., acting out, argumentative, verbally or physically aggressive, overly shy, worried or withdrawn, anxious, unusually sad; see Fig. 2). The identification process is designed to promote early access to readily available interventions before student problems develop to a level that requires intensive intervention. It also includes clear criteria for indicating which students require immediate, intensive assistance (e.g., safety concerns for student or others). In addition, the identification approach is intended to identify students with internalizing (e.g., social withdrawal, anxiety) and externalizing (e.g., disruption, aggression) concerns regardless of the presence or absence of risk factors (Kamphaus et al. 2010).

Fig. 2
figure 2

Multi-method approach for early identification of social–emotional and behavioral risk

Office Discipline Referral Data

Within an SW-PBS framework, one commonly used measure of student social behavior performance is an office discipline referral (ODR), which is a standardized record of events of problem behavior. Documentation of disciplinary events is carried out consistently across students in that behavioral infractions are identified by common definitions and a standard set of information about these incidents is recorded and collected (i.e., type of problem, location, time of day, others involved, and possible motivation). Parallel to the tiered public health model that matches level of need with level of support (Gordon 1983), SW-PBS researchers and school teams frequently use number of ODRs accrued to define the level of support each student may require (McIntosh et al. 2009b). Regular review of ODR data is the primary method for determining impact of tier 1, core behavioral instruction. Review of ODRs also serves as a way to identify individual students who continue to display social behavioral problems even with exposure to high-quality core instruction (McIntosh et al. 2009). Considering that ODRs are already commonly collected, easily available and cost-efficient, use of these data for identifying students in need of tier 2 intervention may be more likely than use of other measures that may require additional funding or collection procedures (McIntosh et al. 2009). The following criteria, derived in part from proportions of ODR distributions in a large sample of schools, are currently viewed as a method for monitoring level of support required (Horner et al. 2005). Students receiving zero to one ODR per year are considered as responsive to tier 1. That is, the foundational level of support (i.e., structuring of environment to include instruction for behavioral expectations and recognition for successful demonstrations) provided to all students is sufficiently meeting the needs of students with low or no frequency of disciplinary events. Students with documented disciplinary events in the range of two to five incidents are considered for receiving tier 2 supports. Students with six or more ODRs are recommended for tier 3 interventions. Although ODRs offer valuable information about the frequency, topography, location, and potential motivation for behavior, they tend to be more representative of students with externalizing, rather than internalizing, behaviors (Walker et al. 2005).

Universal Screening

Another method for systematically identifying students who may require additional support is use of a brief, behavioral screening instrument. Typically, screening instruments require a response to short statements about emotional or behavioral characteristics of a student. These instruments can be used to generate risk scores for all students in a classroom, grade level, building, or district.

There are several potential advantages for developing a systematic identification process that incorporates universal screening. First, screening instruments are generally perceived as a quick, accurate, and respectful process with capacity to include all children and youth of interest. Second, if an error occurs most often it is on the side of caution with the tendency to overidentify rather than missing students who are in need of support. Third, use of screening scores also informs schools about the needs of their particular student population which can assist with planning and resource mapping by finding groups of students with common needs. Finally, universal screening is recommended as an evidence-based practice by a number of different influential groups associated with educational policy and practice (e.g., President’s Commission on Special Education 2002; No Child Left Behind Act 2001; U.S. Department of Health and Human Services, 2001).

Unfortunately, there are a number of reasons why universal screening for behavioral risk has not become a more common practice yet. The following list represents concerns that often are expressed (Levitt et al. 2007):

  • Behavior is viewed as purposeful rather than as associated with environmental arrangements.

  • Historically, schools tend to be reactive rather than proactive with respect to behavior.

  • There is a widespread impression kids will “grow out of it” regarding problem behavior displayed during the early years of child development.

  • Concerns about profiling or stigmatizing children and youth who meet risk criteria.

  • Fear of costs and potential for identifying large numbers of students with EBD.

  • General perception that it is easier to screen for vision and hearing concerns because the response falls in the realm of families.

  • Political realities of managing parent reactions to behavior screenings and addressing issues of confidentiality.

  • Lack of needed skill set. Educators often are not trained to respond to behavior with the same confidence that they are able to respond to academic concerns.

Within a tiered framework of support one important goal is to “catch” students before academic and/or behavioral challenges become severe. Universal screening provides an opportunity for all children to be considered for risk against identified criteria. This approach shifts focus from a traditional “wait to fail” service delivery model toward proactively seeking out children who may be at risk of academic failure and/or behavioral difficulties that would potentially benefit from specific instruction or intervention (Glover and Albers 2007). Use of a universal screening process also has the potential to minimize impact of risk and/or impede further development of more severe problems by identifying students before problems become severe.

The process for determining social behavioral risk within school settings is perhaps less firmly established than is the process for uncovering academic risk. Contemporary thinking in academic screening suggests that multiple concurrent methods are redundant and do not have a value-added effect. Instead, academic screening emphasizes gated or filtered screening with intervention trials in between. Yet related to behavioral concerns schools have long relied on teacher nominations (i.e., referral for special education) as the most common approach for identifying problems (Kamphaus et al. 2010). Increasingly, educators are also making use of commonly collected student behavioral data (i.e., ODR) (McIntosch et al. 2009; Walker et al. 2005). Systematic screening is becoming more prevalent but confronts educators with challenges of instrument selection, interpretation of results, capacity to meet the needs of identified students, and potentially negative perceptions of the process by families and community stakeholders who may view the procedures as invasive and/or stigmatizing (Golver and Albers 2007). The President’s Commission on Special Education (United States Department of Education Office of Special Education and Rehabilitative Services 2002), the No Child Left Behind Act of 2001 (NCLB; United States Department of Education 2001), the US Public Health Service (2000), and the National Research Council (NRC; Donovan and Cross 2002) each indicate the need for early identification and intervention with recommendations to adopt universal, early, behavioral screening, yet limited information is available for schools regarding use of techniques and results (Albers et al. 2007)

The following section provides a brief description and sample items from several different screening questionnaires: (a) the Strengths and Difficulties Questionnaire (SDQ; Goodman 1997), (b) the Behavioral and Emotional Screening System (BASC-2 BESS; Kamphaus and Reynolds 2007), and (c) the Systematic Screening for Behavior Disorders (SSBD; Walker and Severson 1992).

The Strengths and Difficulties Questionnaire

The SDQ is a brief behavioral screening questionnaire for children and youth aged 3–16 years old. All versions of the SDQ ask about 25 attributes, some stated positively and others negatively. These 25 items are divided between five scales: (a) emotional symptoms (five items), (b) conduct problems (five items), (c) hyperactivity/inattention (five items), (d) peer relationship problems (five items), and (e) prosocial behavior (five items). Scales (a) through (d) are added together to generate a total difficulties score (based on 20 items).

The same 25 items are included in questionnaires for completion by the parents or teachers of 4–16-year-old children (Goodman 1997). A modified informant-rated version is available for the parents or preschool teachers of 3- and 4-year-old children. In addition, questionnaires for self-completion by adolescents also are available and ask about the same 25 traits, with slightly different wording (Goodman et al. 1998). The self-report format is appropriate for youth aged 11–16. In general population samples, it is recommended to use a three-subscale division of the SDQ into internalizing problems, externalizing problems, and the prosocial scale (Goodman et al. 2010). The SDQ can be administered by hand and scored by hand or by entering scores on-line. Paper copies of the instrument can be downloaded and photocopies made with no charge. On-line administration and scoring for the SDQ also are available.

The Behavioral and Emotional Screening System (BASC-2 BESS)

The BASC-2 BESS offers a systematic way to determine the level of individual student risk drawn from ratings of behavioral and emotional strengths and weaknesses for considered students. The instrument was designed for children and adolescents in preschool through high school. The process consists of brief forms that can be completed by teachers, parents, or students individually or in any combination. Each rating form ranges from 25 to 30 items, requires no formal training for the raters, and is easy to complete, taking only 5–10 min of administration time per student. The screener assesses behaviors that represent both problems and strengths, including internalizing problems, externalizing problems, school problems, and adaptive skills. It yields one total score with associated risk classification (i.e., normal, elevated, extremely elevated) that is a reliable and accurate predictor of behavioral and emotional problems.

The Systematic Screening for Behavior Disorders

The Systematic Screening for Behavior Disorders (SSBD; Walker and Severson 1992) incorporates three gates, or stages. The screening takes into consideration both teacher judgments and direct observations in order to identify students at risk for developing ongoing internalizing and externalizing behavior concerns. Stage 1 of the SSBD involves teacher nomination. Stage 2 requires that teachers complete a critical events inventory and a short adaptive and maladaptive behavior checklist for each of the nominated students. Students whose scores on these checklists exceed the established cutoff are then candidates for stage 3. This final stage involves a 15-min interval observation in both the classroom and on the playground to determine a student’s actual performance in social and classroom interactions.

Intervention Matched with Student Need

Once students are identified as nonresponsive to tier 1 instruction, the tier 2/3 team continues the SW-PBS logic by matching students with an appropriate level or tier of support. This decision is based on intensity, chronicity, and nature of the problem, as specific intervention strategies are selected to reflect individual student need (e.g., social skill deficit, self-management issue, emotional concerns). For example, a team may find a student has four ODRs and scored in the borderline range of total difficulties on the SDQ Goodman 1997), thus indicating the need for a tier 2 intervention. Or, a team may choose to forego tier 2 in favor of tier-3-level intervention for a student with more significant and intense problems (e.g., more than six ODRs, abnormal score on the SDQ). Beyond matching the intensity level (i.e., tier 2 or tier 3) of support based on student data, it is beneficial to consider particular attributes of the problem (e.g., function, location, setting, time of day, behavioral topography, acquisition deficit, performance deficit) prior to selecting an intervention (Hansen et al. 2014).

Accordingly, the tier 2/3 team should gather relevant information in a timely manner so that features of the problem are accurately identified, while still allowing for rapid access to interventions that are readily available. For a majority of identified students, screening results along with existing school data can be used to facilitate decisions that accurately inform intervention selection. Data that are easily accessible and generally useful for pinpointing aspects of social, emotional, or behavioral challenges may include documented disciplinary events (e.g., ODRs), student attendance patterns, grade point average and/or course grades, academic performance scores, and/or frequency of visits to the school nurse or counselor. These data can indicate when, where, and under what condition problem behavior is most likely to occur. For example, data may indicate a student is (a) reading below grade level, (b) often complains of sickness during reading class and thus frequently asks to see the nurse, and (c) when referred to the office for behavioral infractions the teacher consistently indicates “task avoidance” as a possible motivation for the problem behavior. Collectively, these data indicate the student may need both targeted small-group reading instruction, as well as behavioral supports to keep the student engaged during instruction. Similarly, universal screening tools originally used to identify students for tier 2/3 supports may provide relevant information for selecting an intervention. For instance, the SSBD may indicate a student has critical internalizing issues such as extreme anxiety. This information should be coupled with other data to determine the root of the anxiety, and in turn, select an intervention that confidently reflects the student’s needs. When tier 2/3 teams triangulate data rather than relying solely on one tool or one statistic, they can more precisely capture the student’s abilities and deficits, and thus match intervention accurately (Bruhn, Lane, & Hirsch, 2014; Marchant et al. 2009)

Beyond data collected as part of regular school practices, brief assessments such as the Functional Assessment Checklist for Teachers & Staff (FACTS; March et al. 2000) and the Functional Assessment Screening Tool (FAST; Iwata and DeLeon 1996) can be used to identify possible functions of students’ problem behavior (e.g., access or avoidance of attention or tasks, sensory stimulation), which in turn, can inform the appropriate intervention selection. Although identifying function is generally regarded as part of the tier 3 process, quick tools such as the FACTS and FAST may help guide “function-based” thinking at tier 2. For example, if a student is misbehaving to gain peer attention, the student might be placed in a self-management intervention whereby if the student meets his/her goal of remaining on task they can access free time spent with peers (Briere and Simonsen 2011; Bruhn et al. 2014). Similarly, much of the research on the popular tier 2 intervention, check-in/check-out (CICO), has suggested students who are attention motivated are more likely to respond positively to CICO than students with escape-motivated behavior (Hawken et al. 2011; McIntosh et al. 2009).

For the few students (approximately 5 % of the student population) who display intense and chronic problem behavior, a comprehensive functional behavioral assessment (FBA) should be conducted to create an equally individualized and intensive behavior support plan (BSP). Generally, students needing an FBA have long, complex histories of behavior problems and have been exposed to multiple risk factors (e.g., transiency, academic failure, substance abuse, poor parenting). Ideally, the team conducts interviews with the student’s teachers, parents, and the student him/herself; reviews archival student records; and directly observes the student to identify antecedent conditions occasioning the target behavior and the consequences maintaining that behavior (Cooper et al. 2007). Data gleaned from the FBA are then used to design a highly individualized BSP. Tier 3 interventions such as function-based BSP’s tend to require more time, effort, and resources than tier 2 interventions because they are acutely tailored to the individual rather than addressing a small group of students with comparable problems through more generalized supports as in tier 2. Function-based interventions have a record of success that began in clinical settings with students who had low-incidence disabilities, but more recently, positive outcomes have been documented across a range of school-based settings for students with persistent behavioral problems (McIntosh et al. 2008; Lane et al. 2009).

Regardless of the interventions the team has to choose from, teams must have adequate time to consider and plan for all students identified as needing support beyond tier 1. Use of a specific format for collecting, reviewing, and discussing relevant student information can keep conversations directed toward existing school-based supports and/or accessible community agencies and programs while avoiding discussion of factors that are beyond influence of the support team (e.g., homelife, community circumstances, previous experiences with related families). Maintaining a problem-solving approach that is focused on alterable indicators of risk through the use of research-based interventions will maximize time allotted for the behavior support team to address all students identified for additional support.

Research-Based Interventions

Following student identification and data review to determine appropriate intervention and level of environmental supports, intervention implementation, and ongoing supports should be provided. The tier 2/3 team should first determine their capacity to provide a range of interventions and supports, start with those currently in place, and expand to create a full range within each tier of the continuum. Key factors related to intervention implementation within a continuum of supports include: (a) clear alignment with core behavioral instruction and procedures (e.g., addresses school-wide behavioral expectations, uses same or similar language for recognizing or correcting behavior, follows school-wide reinforcement system, but provides a greater frequency of support), (b) a designated coordinator for each tier 2 strategy and designated case managers for students receiving tier 3 supports, (c) ongoing data review to assess implementation fidelity and student progress, and (d) a plan for fading or intensifying supports. Within tier 2, the current research base advocates (a) additional small-group social skill instruction addressing skill and performance deficits (Elliott and Gresham 2008), (b) empirically validated self-management strategies that have demonstrated improvements in social and academic behaviors such as CICO (Crone et al. 2010), Check and Connect (Christenson et al. 2012), or Check, Connect, and Expect (Cheney et al. 2009), and (c) academic supports either through differentiated and supplemental instruction ideally within an MTSS with progress monitoring (Epstein et al. 2008).

Research studies reporting positive impact of tier 2 interventions are evident. For example, outcomes from use of social skill instructional groups have shown decreases in disruptive behavior, increases in on-task behavior, improved scores from teacher ratings of social behavior and academic competence, along with increases in prosocial play skills, peer interactions, and communication (Kamps et al. 2011; Gresham et al. 2006; Marchant et al. 2007). Evidence for the group-oriented self-management program show similar results. For example, investigations of the CICO intervention consistently demonstrate decreases in ODR for student participants (e.g., Hawken et al. 2007; Todd et al. 2008), increases in academic engagement (Hawken and Horner 2003; Campbell and Anderson 2011), and reduced frequency of disruptions or negative social interactions (e.g., Campbell and Anderson 2008; McIntosh et al. 2009). Finally, a recent study that documented use of an academic support intervention that showed improvements in “assignment attack” behaviors such as task persistence and organization (Ness et al. 2011).

Tier 3 supports are guided by an FBA whereby an individual support plan is developed to teach a functionally equivalent replacement behavior and environmental modifications are made to reinforce replacement skill use (Gage et al. 2012). Additional mental health, family, and other noneducational supports are also included in tier 3 supports when indicated (e.g., RENEW; Malloy et al. 2010).

In addition to building small-group strategies within tier 2 supports, the tier 2/3 team can also include simple alterations within classroom and other school environments as an additional tier 2 option. It may be the case that a small-group intervention is not necessary to address the presenting concerns, rather intensifying universal practices in response to the problem may be sufficient to produce behavior change. For example, increasing universal strategies on targeted behaviors such as (a) providing more frequent prompts for expected behavior, (b) increasing the rate of specific performance feedback for correct behavior, (c) reteaching classroom rules and routines, or (d) altering the environment to increase supervision or the add effective instructional practices may be sufficient to meet the student’s need.

Monitoring Student Progress

In the same manner that curriculum-based measures are one type of progress monitoring tool to assess academic performance over time, behavioral assessment data within SW-PBS can also be collected and used as the basis for determining students’ RTI. In academic progress monitoring, there are explicit decision rules to guide service delivery and determine student responsiveness (VanDerHeyden et al. 2007). Behavioral progress monitoring, on the other hand, lacks an equivalent standard protocol. Rather, instructional programming decisions may be made based on data collected via systematic direct observation, direct behavior ratings (DBR; Chafouleas 2011), and review of archival data (e.g., documented disciplinary events, time out of instruction due to problem behavior, academic work samples). These are common techniques for monitoring student progress before, during, and after intervention.

Direct observation is certainly the most time and labor intensive progress monitoring option, as it requires an individual to watch a student’s behavior for a set period of time and record either the numerical (e.g., frequency) or temporal dimension (e.g., duration) of the behavior. This process must be done frequently and data should be graphed for analysis. Additionally, to establish reliability of direct observation data, another observer may simultaneously, but independently, collect data using the same preestablished procedures. Then, the data may be compared for interobserver agreement (IOA). Although direct observation of both problem and appropriate behavior would provide schools with the best data source to inform instructional decisions, the challenges of (a) observing behavior expected across multiple educational settings, (b) the costs in terms of time to complete, and (c) the expertise required to collect reliable data often limit direct observation as a viable option for progress monitoring. This may be true especially for teachers who are asked to simultaneously teach and collect data, as many teachers view this as an impossible task (Epstein 2010). Instead, teacher perceptions as measured through informal and formal behavioral rating scales, and DBRs are more common approaches to monitoring behavioral interventions. A DBR is similar to direct observation in that it requires the teacher to directly observe the student. However, observation is not continuous and is accumulated over a set period of time (e.g., first period, circle time, 45 min). The DBR can be used in the pre-populated form, which asks teachers to rate students on a scale of 0–10 on three behaviors—academic engagement, disruptive, and respectful. Or, the teacher can tailor the DBR to reflect the behavior(s) of interest. Researchers have suggested the pre-populated DBR is moderately to highly correlated with frequency and duration data obtained via direct observation (Chafouleas et al. 2009; Riley-Tillman et al. 2008). Thus, DBR may be a more practical option than direct observation for teachers.

Data-Based Decisions for Movement Between Tiers

Whether the team selects direct observation or a DBR, regular review of student data is key to making decisions about movement within the tiered model. For example, in some cases, review of student data will demonstrate a student has consistently met a specific behavioral performance target over a period of several weeks. In this event, the Tier 2 team would have sufficient evidence to advocate for a gradual reduction in intervention supports that leads the student to a self-management and maintenance phase (e.g., Campbell and Anderson 2011). Or, regular review of student data might reveal highly variable performance or performance that is consistently below the expected level, representing a questionable RTI. Under these circumstances, the team will first verify fidelity of intervention implementation and then make decisions that could include simple modifications to the existing support or recommendation for a higher intensity intervention (e.g., Campbell and Anderson 2008; Fairbanks et al. 2007). Finally, in some cases, review of student performance data will indicate an overall poor response to the intervention that is demonstrated by an increase in frequency or intensity of problems and/or failure to improved at the expected rate or to the desired level of performance after sufficient exposure to an intervention matched with need and implemented with integrity. This type of data can serve as the basis for recommending additional information gathering (i.e., FBA) and development of an individualized support plan in tier 3 (Crone and Horner 2003; March and Horner 2002).

Monitoring Fidelity

With respect to fidelity of treatment, the tier 2/3 team has two foci. First, the tier 2/3 team must routinely consider how well or to what extent interventions are being accurately implemented. Second, the team should conduct, at least annually, a review of their SW-PBS tier 2/3 system that includes evaluation of overall effects and impact toward improvement of social and academic outcomes.

Tier 2/3 Intervention Fidelity

Only when an intervention is implemented as designed can conclusions about behavior changes be made accurately (Gansle and Noell 2007; Yeaton and Sechrest 1981). That is, school teams that measure implementation quality and accuracy can have greater confidence in decisions made when reviewing student data. For example, if a student demonstrates limited or poor response during an intervention but the school team has measured and determined fidelity of implementation to be high, then it is likely the student truly is in need of adapted, alternate, or more intensive support. Conversely, if a student did not have the opportunity to benefit from intervention because it was not implemented with fidelity, then the team may need to provide additional training and resources to the interventionist and allow time for the intervention to be implemented with fidelity prior to placing the student in a more intense level of support (Bruhn et al. 2014; Mitchell et al. 2011). Additionally, teams must realize there are multiple factors such as intervention complexity, training adequacy, and skill of the interventionist that affect intervention integrity (Yeaton and Sechrest 1981). Regardless of the reason for low treatment integrity, performance feedback is critical for improvement and accurate implementation (Duhon et al. 2009; Keller-Margulis 2012). Performance feedback should involve the team reviewing fidelity data and participating in a constructive dialogue about setting goals for improving implementation and specific steps to take toward meeting those goals (Keller-Margulis 2012). Determining the extent to which all parts of an intervention are implemented accurately is a key priority for the tier 2/3 team. SW-PBS tier 2/3 teams are challenged with developing a process that (a) provides adequate evidence the intervention was implemented as intended, and (b) can feasibly be conducted on a regular basis. Three common methods for measuring accuracy of intervention implementation include: (a) intervention-specific product review (e.g., monitoring forms, rating scales, daily progress report (DPR), self-management records), (b) direct observation of intervention implementation, and (c) self-report measures completed by intervention providers. Each method includes both benefits and potential limitations.

Intervention Product Review

One way to verify delivery of intervention is by completing a review of products associated with the intervention (Crone et al. 2010). For example, many tier 2 behavioral interventions include daily or weekly documentation of student performance such as a DPR system, as is used in CICO. In these cases, the behavior support team can examine three to five of the most recent progress documents to determine whether some of the specific elements of the intervention occurred (e.g., progress was recorded, progress was calculated and evaluated using specified criteria, data were shared with relevant stakeholders such as classroom teacher and family as evidenced by stakeholder signatures). When review of student products provides evidence that specified components of the intervention are in place and being delivered consistently, the team can have greater confidence in the accuracy of progress monitoring data and subsequent decisions made from use of those data. Alternately, when review of intervention-related products identifies an area of low implementation, a member of the support team can be designated to provide coaching, remediation of skills, and feedback about intervention delivery as needed (e.g., student, teacher, and/or parent). However, one limitation of the product review is that it does not take into account components of the intervention that are not part of the product, or in this case, the DPR. For instance, when students and interventionists use the DPR, it is expected that students will receive verbal feedback from an adult about their performance as well as a predetermined reward for meeting goals. These components, which may be critical to the success of the intervention, may need to be documented via direct observation or self-report (Bruhn, McDaniel, & Kreigh, in press.)

Direct Observation

A second method for verifying accuracy of intervention integrity is direct observation. In these cases an observation checklist may be especially useful both for documenting features that occurred and for providing feedback to implementers. Direct observation tools can be developed to reflect the essential features of intervention delivery. For example, a tier 2 social skills intervention observation checklist might include: (a) instructor introduced, defined, and discussed skill use and its importance, (b) instructor demonstrated at least two examples and nonexamples of the skill, (c) students correctly demonstrated skill use in prompted role plays, (d) students correctly demonstrated skill use in nonprompted role plays, and (e) instructor reinforced occurrences of the appropriate skill and corrected incorrect skill use (Elliott and Gresham 2008). As a second example, within a tier 3 FBA-BSP intervention, direct observation targets might include: (a) teacher prompted student regarding correct skill use, (b) teacher consistently reinforced desired behavior, and (c) teacher minimized reinforcement of problem behavior. Like direct observation used to monitor student progress, direct observation of fidelity is also time and labor intensive, but is the most accurate method for estimating intervention integrity.

Self-report Measures

A final option for collecting fidelity of implementation data is asking intervention providers to record components they provide and/or self-assess their level of accuracy according to a list of described features. Self-report measures can be organized to collect implementer ratings (e.g., five-point scale) or as a simple “yes” or “no” checklist. Interviews and questionnaires may be used to give informants an opportunity to elaborate beyond the scope of what is covered in a checklist or rating scale. Although self-report measures are often used to assess implementation because they are easy and require little time to complete, teams using this technique should be cautious when reviewing and evaluating data. Some research on self-reporting has indicated implementers tend to overrate their performance and accuracy (Wickstrom et al. 1998). However, more recently, some researchers have demonstrated self-report of fidelity can be an accurate and more efficient alternative to direct observation (Hagermoser Sanetti and Kratochwill 2009). Regardless, a multi-method, multi-informant approach to measuring fidelity that includes both direct and indirect observation can only increase the accuracy of conclusions drawn about intervention effects (Lane et al. 2004).

Tier 2/3 Systems Fidelity

A hallmark of the SW-PBS process is coordinated universal, tier 2, and tier 3 systems of support (Lewis et al. 2010). Similar to the work of VanDerHeyden et al. (2007) who evaluated fidelity of an entire RTI system including not only intervention fidelity but also fidelity to the decision-making process; recent advancements in SW-PBS related to effective tier 2/3 intervention efforts led to the development of instruments designed to assess the fidelity of the overall systems, or process, of implementation. These tools are used both as metrics of fidelity as well as data sources for making data-based decisions to continually build and refine tier 2/3 systems. One example, the Benchmarks for Advanced Tiers (BAT; Anderson et al. 2010) was created to answer three main questions: (1) Are the foundational (organizational) elements in place for implementing tier 2 and 3 behavior support practices? (2) Are the essential features of a tier 2 support system in place? and (3) Are the essential features of a tier 3 system in place? The BAT is completed by the tier 2/3 team and reflects the consensus or majority of team member perceptions. A second example, the Individual Student Systems Evaluation Tool (ISSET; Anderson et al. 2012) is also used to measure the implementation status of tier 2/3 systems within a school (e.g., Debnam et al. 2013). The ISSET consists of 35 items and is divided into three parts: foundations, tier 2 interventions, and tier 3 interventions. A summary score is obtained for each of the three parts and is administered/completed by an external evaluator. Two data sources are used to score the ISSET: Administer and teacher interviews and a review of permanent products/documented procedures. Results from either instrument are used to identify areas of strength and needed improvements.

Evaluating Impact

Tier 2/3 teams, in concert with the universal SW-PBS, adhere to the basic logic of SW-PBS (i.e., data–practices–systems) and also conduct informal evaluations of the effects of their efforts. Teams focus on three questions: (a) Did the interventions we put in place lead to improved student outcomes? (b) If interventions were not effective, what else do we need to know to increase the likelihood of success (e.g., what training and technical assistance do we need)? (c) Could we have worked more efficiently (see Fig. 3 for a more comprehensive list of system evaluation points). To answer these questions, the tier 2/3 and universal SW-PBS teams determine what change has occurred across the variables or behaviors of interest. In the case of behavioral intervention, programs were likely selected with the expectation of impacting problem behavior and student engagement, which in turn may lead to improvements in academic achievement. Evaluating features of tier 2 and 3 program impact can occur using a variety of methods. For example, in some cases compilation of existing school data provides valuable evidence of program impact and can be used to evaluate system outcomes.

Fig. 3
figure 3

Sample evaluation questions to conduct tier 2/3 system review

Under other circumstances, a more formalized assessment of behavior support team member perceptions or measures of system-wide implementation may be warranted when student outcomes are inconsistent (e.g., Self-assessment Survey, SAS; School Safety Survey, SSS; Team Implementation Checklist, TIC). School staff members typically complete the SAS survey at minimum annually. After reviewing outcomes from tier 2 or 3 interventions that demonstrated questionable positive impact for multiple students, the team could review SAS survey results to get a better understanding about staff perceptions of implementation at the school-wide, nonclassroom, classroom, and individual student levels (Safran 2006). SAS results also demonstrate staff perceptions of priorities for improvement. Regular review of these data lends itself to ensuring the SW-PBS efforts across all tiers address the needs identified by staff that completed the survey.

In addition to the outcomes derived from existing school data, teams also gather social validation data. Social validity data typically provides a picture of the extent to which particular stakeholder groups (i.e., students, families, and teachers) value identified practices and outcomes. Social validity data are commonly gathered through use of a survey or asking personnel to respond to items on a brief questionnaire. Example statements or questions may include: Overall problem behaviors have decreased for this student during participation in the behavioral intervention program; I think this behavioral intervention program may be good for other students in our school; Having my child/student in the behavioral intervention program is worth my time and effort. If social validity results are low it may be difficult to continue implementation of the support “as is.” Instead, teams will investigate why the practice is perceived poorly and make adjustments either by providing additional information and technical assistance and/or by making changes to features that perhaps were not feasible to maintain (Fixsen et al. 2009; Guskey 2002).

Finally, timelines for conducting the system evaluation must be considered also. An annual review that occurs near or after the end of each school year may be practical and make sense for many school teams. This time frame allows widespread participation in the system across many staff members, students, and parents throughout the school year, concludes during a period when student data are already commonly collected, and guides decisions the team will make for refining the system the following year. Annual evaluation allows time before the start of the next school year for making adjustments to the existing system such as improving the communication system to all staff regarding implementation if student outcomes are inconsistent across the school day or additional training for intervention implementers if fidelity is low. Systems evaluation requires thoughtful but realistic consideration. The process should use existing data collection strategies (e.g., student outcomes) and fidelity/planning tools (e.g., the BAT) to insure teams engage in the process. At the same time, evaluations also should not be so simplistic that valuable outcomes are overlooked or never uncovered.

Empirical Support for Tier 2/3 SW-PBS

To date, there is empirical evidence on the social, emotional, behavioral, and academic impact of tier 1 SW-PBS supports (Barrett et al. 2008; Bradshaw et al. 2010; Bradshaw et al. 2008; Curtis et al. 2010; Horner et al. 2009; Simonsen et al. 2010). Likewise, there is a strong evidence base demonstrating positive effects from use of particular interventions that are commonly implemented within tier 2 and 3 levels of support. For example, social skill instruction, self-management strategies, academic supports, and use of FBA data to guide individual BSPs have been verified as effective treatments for changing behavior (e.g., Bessette and Wills 2007; Christensen et al. 2007; Skinner et al. 2009; Sumi et al. 2013). However, empirical demonstrations of the impact of tier 2 and 3 supports within the context of a complete continuum of SW-PBS to date are limited. The following provides a brief summary of what has been concluded from the existing body of work.

Two recent reviews of tier 2 implementation studies examined questions associated with identification of students, delivery of supports, and overall impact (Bruhn et al. 2014; Mitchell et al. 2011). From the two reviews several themes emerged. First, on balance, the overall number of studies investigating tier 2 level of implementation has increased recently. Ranging from approximately 2002 through 2012 researchers conducted and published nearly 30 separate investigations concerning tier 2 interventions and delivery within a tiered framework (Bruhn et al. 2014). Second, demonstrations reporting sufficient fidelity of tier 1 implementation prior to initiation of tier 2 intervention, particularly at the classroom level, were fewer than expected (Bruhn et al. 2014). Third, although there is evidence of reliance on a few commonly collected data such as disciplinary events, screening scores, and teacher nominations as primary sources for determining which students will access intervention, wide variation in identification approaches were still clearly evident (Bruhn et al. 2014; Mitchell et al. 2011). Fourth, across the existing body of work, evidence generally indicated high social validity ratings for the variety of interventions that have been included. The remainder of this section provides further detail for some of these findings.

Core Instruction Delivered With Fidelity

Within the SW-PBS literature, several instruments have been developed and used for monitoring implementation efforts. Examples include the TIC (Sugai et al. 2009), the SSS (Sprague et al. 2002), and the Benchmarks of Quality (BoQ; Cohen et la. 2007). Two of the most commonly used tools are the School-Wide Evaluation Tool (SET; Horner et al. 2004) and the Effective Behavior Support/Self Assessment Survey (EBS/SAS; Sugai et al. 2000). The SET is designed to assess features of tier 1 that are in place and evaluate ongoing efforts for developing a school-wide approach for behavior management and discipline.

The SET consists of interviews, observation of the setting, and a review of products such as teaching plans, signage, and data collection. An overall score is calculated and results of 80 % or higher indicate adequate implementation. In some of the initial publications for tier 2/3 interventions, it was common to simply state that features of tier 1, core instruction such as clearly defined expectations and rules, explicit behavioral instruction across settings, use of a recognition system, and consistent documentation of behavioral infractions were in place (e.g., Lane et al. 2002; Lane et al. 2003; Gresham et al. 2006). However, more recent investigations reflect the importance of high-quality core instruction as part of the continuum of supports by attending to and reporting scores from the SET (e.g., Campbell and Anderson 2011; Hawken et al. 2011; Mong et al. 2011; Simonsen et al. 2011). While the SET is a research-validated instrument with strong psychometric properties demonstrating consistency in measurement of tier 1 features, one aspect of tier 1 that is not addressed within the SET is how well the core behavioral instruction is provided within individual classrooms (Horner et al. 2004). Likewise, acknowledgment of students demonstrating expected behaviors and matching instruction with student ability are associated with sustained implementation and positive student outcomes (e.g., ODR) (Matthews et al. 2014). However, the existing literature related to tier 2 has not explicitly reported that scores from the SAS showed integrity of classroom-level implementation prior to delivery of behavioral supports for identified students (Bruhn et al. 2014; Mitchell et al. 2011).

Identification Practices and Tools

The existing evidence also indicates some commonalities among tools that are used to identify students for tier 2 behavioral interventions, but still shows wide variation in decision criteria. For example, Mitchell et al. (2011) reported participation in a tier 2 intervention was based on one or a combination of the following: (a) a nomination process in which a classroom teacher, a parent, or a problem-solving team identified the student as at-risk; (b) use of existing behavioral performance data—typically, ODR information—to indicate that the student was unresponsive to the tier 1 prevention efforts or continuing to demonstrate difficulties meeting social behavioral expectations; or (c) use of a behavioral screening score. Similarly, Bruhn et al. (2014) determined (a) ODR; (b) teacher nomination; (c) academic performance; or (d) other methods such as parent nomination, attendance data, or a specified behavioral function served as tools for identifying student candidates for intervention. Yet, specific details regarding exact criteria differed across studies. Descriptions of several examples follow.

Nomination

Several studies asked teachers to identify, without a quantifiable index, which students needed intervention (Bruhn et al. 2014). In one study, participants were identified for intervention based on teacher nominations because of behavioral difficulties in the classroom and/or existence of a behavior plan (McCurdy 2007). A different example indicated teachers nominated students because of classroom problem behavior and lack of responsiveness to the tier 1 prevention efforts (McIntosh 2009). In a third example, a multi-informant process was used, which included administrator nomination, teacher verification of problem behavior, parental consent, and student willingness to participate, to determine which children would receive intervention (Todd et.al. 2008).

Student Data

The existing research base also provides examples for using specified cut points of student academic and/or behavioral data that are commonly collected in school settings to identify intervention candidates. For example, students who met a predetermined threshold of risk such as five or more disciplinary incidents or two or more documented events of problem behavior were selected for intervention (Hawken 2006; Hawken et al. 2007). Other studies have also included use of academic data such as grade point averages (e.g., Robertson and Lane 2007) or Dynamic Indicators of Basic Early Literacy Skills (DIBELS) scores (e.g., Wills et al. 2010) along with the collected behavioral data (Bruhn et al. 2014).

Behavioral Screening

In the most recent comprehensive review of tier 2 within SW-PBS, use of one or more screening instruments to identify students at risk was evident in less than half of all the reviewed studies (i.e., 13 out of 28, Bruhn et al. 2014). When a systematic screening process was used, the most commonly employed instruments were the Student Risk Screening Scale (SRSS; Drummond 1994), the SSBD (Walker and Severson 1992), and the Social Skills Rating System (SSRS; Gresham and Elliott 1990). Several examples combined use of the SSBD with the SRSS (e.g., Lane et al. 2008; Lane et al. 2010; Little et al. 2010) which is interesting considering the SSBD has been shown to identify both externalizing and internalizing concerns and at this time the SRSS is best known for detecting externalizing attributes alone. One study demonstrated use of a behavioral screener, the SSBD, but also included DIBELS as part of the identification method (Wills et al. 2010). Demonstration of screening as a method for tier 2 identification appears to be increasing within recent years. Among the 13 studies from the Bruhn et al. (2014) review, all studies that included a behavioral screening process were published within the past 10 years (i.e., 2003–2011) and a majority occurred in studies published in 2007 or later. While use of screening to identify students at risk in academic domains is well established, teams working within SW-PBS have continued to rely on more of a multi-method approach that often includes use of a psychometrically sound rating instrument, but may review other data such as informal teacher or school personnel perceptions of problems and/or recorded disciplinary events (Bruhn et al. 2014)

Social Validity

A final topic that was evident within both of the recently published tier 2 literature reviews was the issue of social validity. In a 1976 address presented for the Division of Experimental Analysis of Behavior within the American Psychological Association, Montrose Wolf described, with humor, how Don Baer helped prioritize “social importance” as it relates to the mission of applied behavior analysis and how subsequently ABA came to “have a heart” (Wolf 1978). Not quite 40 years later, the measurement of behavioral change that is perceived as socially valued or important is more common than not at least in the area of SW-PBS research. Currently the majority of studies conducted with tier 2 level interventions have reported social validity data (Bruhn et al. 2014; Mitchell et al. 2011). Most of the studies that included social validity measurements at minimum gathered these data from participating teachers who perhaps most closely detected the impact of school-based interventions (e.g., Fairbanks et al. 2007; Hawken et al. 2007). Several studies also demonstrated measurement of student beliefs about interventions that were delivered (e.g., Lane et al. 2002; Lane et al. 2003). Commonly used tools included the Intervention Rating Profile (IRP; Witt and Elliott 1985), the Children’s Intervention Rating Profile (CIRP; Witt and Elliott 1985), and the CICO Program Acceptability Questionnaire (Crone et al. 2010). Results from these assessments across a variety of stakeholders showed generally favorable outcomes for indicators like ease of implementation, perceived benefit, and potential sustainability of practices over time (Bruhn et al. 2014).

Recommendations for Future Research

There is agreement about the types of practices to include in a full continuum of behavioral supports and there is understanding of the essential system features that will facilitate high-quality implementation. What is needed is a more expansive body of empirical evidence indicating positive outcomes (e.g., student academic and behavioral indicators, organizational health, school safety and climate) are attained and can be sustained over time when interconnected tiered support is delivered. In addition, research is needed on several factors within the tier 2/3 system including (a) impact on students with internalizing problem behavior (e.g., depression, anxiety), (b) the ability of teams to effectively and efficiently match intervention to presenting problem with existing commonly collected school data, (c) the impact of academic and behavioral supports guided by student need, (d) essential implementation steps that lead to sustained outcomes, and (e) the mediating and/or moderating effect of related environmental variables such as school district/state support, community and cultural factors, and prior student learning history. This will require sophisticated investigations demonstrating a clear linkage between tier 1 systems, including classroom, implemented with high-fidelity and student-need-directed tier 2/3 supports. While the emerging literature on impact of tier 2/3 supports within the context of SW-PBS is encouraging, the true “value add” has yet to be empirically demonstrated. The logic of nesting tier 2/3 supports within a connected and related instructional environment is built on decades of research on promoting maintenance and generalization of intervention outcomes and to identify and support students at the first sign of risk. Two recent studies provide a starting point to empirically validate the impact of both the prevention logic of SW-PBS and the impact of a complete continuum of supports.

First, Reddy and colleagues completed a meta-analytic review of the types of prevention and intervention programs available for students with or at risk for emotional and behavioral disorders (Reddy et al. 2009). Findings indicated that prevention and early intervention programs can be effective for reducing risk of onset and for minimizing impact of already existing symptoms. Second, Nelson et al. (2009) assessed child outcomes when a full continuum of behavioral supports was provided. The tiered system of supports consisted of Behavior and Academic Support and Enhancement (BASE) at the universal level (Nelson et al. 2002), First Step to Success as the tier 2 level intervention (Walker et al.1997), and multi-systemic therapy (MST) as the tier 3 treatment (Henggeler et al. 1998) and were tested with 407 children in grades K–3. Effects on problem behavior, social skills, and academic competence across levels of intervention (tier 1, tier 2, and tier 3) were measured. One finding of particular importance was that children who participated in the tier 2 or tier 3 interventions early in their school career showed immediate social behavioral improvements that could be sustained over time with continued implementation of the tier 1 supports (Nelson et al. 2009). However, the results did not demonstrate a similar outcome for teacher ratings of academic competence among students who received the advanced tier interventions. The results offered initial evidence to support the idea that a tiered service delivery model could impact behavioral performance of children with or at risk for behavioral problems (Nelson et al. 2009).

Implications for Practice

Across this chapter, essential features and examples to build SW-PBS systems at the tier 2/3 level of support have been provided. The limitations of a single chapter and the complexities of both specific intervention strategies and the intensity of support across school environments preclude simple recommendations for practice. The encouraging news as previously discussed, intervention and data collection strategies along with the necessary systems of support at the universal level are well established and comprehensive implementation guides are readily available for practitioners (see pbis.org). In addition, many districts have adopted integrated frameworks to tie in community supports at the individual student level. Unfortunately, schools and districts continue to struggle with establishing effective tier 2 systems of support. While the research base is limited with respect to systems, as noted above there is strong empirical evidence for specific intervention strategies appropriate for tier 2 level of supports (e.g., CICO, social skill instruction).