Introduction

Special education teachers and paraprofessionals can be trained to use differential reinforcement procedures as integral features of behavior intervention plans (Countenance et al. 2014; DiGennaro et al. 2007; Plavnick et al. 2010; Petscher and Bailey 2006). For example, DiGennaro et al. (2007) used performance feedback to train four change agents to implement behavior support plans that incorporated differential reinforcement of alternative behavior (DRA) for one participant and differential reinforcement for other behavior (DRO) for another participant. More broadly, skilled practitioners have used behavioral skills training (BST) and pyramidal training to teach support staff to correctly implement discrete-trial teaching (Sarokoff and Sturmey 2004), preference assessments (Pence et al. 2012), functional analyses (Pence et al. 2014), and other behavioral programming (e.g., Kuhn et al. 2003). Although BST and pyramidal training both improve staff performance, these training modalities may not be widely embraced in school settings due to the substantial cost of hiring a skilled practitioner in behavior analysis (Parsons et al. 2012). Specifically, public schools may not have funds to conduct optimal training.

As previously noted, the ability to implement differential reinforcement procedures should be a core skill set for change agents employed in school settings (Wong et al. 2015). When change agents implement DRA and DRO procedures with high procedural integrity, these procedures can increase appropriate behavior, decrease problem behavior, or both (Codding et al. 2005; St. Peter Pipkin et al. 2010; Vollmer et al. 1999). Nevertheless, treatment errors, particularly those made early in the treatment process, could lead to the reemergence of problem behavior and impede later treatment progress (e.g., St. Peter Pipkin et al. 2010). Thus, further research is needed to identify instructional strategies to effectively train change agents in school settings to implement common behavioral reduction interventions.

One training method, self-instruction packages (SIPs), does not require a skilled practitioner (Graff and Karsten 2012) and might offer practical advantages in settings such as public schools. By design, SIPs are antecedent-only strategies that can be used to train change agents without an expert trainer. Studies have shown that SIPs can teach change agents to correctly implement preference assessments (Deliperi et al. 2015; Delli Bovi et al. 2017; Graff and Karsten 2012; Hansard and Kazemi 2018; Lipschultz et al. 2015; Nottingham et al. 2017; Ramon et al. 2015; Weldy et al. 2014), most-to-least prompting (Giannakakos et al. 2016), discrete-trial instruction (Cardinal et al. 2017; Thiessen et al. 2009; Vladescu et al. 2012), motor imitation programs (Du et al. 2016), picture-exchange communication systems (Martocchio and Rosales 2017), and three-step prompting (Spiegel et al. 2016).

Specifically, Spiegel et al. (2016) evaluated the extent to which a SIP trained three caregivers of children with ASD to implement three-step prompting during tasks in which their children were typically noncompliant. The SIP, which could be completed in 12 min 36 s, improved caregivers’ implementation of three-step prompting with their children and promoted their application of the skills to novel tasks and settings. Notably, researchers remained within 3 m of participants when viewing the SIP, aligning with prior investigations in which researchers remain in close proximity (e.g., Delli Bovi et al. 2017).

Although the findings of the Spiegel et al. (2016) study are promising, additional research is needed to evaluate the extent to which SIPs can train change agents to implement behavioral interventions in settings such as schools. School consultation models may rely on group trainings provided by noncredentialed personnel. If SIPs can promote acquisition of change agents’ implementation of behavioral reduction strategies when delivered in a group format, practitioners in school settings could train more change agents to deliver empirically supported behavioral interventions with less time.

The current study aimed to evaluate the extent to which change agents learned to implement DRA and DRO procedures after viewing written instructions and PowerPoint™ presentations containing voice over instruction and video modeling in a group. This SIP method, when delivered in a group, has the potential to decrease training costs and increase the probability children in special education receive their prescribed behavioral services.

Methods

Participants

As part of a week-long in-service training on evidence-based practices for students with special needs, we recruited six special education staff members (five females and one male) to participate in this study. Participants, on average, were 35 years old (range 22–66 years old) and had varying levels of experience in special education (less than 1 year up to 10 years) as either a teacher or a paraprofessional. Table 1 displays demographic information for each participant. The first author obtained informed consent from participants at the start of the training. Sessions took place in a conference room at a public school’s administrative building. All participants reported some experience with behavioral intervention strategies but no formal training with DRA or DRO. Additionally, participants reported they had or were currently working with children in special education who engaged in problem behavior.

Table 1 Participant demographic and training duration information

Setting and Materials

Participants used the following materials across all phases of the study: a MotivAider® (to monitor the 30-s interval in the DRO procedure); flash cards depicting common special educational targets (e.g., letters, numbers, shapes, animals, and colors); edibles to use as reinforcers (e.g., skittles, crackers); and a token board with four tokens. The lead author developed and created the PowerPoint™ SIP in approximately 10 h. This time included filming the video clips, editing the video clips, recording the voiceover instruction, and creating the final presentation.

Experimental Design

We used a nonconcurrent multiple probe design across participants to evaluate the effects of the SIP training on staff members’ implementation of DRO and DRA procedures (e.g., Jenkins et al. 2015; LeGray et al. 2010). Participants watched the SIP in triads. We assessed skills for each participant individually (i.e., without the other participants present). Although we conducted sessions for all participants in close temporal proximity, we did not systematically alternate sessions across participants and tiers as required for a concurrent multiple baseline design (e.g., Carr 2005; Coon and Rapp 2018).

Training Simulated Students and Scripts

Prior to conducting the study, the first author used BST to train the second and third authors to act as simulated students. During the instructions component, the first author reviewed the intervention and data collection procedures, and then outlined the programmed behaviors (i.e., five instances of disruption and self-injurious behaviors; four instances of compliance). Following instructions, the first author modeled acting as the simulated student. Simulated students practiced complying with demands and engaging in problem behavior while collecting data during the role-play component. The first author delivered feedback following each role-play simulation.

The first author trained the simulated students to engage in five instances of disruptive and self-injurious behavior (defined below), spaced a minimum of 15–30 s apart, during each 5-min session. In addition, the first author trained the simulated students to engage in appropriate behavior to allow the opportunity for the participant to deliver a token four times during each session. For example, for the DRO 30-s procedure, simulated students refrained from problem behavior during four, 30-s periods. For the DRA procedure, simulated students engaged in independent compliance four times. For the remaining trials, the first author trained the simulated students to engage in compliance with either a gestural prompt or physical prompt. The first author trained simulated students to 90% criterion in engaging in all of these behaviors during their training.

Following training, the simulated students engaged in target behavior, which included self-injurious behavior and disruptive behavior, during the sessions with each participant. Self-injurious behavior was defined as any instance of the simulated student contacting her own body with an open or closed hand from a distance of 15 cm or greater or any instance of a person biting any part of her own body. Disruptive behavior was defined as any instance of the simulated student throwing an object, swiping objects off a surface, ripping or tearing items, or contacting hands or feet to a surface from a distance of 15 cm or greater. Simulated students did not display target behavior while simultaneously engaging in compliance. Given the broad goals of the in-service training were to increase skill sets in implementing evidence-based behavioral interventions, we selected these target behaviors due to the district’s history of referring students who engaged in these behaviors for behavior-analytic services.

Response Measurement

The instructor (OL) and simulated students collected data on each participant’s correct and incorrect implementation of each component using specific data sheets for the DRA and DRO procedures. The dependent measure was the percentage of correct responses per session, which was calculated by dividing the sum of correct responses by the total number of opportunities to respond and then multiplying by 100%. For both the DRA and DRO procedures, we required participants to implement the six components as indicated by Table 2. To be scored as correct for each of the components, participants must have engaged in the correct response within 5 s.

Table 2 DRA and DRO components

Observers scored component #1 (use three-step prompting including a directive, gesture, and guidance) as correct if the participant followed a no response or incorrect response with a gesture and then physical guidance within 5 s. If the participant, continued to repeat the vocal directive without the moving up the prompt hierarchy, observers scored this incorrect. If the participant did not follow the prescribed sequence (i.e., going from vocal directive to guidance) or repeated a prompt (e.g., multiple gestures), observers scored this as incorrect. Observers scored component #2 (“providing behavior-specific praise following compliance”) as correct if the participant stated the task the simulated student completed before 5 s elapsed (e.g., “Excellent job finding the letter ‘A’”). If the participant stated, “Good!” and did not pair this with an observable behavior, the observer scored the response as incorrect. Additionally, if the participant delivered the praise following 5 s, this was also scored as incorrect. Observers scored component #3 as correct if the participant delivered a token following an independent response (e.g., simulated student did not require gestural or physical prompting) within 5 s for the DRA procedure and if the simulated student did not engage in problem behavior for 30 s for the DRO procedure. For component #4 (refrain from commenting or changing facial expressions following the occurrence of problem behavior), the participant needed to continue presenting the demands without altering their facial reaction (e.g., grimace) or vocalizing on the problem behavior (e.g., “Don’t do that!”) for observers to score this component as correct. For component #5, observers scored this as correct if, following the occurrence of problem behavior, the participant delivered a demand. Finally, for component # 6 (allowing for the token exchange for edibles), observers scored this as correct if the simulated student earned four tokens and the participant prompted the student to exchange (e.g., “You can trade in.”, gestures to the board, “Remember you earned all your tokens.”).

We planned to have at least five opportunities in each session for components 1, 2, 4, and 5 in both DRA and DRO sessions. We also planned for four opportunities for component 3 and one opportunity for component 6.

During each 5-min session, simulated students (a) displayed five instances of problem behavior spaced a minimum of 15–30 s apart, (b) produced four opportunities to earn a token by either engaging in an independent compliant response in the DRA or refraining from problem behavior in the DRO, and (c) produced one opportunity for the participant to allow for a token exchange for edibles. The fluency in which participants implemented the procedure dictated if all the opportunities for token delivery and token exchange could occur in the 5-min session. Said differently, if the participant responded slowly to problem behavior, the delay could have prevented the simulated student from engaging in appropriate behavior, and thus precluded the opportunity for her to earn all the tokens needed to exchange.

One of the major goals of the broader district training was to instruct special education staff members (teachers and paraprofessionals) to interact with special education students with behavioral challenges during instructional time. Thus, for both DRA and DRO, we included demand delivery (e.g., discrete trials of identifying letters, numbers, shapes, animals, and colors and one-step motor directives).

Interobserver Agreement

Simulated students also collected data in vivo during baseline and SIP trainings for 18.5% of sessions across both DRO and DRA procedures. They were not blind to whether the participants were in baseline or treatment. We calculated interobserver agreement (IOA) scores on a step-by-step basis for each participant by dividing the number of total agreements by the number of agreements plus disagreements and converting the result to a percentage. Mean IOA scores across participants for DRA and DRO were 79.4% and 83.9%, respectively. It is possible we obtained lower IOA than expected due to the multiple roles of the simulated students, who collected data on participant performance while simulating a child with problem behavior.

Procedures

Prior to conducting baseline sessions, the first author provided each participant with the materials, as described above, and one-page written instructions for the DRA and DRO procedures (“Appendix A”, Flesch Reading Ease Scores 46.9 and 45.1, respectively). She then instructed the participants to read the materials and provided a 10-min period for them to do so. The first author did not deliver feedback to or answer questions from the participants. Instead, she redirected all questions with nonspecific statements (e.g., “Just try your best.”). All baseline and SIP training sessions occurred across two training days. As mentioned above, we conducted this study within a larger school district training on using evidence-based practices with children with special needs.

Baseline

Following the 10-min reading period, authors instructed participants to implement DRA and DRO procedures with a simulated student. The simulated student terminated the session if the participant indicated that she or he was done (e.g., a participant stating “I’m done.”) or after 5 min elapsed. No participant terminated the baseline sessions, and each participant engaged with the simulated student for the duration of each 5-min session. Neither the instructor nor the simulated students provided feedback to participants. The instructor did not prompt the participants to read the procedures prior to each baseline session. However, participants did have access to the written procedures between baseline sessions.

SIP Training

Prior to each SIP training session, the instructor set up the laptop for the two groups of three participants. The instructor stated to the participants:

It’s time to watch the presentation. Please do not skip ahead. The presentation will change slides by itself. After it is complete, one of you please raise your hand, and we will begin practicing the procedure.

The instructor began the presentation, then stepped away from the participants. She returned to the participants when someone raised their hand.

The participants viewed a PowerPoint™ presentation SIP (a total of 16 slides) on a laptop in their assigned group. Each PowerPoint™ included a corresponding video model for all the steps of the task analysis for the designated reinforcement procedure. For example, on the seventh slide of the DRA-SIP, the audio stated “provide behavior-specific praise following compliance,” and the subsequent slide provided a brief video model of this component. The teacher in the brief video model demonstrated a minimum of one exemplar. For example, following the verbal explanation of the component “use three-step prompting,” the teacher demonstrated the use of this prompting hierarchy two times.

After each step was reviewed and modeled, the final video depicted all the steps required to complete a 5-min session. For the DRO-SIP’s final video model, the teacher demonstrated the use of three-step prompting 35 times; providing behavior-specific praise following compliance 26 times; delivering a token following 30 s in the absence of problem behavior 5 times; refraining from commenting on problem behavior, resetting the timer, and placing a demand 9 times; and exchanging four tokens 1 time. For the DRA-SIP’s final video model, the teacher demonstrated the use of three-step prompting 28 times; providing behavior-specific praise following compliance 18 times; delivering the token following independent compliance 7 times; refraining from commenting on problem behavior and placing a demand 9 times; and exchanging four tokens 1 time.

In the SIPs, the first author served as the teacher and the third author served as the student. Participants watched either the DRA-SIP or the DRO-SIP in a group format following baseline sessions. Each SIP lasted approximately 12 min. “Appendix B” contains an example of a PowerPoint™ Slide and transcribed text of the corresponding voice-over instruction.

Following each presentation viewing, the authors conducted role-play simulations with each participant, individually (i.e., other group members were not present), to assess skill acquisition. During the simulations, the first author instructed the participant to conduct either the DRA or DRO procedure with the simulated student. The simulated student’s behavior during the role-play assessment matched what the student did during the video models. The authors did not provide performance feedback.

The mastery criterion was set at one session with 90% or higher correct implementation of the components for each procedure. Researchers allowed a maximum of four group viewings of each SIP. In an attempt to avoid indirect performance feedback, researchers continued running sessions with participants who reached mastery criterion. This practice ensured participants in the group could not determine who performed correctly by virtue of who was not required to view the SIP again. Following mastery of the DRA or DRO procedure, the authors assessed the untrained procedure prior to introducing its corresponding SIP. If a participant’s performance decreased below mastery level in a subsequent assessment session, her behavior was still considered mastered by virtue of prior performance (e.g., see Cathy with DRO); however, she was exposed to additional assessment and training sessions. The fourth SIP view occurred on the second day of training, and it did not differ from procedures described above. Given the time constraints of the broad school training, we could not devote more time for additional SIP presentation or assessment sessions with the simulated student.

Results

Figure 1 shows the results for the six participants. The left column of panels depicts the percentage of correctly implemented components for all the participants when implementing the DRO procedures. During baseline, none of the participants demonstrated mastery DRO procedures; however, Cathy’s performance improved across sessions. Following the DRO-SIP, Brittany, Callie, Brandy, and Cathy reached the mastery criterion. Abby and David did not achieve mastery of the DRO procedures; however, both participants’ performance improved markedly in the final session.

Fig. 1
figure 1

Percentage of correctly implemented steps by participants across baseline (BL) and training sessions for differential reinforcement of other behavior (DRO; left column of panels) self-instruction package (SIP) and differential reinforcement of alternative behavior (DRA; right column of panels) SIP

The right column of panels depicts Brandy, Cathy, David, Abby, Brittany, and Callie’s performance when implementing the DRA procedures. During baseline, none of the participants demonstrated mastery of the DRA procedures; there were increases in performance in Brandy, David, and Abby. Following the DRA-SIP, Brandy, Abby, and Brittany met the mastery criterion, whereas Cathy, Callie, and David improved following the SIP training but did not reach mastery level.

Table 3 displays each participant’s average percent correct on component skills 1–6 (defined in Table 2) during baseline sessions and following the SIP training for both DRO and DRA procedures. To aid in visual analysis of component skills (e.g., Higgins et al. 2017), black boxes denote that the participant conducted the component with an average of 100% accuracy, gray boxes denote that the participant performed that skill with an average of 70.0–99.9% accuracy, white boxes denote the participant performed the skill with less than an average of 70% accuracy, and “X” boxes indicate the simulated student did not provide an opportunity for the participant to demonstrate the skill.

Table 3 Average percent correct Implementation during BL and SIP sessions across each Component Skill

For Abby, average percent correct responding on all component skills improved following the DRO-SIP, with the highest levels on components four and six. Following the DRA-SIP, Abby’s average percent correct responding across sessions also increased for all components, with the highest levels for components one, three, four, and six. Following the DRO-SIP, Brittany’s average percent correct responding improved for five component skills with components one, three, four, and five averaging over 90% correct. Following DRA-SIP, Brittany’s performance on all component skills increased, with notable increases with components one, three, four, and six. For Callie, her performance across most DRO and DRA component skills increased following the SIP trainings. Specifically, Callie’s performance improved to high levels for component one (DRO only) and components three, four, and five (both DRO and DRA).

For Brandy, average percent correct responding on five of the six component skills improved or remained stable following the DRO-SIP. Following baseline for the DRA procedure, Brandy’s average percent correct responding increased for components one, three, four, and five. Following the DRO-SIP, Cathy’s average percent correct responding on five component skills increased, with subpar mastery on component three. For the DRA procedure, Cathy’s performance on five component skills increased, with the exception of a slight decrease for component one. Following DRO-SIP, David’s performance across five component skills increased (with the exception of a minimal decrease for component one). However, only one of these component skills (component six) increased to a high level. On average, David’s performance improved across all component skills during the DRA procedure, with the highest levels achieved for components one, four, and six.

As shown in Table 1, Abby and Brandy required several SIP views, and Brittany required only one view to achieve mastery for the DRA procedures. Callie, Brittany, and Brandy needed three to four SIP views to achieve mastery on the DRO procedures, whereas Cathy required only one SIP view to perform at mastery levels.

Discussion

Overall, the SIPs produced mastery performance for seven of 12 training opportunities across participants and differential reinforcement procedures. Notably, two participants (Brittany and Brandy) mastered both procedures and every participant displayed improved performance for both procedures following SIP training. By contrast, two participants (Abby and David) did not master either procedure with the SIP training, though their performance increased across sessions in the SIP phases. Interestingly, most participants’ performance for providing behavior-specific praise for compliance (component two) did not improve appreciably (with average increases for 3 of the 12 training opportunities). Relatedly, several training opportunities had at least one session in which there was no opportunity to score component six. It is possible that participants’ slow responding contributed to fewer opportunities to observe token exchanges.

As a whole, findings from this study important for two reasons. First, this investigation extends prior literature by evaluating a SIP under environmental arrangements in which a school consultant may encounter. As in the Weldy et al. (2014) study, we presented the SIP to groups of participants without oversight from a trainer. By decreasing the time demands on qualified trainers, SIPs may help broader training become more feasible in settings (e.g., public schools) that have limited access to qualified trainers. Although not all participants reached mastery for both procedures, we extended prior studies (Deliperi et al. 2015; Delli Bovi et al. 2017; Giannakakos et al. 2016; Lipschultz et al. 2015; Spiegel et al. 2016) by having participants view the SIP in groups, thereby increasing the efficiency of this training approach. Given our outcomes, future research should continue to determine whether SIPs are effective training tools when presented to groups of individuals simultaneously.

It is worth noting differential reinforcement procedures taught in the current study were likely more complicated than procedures taught in prior studies. For example, preference assessments involve a sequence of discrete responses for which completion of a step sets the occasion for the subsequent step. To illustrate, Weldy et al. (2014) reported seven steps for setting up and conducting the multiple-stimulus preference assessment, and the simulated students’ behavior determined whether the participant (change agent) conducted four of the seven steps. With the DRA and DRO procedures, participants implemented six steps; however, simulated student behavior influenced whether and when the change agent implemented every step. During our procedure, a participant’s implementation of a three-step prompting procedure was dependent on whether the simulated student complied with a vocal directive, complied with a gesture prompt, or engaged in problem behavior. In addition, participants needed to implement multiple components (e.g., refraining from commenting or changing facial expression paired with placing a demand following problem behavior) within close temporal proximity. In short, teaching change agents the intricacies of DRA and DRO procedures might require some individualized training time with a skilled practitioner.

Some limitations of the current study warrant discussion. First, the potential carryover effect of the SIP presentation to the untrained differential reinforcement procedure is a limitation of this study. Specifically, performance by Brandy and Cathy increased during the DRA baseline and DRO baseline, respectively. The similarity of some DRA and DRO components may have contributed to increased level of responding in baseline conditions for these participants. Future research should investigate the extent in which teaching one behavioral reduction strategy to participants facilitates response generalization to novel behavioral reduction strategies. Researchers conducted this study in an applied, in-service training setting with the ultimate goal to exposing all participants exposed to both SIPs to promote acquisition.

Second, observers produced some suboptimal IOA scores. This may be due to participants’ frequent, incorrect engagement in nonspecific contingent and noncontingent praise. Though providing noncontingent attention is not as problematic as delivering contingent attention for problem behavior, this error may have contributed to some participants’ failure to improve in component two for DRA and DRO procedures. Beyond this training, it is unclear if this error would have undermined the effects of DRA, DRO, or both when aiming to reduce problem behavior in school settings.

A third limitation is that we set the mastery criterion at one session at 90%, which does not allow for steady state responding and is a weaker demonstration of the skill in the participant’s repertoire. Global measures of staff performance may mask repeated errors on specific steps (Cook et al. 2015). By doing a further analysis of the participants’ component skills, we saw participants did make repeated errors on specific steps while still meeting the mastery criterion. Previous staff training studies have addressed this concern by requiring 100% correct implementation, only scoring a step as correct if the participant implemented it correctly across all possible opportunities per session, or both (e.g., Lipschultz et al. 2015; Weldy et al. 2014). Though 100% correct implementation across three consecutive sessions is ideal, school consultants may have to train 20 individuals within 2 h., potentially making such performance difficult to achieve. Future research should evaluate the efficiency of differing levels of mastery of trainee performance within specific time frames and the extent to which this influences change agents’ performance when working in the school setting.

Another limitation to this study is that we did not evaluate participants’ performance with actual consumers of behavior-analytic services. Nevertheless, simulated consumers provided the opportunity to expose trainees to a wide range of possible responses by consumers. In turn, trainees may be in a better position to implement the technologies correctly when actual consumers engage in similar responses. Moreover, a number of studies have demonstrated generalization of acquired skills from simulated to actual consumers following training with SIPs (Deliperi et al. 2015; Delli Bovi et al. 2017; Du et al. 2016; Giannakakos et al. 2016; Lipschultz et al. 2015; Martocchio and Rosales 2017; Vladescu et al. 2012). However, there are very few that have evaluated and included these generalization measures when training behavioral reduction strategies (Spiegel et al. 2016). Without such data, school consultants should use caution if considering this training strategy for instructing school personnel. Future researchers should study specific parameters of SIPs (e.g., duration of the package, use of colloquial language, number of procedure components) that are necessary to achieve correct responding from when implementing behavioral reduction programs. From here, researchers could conduct evaluations of efficiency, preference, maintenance, and generalization of SIPs when compared to more time and resource intensive methods (e.g., BST) in school settings.

It is also worth noting Cathy’s correct responding during the DRO session decreased after she had demonstrated mastery performance in prior sessions. She ultimately recovered mastery-level performance; however, this finding suggests that the skills acquired from a SIP might not maintain over time without a feedback process. That is, correctly imitated behavior by the change agents may require a supporting consequent event (e.g., praise from a supervisor) to maintain over time. For example, Nottingham et al. (2017) found participants mastered preference assessments and generalized their skills to consumers when experimenters delivered brief feedback. Alternatively, brief group feedback (see Luna et al. 2018) would have been a relatively easy way to enhance performance without much additional time. Future researchers should continue to investigate how to deliver feedback when instructing groups.

This study examined if a SIP would increase special education staff members’ correct implementation of differential reinforcement strategies, arguably more complex procedures than prior investigations using SIPs to train preference assessments (Giannakakos et al. 2016; Graff and Karsten 2012; Weldy et al. 2014). School consultants are often tasked with training novice staff members to implement behavior plans (Hogan et al. 2015). Consequently, it is imperative to continue to evaluate potential cost-efficient training strategies, regardless of procedural complexity, so students in special education receive quality behavioral services.