Abstract
This randomized controlled trial examined the influence of two strategies (informational packets alone and in conjunction with Webinars) aimed at increasing the adoption of motivational interviewing (MI), a patient-centered behavioral health practice supported by evidence from comparative effectiveness studies, among community health organizations responsible for delivering mental and behavioral health services. Data were obtained from 311 directors and staff across 92 community organizations. Hierarchical linear modeling was used to examine changes in decision to adopt MI. The mediating effects of multiple contextual variables were also examined. Results showed that both strategies positively influenced the decision to adopt. The positive impact on decision to adopt was significantly greater among individuals that received informational packets in conjunction with Webinars. Baseline attitudes toward evidence-based practices and pressures for change appeared to mediate this effect.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
If effective public health programs, products, and practices are not widely and effectively disseminated and implemented, they will not achieve their potential impact to improve the public's health.1 Recent studies in comparative effectiveness or translational research (CER) have focused on improving the dissemination and implementation of evidence-based practices (EBPs) to healthcare settings. Despite the acknowledgment that dissemination and implementation of evidence-based health behavior change interventions is a high priority across many sectors, there appears to be no consensus on the best methods for accomplishing it.2 This randomized controlled trial evaluates two strategies aimed at increasing the adoption of motivational interviewing (MI), a patient-centered behavioral health practice, among community health organizations responsible for delivering mental and behavioral health services across the USA.
There are multiple frameworks for examining the translation of EBPs into practice. Consistent with current definitions set forth by the National Institutes of Health,3 these frameworks often rely on the constructs of diffusion, dissemination, and implementation to understand how EBPs are adopted, implemented, and sustained by organizations and individuals. Diffusion is a broad, overarching concept which generally refers to the process by which members of a social system (e.g., a community health care provider) learn about, decide about, and act on new ideas, practices, or objects, including EBPs.4 Recent literature has acknowledged the complexities involved in the diffusion process and the need to further distinguish the specific processes involved to facilitate adequate understanding.3 , 5 As such, it is critical to consider more narrowly defined constructs, like dissemination and implementation. Dissemination involves targeting the distribution of information necessary for the implementation of a new program or practice to the intended audience.3 Implementation refers to the strategies that can be used to adopt, integrate, and sustain practice changes within a particular setting.3 A better understanding of how dissemination and implementation strategies influence the adoption of EBPs is critical in reducing the time it takes for new evidence to be routinely implementing in practice.
The Diffusion of Innovations Theory (DIT) was used to guide the development of this study.4 DIT provides a descriptive theoretical framework based on an empirical review of diffusion research studies from disciplines such as agriculture, medicine, sociology, and marketing. The theory attempts to explain how and why innovations (e.g., EBPs) are disseminated from research settings and implemented into standard clinical care. It takes into account not only the broad process of diffusion but also the importance of prior conditions within a social system.4 Research suggests that an understanding of the variables that affect the diffusion process can help to inform the effective dissemination and implementation of EBPs.6 Four key elements of DIT are expected to affect dissemination and diffusion: the innovation itself, the communication channels through which information about the EBP is transferred, the timing of the diffusion, and the social system where the diffusion takes place. Embedded in this is the innovation–decision process which postulates that members or units of a social system go through a five-stage innovation decision-making process consisting of: (1) knowledge; (2) persuasion; (3) decision; (4) implementation; and (5) confirmation. Dissemination and implementation strategies can target any stage in this process; however, little is known about which strategies are most effective at which stage in the process.1 , 7 , 8 For example, little is known about the kinds of strategies that are effective in helping organizations make an appropriate decision as to whether to adopt a particular EBP. To address this gap, this study focused on the first half of the innovation–decision process; that is, the effects of strategies targeting knowledge and persuasion on decision making.
Chambers and Kerner9 set forth a continuum of dissemination strategies ranging from the objective (e.g., scientific literature summarized in systematic reviews, individual scientific studies) to the more subjective (e.g., qualitative data, media or marketing data, word of mouth, personal experience). Some studies have concluded that active dissemination strategies (e.g., interactive workshops, educational outreach, seminars) are critical to successful diffusion,2 , 10 , 11 while others have found that passive strategies (e.g., printed educational materials) are equally effective as active or more tailored ones.12 – 16 Several systematic reviews have been conducted evaluating the effectiveness of passive strategies.17 – 23 The findings of these reviews indicate that when used alone, passive strategies are effective in increasing knowledge but have minimal effects on behavior change. The combination of passive and active strategies is much more likely to produce change.
In addition to particular strategies for dissemination and implementation, it is important to account for the multiple factors that may influence this process. Research suggests that an understanding of the variables that affect the transfer process can help to inform the adoption and implementation of the EBPs.6 This literature, including literature based on DIT, indicates that the diffusion of an innovation can be affected by the characteristics of the EBP itself, the characteristics of the organization that may adopt it, and the individuals involved in the decision to adopt and the implementation of the innovation.1
Features of intervention programs (i.e., innovations or EBPs) found to affect their diffusion include the innovation's relative advantage over existing practice, compatibility with current practices, complexity, trialability (i.e., the degree with which the innovation can be experimented), and observability (i.e., the visibility of the results to others).1 , 2 , 4 Importantly, it is not only the features of the innovation that affect its transferability into practice but also how decision makers and practitioners perceive those features. Dissemination is therefore a key component in communicating program features to those who may implement the EBP.
Organizational characteristics that affect dissemination and diffusion can include a climate that values innovation and/or shared decision making between administrators and frontline staff, and active support by leaders for a new EBP.24 , 25 In addition, the extent to which time, money, and human resources devoted to implementation are seen as available has a considerable influence on the decision to adopt. Organizations are more likely to choose innovations that fit well with their strategic plans, missions, and philosophies of treatment.26 Organizational readiness to adopt an innovation is a critical factor in the decision to adopt and in subsequent implementation.27 Indicators of organizational readiness include the perceived risk of adopting a practice, the organization's capacity to manage risk (including past experience implementing new innovations), resource availability, and staff capacity.28 – 31
Individual-level factors, or characteristics of decision makers and practitioners, are also critical to the diffusion of an EBP. Characteristics of individual providers such as gender, age, ethnicity, educational status, and years of professional experience have been associated with the adoption of EBPs,32 – 34 and these factors may vary by position within the organization.35 Provider attitudes toward EBPs and their views about their organization as well as the new EBP under consideration may serve as facilitators or barriers to EBP implementation.32 Adoption is often facilitated if staff are familiar with the EBP, recognize its utility, and see it as similar to other practices they currently implement.36 , 37 Feeling supported by one's organization through positive leadership and allocation of resources can also facilitate the adoption process.28 , 29 , 31 , 38 , 39
There are a number of gaps in the literature about effective dissemination and diffusion of EBPs, and more research is needed to determine if the effectiveness of dissemination and implementation strategies is influenced by the target population, intervention or practice being disseminated, or intended outcome.40 – 43 Few studies have examined the mediating effect of programmatic, individual, and organizational factors on the impact of various dissemination strategies. Damanpour and Schneider44 examined the input of varying contextual factors on effective dissemination strategies and found that organizational characteristics and managers' receptiveness toward innovations were more predictive of EBP diffusion than environmental factors or managers' demographic characteristics. However, more research is needed to explore the interaction among individual, organizational, and programmatic characteristics that affect the adoption and implementation of EBPs.1 , 6 , 45
The current study addresses several of the limitations in our understanding of effective dissemination and implementation strategies on the diffusion of EBPs. The purpose of this study was to examine the influence of two different strategies—one passive and the other active—on the decision to adopt MI. This study targets individuals and organizations which are early on in the adoption decision-making process; that is, those who are not familiar with MI or are contemplating its use, rather than those who have already made a decision to implement. The potential mediating effects of multiple contextual variables are also examined through this work.
Methods
This section provides a brief description of the methodology used in this study. More specific information regarding the study design and methodology is published elsewhere.46
Intervention study design
This study used a nested, experimental design to examine the influence of two different dissemination strategies on the decision to adopt a patient-centered behavioral health practice among organizations responsible for delivering community-based behavioral health services, namely, community behavioral health organizations (CBHOs) and community health centers (CHCs). Participating organizations were matched based on organization type, size, revenue, and geographic location. One organization from each matched pair was randomly assigned to receive either strategy 1 (informational packets alone—control) or strategy 2 (informational packets in combination with Webinars—intervention). The other organization in each matched pair was, by default, assigned to the other group.
MI was the EBP chosen for this study because of its solid evidence base in empirical trials, including CER trials, and consistency with the current practice structure of both CBHOs and CHCs. MI is a counseling approach that attempts to increase the patient's or consumer's awareness of the potential problems, consequences, and risks resulting from the particular behavior in question. It can be adapted to treat different conditions across different population types. MI is also relatively inexpensive to implement, which helps control for this potential influence on the study outcomes.
Sample
The target population included CBHOs and CHCs responsible for providing mental and behavioral health services in the USA. CBHOs and CHCs were initially contacted via email by the national associations representing these organizations (the National Council for Community Behavioral Healthcare and the National Association of Community Health Centers). Emails were sent to points-of-contact at approximately 3,209 organizations (1,953 CBHOs; 1,256 CHCs). In addition, the study was advertised on the national associations' Web sites and via newsletters. Three hundred forty-five organizations responded with interest in the study. Of these, 117 organizations met the eligibility criteria and were asked to participate in the study.
To be eligible, an organization had to be a CBHO or CHC and not currently implementing MI systematically throughout the organization. Eligible organizations were asked to provide the names and contact information for individuals who would be participating in the study. They were told that participants should be individuals within the organization who are responsible for making decisions regarding the adoption of new practices. No restrictions were placed on an individual's position within the organization (e.g., director, practitioner) or the number of individuals that could participate. The rationale for allowing multiple individuals to participate from each organization was to better understand the “real world” decision-making process (e.g., number and characteristics of decision makers involved, agreement among participants within the same organization) that takes place within these organizations when considering the adoption of a new EBP. Detailed analyses and results of this decision-making process are presented elsewhere.35
One hundred nine organizations provided the contact information for study participants, all of which were sent consent forms. Upon completion of the consent form, participants were sent a baseline survey. Three hundred eleven participants representing 92 organizations (43 CBHOs; 49 CHCs) provided consent and completed baseline surveys, comprising the final sample for the study. Of the 92 organizations, 65 (70.7 %) had more than one participant, with an average of three participants per organization (range 1–11).
Description of interventions
Two different strategies were evaluated in this study. The first was a passive strategy and consisted of an informational packet administered to participants in the control group. The purpose of the packet was to disseminate information to organizations to assist them in making a decision about the adoption of MI. The second strategy was active and consisted of participation in two Webinars: one focused on implementation and the second on coaching. While traditionally, Webinars have been used more as a passive strategy delivered in a lecture style format, one of the main purposes of this study was to demonstrate how Webinars can be used more “actively” as a dissemination and implementation strategy. The purpose of the implementation Webinar was to reinforce the information provided in the packet and provide participants with the opportunity to ask general questions about MI. The purpose of the coaching Webinar was to provide one-on-one guidance to organizations regarding the adoption and integration of MI through interaction with an expert in MI implementation. Because the purpose was to examine the effects of these interactive Webinars above and beyond that of the packets, those in the intervention group received the same packet as those in the control group in addition to the Webinars. More detailed information about the development of these dissemination strategies follows.
Informational Packets
The informational packet consisted of a colored two-page quick reference sheet and a 12-page document, similar to informational packets typically made available to organizations and practitioners by entities such as the Agency for Healthcare Research and Quality. It included information related to the core components of the practice, adaptations, evidence from the literature on harms and benefits including CER evidence, dissemination, and implementation resources, and examples of interventions utilizing the practice. The content of the packets was tailored to meet the needs of the study participants so that two versions of the packet were developed: one for behavioral health providers and the other for community health providers. The rationale for developing the packets was to provide organizations with information to assist them in making a decision about the adoption of MI.
The packet was reviewed by members of the study team, three implementation science experts, three active trainers from the Motivational Interviewing Network of Trainers (MINT), and two directors from the National Council for Community Behavioral Healthcare and the National Association of Community Health Centers. Feedback from these reviewers was incorporated into a final draft that was then pilot tested by eight members of the target population (five reviewers from CHCs and three reviewers from CBHOs).
Webinars
Three Webinar leaders, chosen from a list of MINT individuals recommended by Dr. William R. Miller (a developer of MI), were selected to lead the implementation and coaching Webinars based on interest, availability, and experience supporting the implementation of MI within community health and/or behavioral health settings. Webinar leaders used information provided in the MI informational packet to develop detailed presentation outlines for the 1-h Webinars. Interactive PowerPoint presentations including surveys and other group activities for participant engagement were developed for delivery through Microsoft Live Meeting and provided to intervention group participants. Webinars were tailored to meet the individual needs of CBHOs and CHCs. Webinars were pilot tested with CBHO and CHC organizations and refined based on feedback from the pilot tests.
Implementation Webinars
One MI expert presented ten didactic Webinars (five for CBHOs and five for CHCs). The Webinars were designed to reiterate information presented in the MI informational packet including staff selection; determining agency fit; training and support options; fidelity and outcome monitoring considerations; budget; and sustainability. Barriers and facilitators of successful MI implementation were also discussed.
Coaching Webinars
Two MI experts led private, interactive coaching Webinars for each intervention group organization approximately 1 month after completion of the implementation Webinars. Coaching Webinars were designed to encourage participants to discuss specific barriers and facilitators for MI adoption within their organization and incorporated worksheets for Webinar participants to complete in real time. Coaching worksheets allowed participants to expound on the specific facilitators and barriers that applied to their organization, providing opportunities for Webinar leaders to use MI techniques to elicit change talk to encourage behavioral change at the organizational level in the form of MI adoption.
Data collection procedures
Data were collected through the administration of three separate surveys (a baseline survey and two followup surveys). The baseline survey was administered at the beginning of the study prior to receipt of the informational packets. The followup survey was administered 1 month after completing all intervention components and again 3 months after completing all intervention components. Participants submitted their responses for all surveys via Qualtrics, a third-party online Web-based survey platform, and were given 2 weeks to complete. Each survey took approximately 25–30 min to finish. Three hundred eleven individuals completed the baseline survey; 248 (80 %) completed at least one followup survey with 225 (72 %) participants completing the first followup and 228 (73 %) completing the second followup. Both followup surveys were completed by 205 (66 %) participants.
Measures
Dependent measures
Table 1 presents the specific variables used in this analysis by instrument. The primary outcome measure was decision to adopt MI. For purposes of this study, decision to adopt MI was conceptualized as a continuum of stages individuals go through during the decision-making process. As such, the operationalization of this construct is based on a stage of change model47 and was adapted from McGovern and colleagues.48 A single question asked participants to “Please indicate your level of interest in adopting MI into your program.” Responses included 0 = I am not familiar with MI; 1 = I am not interested and do not think this practice would be effective in my program (precontemplative); 2 = I have considered MI but see many pros and cons (contemplative); 3 = I am leaning in the direction of adopting MI in my program (preparation); 4 = I have just begun to implement MI in my work (action); 5 = I have been using MI, and efforts are in place to maintain it (maintenance).
Several additional dependent variables were modeled in this analysis to explore potential mechanisms for observed changes including (1) attitudes toward EBPs, (2) pressure for change, (3) barriers to EBPs, (4) resources, (5) staff attributes, (6) organizational climate, (7) management strategies related to training, and (8) management strategies related to reading materials. These variables and their psychometric properties are described in detail elsewhere.46
Predictor variables
Several variables were used as potential predictors for each dependent variable. Predictor variables included (1) participant demographics (sex, race, position at organization, type of organization, number of years worked in health care); (2) number of practitioners with direct client contact; and (3) number of organizational assessments conducted within the organization focused on readiness (e.g., resources, climate). Instrumentation and psychometric properties of each variable are described further elsewhere46 and summarized in Table 1.
Data analyses
Descriptive statistical analyses were conducted using SPSS 19 (IBM, Armonk, NY, USA). Chi-square and t tests were used to identify differences between intervention and control groups. The association of the intervention with changes in respondents' decision to adopt MI score was assessed using three-level hierarchical linear models in SAS 9.2 (SAS Institute, Inc., Cary, NC, USA). This approach to the analysis appropriately estimates error while accounting for repeated measures on individuals and clustering within organizations. Random intercepts were included in the model, but random slopes were not included due to inclusion of a high number of starting parameters. Individuals who reported that they were implementing MI at baseline were excluded because the intervention was designed to promote adoption rather than maintenance. Changes in decision to adopt MI among intervention participants were compared to the changes among control participants to determine whether the intervention was associated with greater improvement after baseline. Additional candidate predictor variables were chosen based on previous cross-sectional analyses examining decision to adopt as the outcome in a two-level hierarchical linear model.49 To explore potential mechanisms for improved changes in the intervention group, models were also developed using the following secondary dependent variables; attitudes toward EBPs, pressure for change, barriers to EBPs, resources, staff attributes, organizational climate, training, and reading materials.
Two variables were specified as level 1 predictors: (1) postintervention observation (yes/no) and (2) followup 2 observation (yes/no). The estimates for postintervention represent differences between all postintervention observations as compared to baseline observations, while estimates for followup 2 were used to examine effects associated with later followup (3 months after the intervention) as compared to the first followup (1 month after the intervention).
Four variables were specified as level 2 predictors: (1) sex (male/female), (2) race (Black/African American compared to other races), (3) position (staff/directors), and (4) number of years worked in health care (1 = less than 1 year; 2 = 1 to 3 years; 3 = 3 to 5 years; 4 = more than 5 years). Sex, race, and position were entered into the model uncentered. In contrast, number of years worked in health care was centered on the group mean for each organization to model within-group variation. The centering decision does not impact significance of the results but does influence interpretation of the coefficients. In this case, each respondent is measured against the mean of that variable (e.g., number of years worked in health care) for the organization.
Three variables were specified as level 3 predictors: (1) intervention group (yes/no), (2) number of practitioners with direct client contact (0 = less than 10; 1 = 10 to 20; 2 = more than 20), and (3) number of organizational assessments used (range = 0–4). Intervention was entered into the model uncentered. The number of practitioners and number of organizational assessments were centered on the grand mean for all participating organizations to model between-group variation.
The intervention's association with improved postintervention changes as compared to the control group was assessed using two cross-level interactions: (1) postintervention observation * intervention and (2) followup 2 observation * intervention. The first term tests whether postintervention outcome changes among intervention participants were different from postintervention changes among control participants. The second term compares potential changes at the later followup by receipt of intervention.
A parsimonious model was selected by comparing the Akaike Information Criteria (AIC) for each combination of covariates in separate models, using maximum likelihood estimation to support comparisons between models with different fixed effects. AIC provides a method for selecting a model based upon fit while also maintaining parsimony by penalizing models with more parameters. Using the fixed effects from the best fit model, the model was fit using restricted maximum likelihood. Given the relatively small number of level 3 units (organizations), standard rather than robust estimates were used.50 , 51
Results
Participants and organizations
Detailed demographics of the individuals and organizations that participated in this study are presented elsewhere.35 In summary, participants tended to be White (78.8 %), female (75.9 %), directors (71.1 %), in their mid-to-late 40s (M = 47.0, SD = 11.3), with master's (63.7 %) or doctoral degrees (19.6 %), more than 5 years of experience working in health care (86.5 %), and 1–3 years of experience working in their current position (40.8 %). Organizations tended to be located in urban areas (45.1 %), be in operation for more than 10 years (66.3 %), serve between 1,000 and 10,000 clients a year (42.5 %), and have fewer than ten practitioners with direct client contact (46.7 %). Generally, participants felt that their organizations were ready for change, held relatively positive attitudes toward EBPs, and reported few organizational barriers and several facilitators of EBP implementation. Mean scores for the predictor variables included in the current models were generally consistent with those found in other studies. Participants reported a wide range of interest in adopting MI at baseline. Approximately one-fifth of participants reported that they were not familiar with MI (23.1 %); another one-fifth stated that they had just begun implementing MI in their practice (21.1 %) and therefore were excluded from these analyses.
At baseline, there were no differences between intervention and control in the main dependent variable (i.e., decision to adopt MI) or secondary dependent variables. A few differences in demographic characteristics were found between intervention and control group including position within the organization and race. The proportion of staff in the intervention group (38.7 %) was significantly greater than that in the control group (17.5 %) (X 2 = 16.90, p ≤ .001). Further, nearly one-fifth (19.6 %) of the intervention group were Black or African American compared to 7.7 % in the control group (X 2 = 9.08, p ≤ .01).
Association of the intervention with decision to adopt MI
The results of the models are presented in Table 2. Decision to adopt MI scores increased for individuals at all organizations after the baseline assessment regardless of intervention (β = 1.31, p ≤ .001). Individuals from organizations that received the intervention exhibited a significantly greater increase in decision to adopt MI score (β = 0.43, p ≤ .01) when compared to individuals from organizations in the control group. The decision to adopt at followup 2 was not significantly different from the first followup, and the covariate for the second followup was dropped based upon model fit. The model was adjusted for covariates that improved the fit of the model including black race and number of practitioners.
Mechanisms for the association between the intervention and decision to adopt MI
To explore the potential mechanisms through which the intervention promoted a higher decision to adopt MI score, eight additional outcomes were analyzed controlling for the same candidate covariates included in the main analysis. The results of these analyses are provided in Table 2. Only attitudes toward EBPs changed significantly after the baseline time point, increasing at followup (β = 0.59, p < .001). The increase, however, was significantly smaller among individuals from organizations receiving the intervention (β = −0.17, p ≤ .05). A significant difference in the pressures for change at followup was not observed (β = −0.43, p = 0.474), but an observed decline in pressures for change among individuals from intervention organizations was nearly significant (β = −1.61, p = 0.065). Later followup was associated with a significantly lower training score as compared to the first followup (β = −0.08, p ≤ .05), but there were no differences by intervention group. No other significant differences were observed comparing changes in the intervention group to controls.
Discussion
This study presents evidence of the effectiveness of both passive and active dissemination strategies on positively impacting the decision to adopt a patient-centered behavioral health intervention among two types of community-based health organizations. Few randomized controlled trials have been conducted examining the effectiveness of different dissemination and implementation strategies early on in the decision-making process and even fewer targeting community health organizations (see Fixsen et al.52 for summary of experimental research in this area).
The research design utilized in this study was a nested, randomized controlled trial in which organizations were matched based on several characteristics shown in previous literature to be associated with adoption and implementation. Such a study design has several strengths. Aside from the primary advantages of matching and randomization which minimized the potential for confounding by selection bias (important differences at baseline between the intervention and control group), and the prospective nature of the study, the design has the additional advantage that it acknowledges the nested nature of the decision-making process (groups of individual decision makers work within each organization) and allows simultaneous examination of the impact of the interventions at both the individual and organizational level. The use of a three-level hierarchical model to analyze data collected by this study allowed us to efficiently examine the impact of an intervention not only at both levels but also at different time points. Hierarchical models take into account the shared variance in hierarchically structured data and accurately estimate lower level slopes (individuals within an organization) and their implementation in estimating higher level outcomes (the organization).
In the present study, informational packets were administered to all participants, whereas those in the intervention group also received two interactive Webinars. Results showed that both interventions were effective in increasing participants' decision to adopt MI score, though the Webinars were significantly more effective than the packets. The demonstrated effectiveness of both dissemination strategies is not surprising given that previous research indicates that passive and active strategies are both effective in increasing knowledge.17 – 23 It could be said that the interventions examined in this study were successful in increasing the decision to adopt MI by increasing knowledge and awareness of MI among study participants. While research, including this study, indicates that passive and active strategies are effective in increasing knowledge, this does not guarantee that participants will act on this new information. Previous research has found that while passive strategies are effective in increasing knowledge, active strategies are needed for behavior change. Followup research is needed to determine whether participants in this study acted on the information learned through the intervention components (i.e., took steps to implement MI into practice) and whether future implementation of MI varied by intervention group. Followup research could also examine if particular information provided during each intervention component was used by participants in their implementation efforts more than other types of information. For example, other work conducted under this project53 found that participants reported liking information about the MI core components, approach, and tools and resources the most, whereas they liked the presentation of research findings as to the effectiveness of MI the least. Future work should examine the utility of each of these sections in the actual implementation of MI into practice.
Another area that remains unclear involves the cost-benefit of active vs. passive dissemination strategies. Although in this study, the Webinars were found to increase decision to adopt MI scores above and beyond that of the informational packets, they were also more costly to implement. Future work should be conducted to determine if the added benefit justifies the additional costs.
In the current study, followup data were collected at two time points, approximately 1 month after completion of the intervention components and again 3 months postintervention. Analyses were conducted to determine if any differences in outcomes were observed over time. One concern when implementing any kind of dissemination or training intervention is that there will be a “honeymoon effect” where participants are very motivated immediately after an intervention and this motivation wanes over time. No such differences in decision to adopt were found from followup 1 to followup 2, indicating consistency in the intervention effects at least over a short period of time (i.e., 2 months). Future research is needed to examine the consistency of effects over a longer time period.
Further analyses are also needed to explore the mechanisms for the association between the intervention and adoption of MI. Among the potential mediators explored, a differential postintervention change among intervention recipients was only observed for attitudes toward EBPs and pressures for change. The smaller improvement in attitudes toward EBPs among intervention recipients is surprising, but the effect size was relatively small. The stronger reduction in pressures for change among intervention participants may indicate that the pressures for change become lower as individuals adopt EBPs.
Limitations
Several limitations of this study should be noted. Data represented self-reported information from individuals who volunteered to participate in a study aimed at better understanding of how community health organizations decide to adopt EBPs. Therefore, data may be limited by social desirability and recall bias, and generalizability may be limited to those with higher baseline levels of interest in EBPs. All participants were informed that surveys were confidential and responses would not be linked to individuals, potentially reducing the possibility of social desirability bias. In addition, the specific EBP being disseminated during this intervention (i.e., MI) was concealed during enrollment to help prevent at least some self-selection bias.
While matching and randomization were utilized with the aim of reducing selection bias, some between-group differences at baseline were identified. The proportion of staff in the intervention group was significantly greater than that in the control group, and the proportion of Black or African American participants in the intervention group was higher than that in the control group. Nonetheless, no baseline differences were observed for the primary or secondary outcome measures of interest, and the number of analyses of individual baseline variables was large.35 The observation of the between-group differences at baseline may simply be the consequence of chance.
During the study, participants were not blinded to the intervention they were receiving, which may have introduced additional bias. However, participants were not told what intervention the other group was receiving, making it difficult to determine if they were receiving the control group condition or the intervention group condition. To test whether participants were aware of their intervention group allocation, participants were asked during followup to guess which condition they thought they participated in (i.e., intervention or control). Participants' responses to this question were not related to actual group allocation (X 2 = .559, p = .46), indicating that effects of lack of blinding were probably minimal.
Finally, dissemination and implementation research, including that presented here, has been criticized for its bias toward adopting and implementing EBPs,1 with too little attention to the potential benefits of not adopting an EBP. Additional research is needed to understand how and why innovations should be adopted in specific organizational contexts, and under what circumstances it is more beneficial for an organization not to adopt an innovation.
Implications for Behavioral Health
Examination of the effectiveness of dissemination and implementation strategies for the communication of behavioral health evidence to community-based practitioners and organizations is a necessary step in ensuring that community-based behavioral health care practice is based on the best available evidence and best practices. This study focused on one particular EBP, MI; however, the results may be applicable to other EBPs implemented in community-based health settings. Results indicated that both active and passive strategies were effective in increasing awareness and intent to adopt MI. The active strategy was more effective; however, it was also more expensive to implement. Given that cost-benefit analyses are not available to determine if the extra effectiveness is worth the extra cost, individual organizations should determine which strategy would be most useful for them. Dissemination strategies should be informed by the level of training and learning preferences of staff, the varying resources across implementation settings, and the extent to which staff and the organization may be amenable to change.54 One strategy alone may not be enough to take into account all these factors; therefore, different strategies may be needed.
References
Oldenburg B, Glanz K. Diffusion of innovations. In: Glanz K, Rimer BK, Viswanath K, eds. Health Behavior and Health Education: Theory, Research and Practice, 4th ed. San Francisco, CA: Jossey-Bass; 2008:313–333.
Owen N, Glanz K, Sallis JF, et al. Evidence-based approaches to dissemination and diffusion of physical activity interventions. American Journal of Preventive Medicine. 2006;31:S35–S44.
Glasgow RE, Vinson C, Chambers D, et al. National Institutes of Health approaches to dissemination and implementation science: Current and future directions. American Journal of Public Health. 2012;102:1274–1281.
Rogers EM. Diffusion of Innovations, 5th ed. New York, NY: The Free Press; 2003.
Dearing, JW. Evolution of diffusion and dissemination theory. Journal of Public Health Management and Practice. 2008;14:99–108.
Ducharme LJ, Knudsen HK, Roman PM, et al. Innovation adoption in substance abuse treatment: Exposure, trialability, and the clinical trials network. Journal of Substance Abuse Treatment. 2007;32:321–329.
Kerner J, Rimer B, Emmons K. Dissemination research and research dissemination: How can we close the gap? Health Psychology. 2005;24:443–446.
Ellis P, Robinson P, Ciliska D, et al. A systematic review of studies evaluating diffusion and dissemination of selected cancer control interventions. Health Psychology. 2005;24:488–500.
Chambers D, Kerner J. Closing the gap between discovery and delivery. In: Dissemination and Implementation Research Workshop: Harnessing Science to Maximize Health. Rockville, MD: National Institutes of Health; 2007.
Bero LA, Jadad AR. How consumers and policymakers can use systematic reviews for decision making. In: Mulrow C, Cook D, eds. Systematic Reviews: Synthesis of Best Evidence for Health Care Decisions. Philadelphia, PA: American College of Physicians; 1998:45–54.
Lewin Group. Factors Influencing Effective Dissemination of Prevention Research Findings by the Department of Health and Human Services. Final Report. Washington, DC: Lewin Group; 2001.
Brink SG, Basen-Engquist KM, O’Hara-Tompkins NM, et al. Diffusion of an effective tobacco prevention program. Part I: Evaluation of the dissemination phase. Health Education Research. 1995;10:283–295.
Gordon JS, Andrews JA, Lichtenstein E, et al. Disseminating a smokeless tobacco cessation intervention model to dental hygienists: A randomized comparison of personalized instruction and self-study methods. Health Psychology. 2005;24:447–455.
Lewis E, Mayer JA, Slymen D, et al. Disseminating a Sun Safety Program to Zoological Parks: The Effects of Tailoring. Health Psychology. 2005;24:456–462.
Glanz K, Elliott T. Diffusion of an effective skin cancer prevention program: Design, theoretical foundations, and first-year implementation. Health Psychology. 2005;24:477–87.
Hall DM, Escoffery C, Nehl E, et al. Spontaneous diffusion of an effective skin cancer prevention program through web-based access to program materials. Preventing Chronic Disease: Public Health Research, Practice, and Policy. 2010;7(6):A125.
Arnold SR, Straus SE. Interventions to improve antibiotic prescribing practices in ambulatory care. Cochrane Database of Systematic Reviews. 2005;4:CD003539.
Farmer AP, Légaré F, McAuley LM, et al. Printed educational materials: Effects on professional practice and health care outcomes. Cochrane Database of Systematic Reviews. 2008;3:CD004398.
Houts PS, Doak CC, Doak LG, et al. The role of pictures in improving health communication: A review of research on attention, comprehension, recall, and adherence. Patient Education and Counseling. 2006;61(2):173–90.
Nilsen ES, Johansen M, Oliver S, et al. Methods of consumer involvement in developing healthcare policy and research, clinical practice guidelines and patient information material. Cochrane Database of Systematic Reviews. 2006;1.
Soumerai SB, McLaughlin TJ, Avorn J. Quality assurance for drug prescribing. Quality Assurance in Health Care. 1990;2(1):37–58.
Thomas LH, McColl E, Cullum N, et al. Clinical guidelines in nursing, midwifery and the therapies: A systematic review. Journal of Advanced Nursing. 1999;30(1):40–50.
Tu K, Davis D. Can we alter physician behavior by educational methods? Lessons learned from studies of the management and follow-up of hypertension. Journal of Continuing Education in the Health Professions. 2002;22(1):11–22.
Smith DW, McCormick LK, Steckler AB, et al. Teachers’ use of health curricula: Implementation of Growing Healthy, Project SMART and Teenage Health Teaching Modules. Journal of School Health. 1993;63:349–354.
Parcel GS, O’Hara-Tomkins NM, Harris RB, et al. Diffusion of an effective tobacco prevention program, Part II: Evaluation of the adoption phase. Health Education Research. 1995;10:297–307.
Denis JL, Hebert Y, Langley A, et al. Explaining diffusion patterns for complex healthcare innovations. Health Care Management Review. 2002;27:60–74.
Panzano PC, Roth D. The decision to adopt evidence-based and other innovative mental health practices: Risky business? Psychiatric Services. 2006;57(8):1153–1161.
Gagnon M, Labarthe J, Legare F, et al. Measuring organizational readiness for knowledge translation in chronic care. Implementation Science. 2011;6(72):1–10.
Meijers JM, Janssen M, Cummings GG, et al. Assessing the relationships between contextual factors and research utilization in nursing: Systematic literature review. Journal of Advanced Nursing. 2006;55(5):622–635.
Panzano PC, Seffrin BA, Chaney-Jones S, et al. The innovation diffusion and adoption research project (IDARP): Moving from the diffusion of research results to promoting the adoption of evidence-based innovations in the Ohio mental health system. New Research in Mental Health. 2004;16.
Saldana L, Chapman JE, Henggeler SW, et al. The organizational readiness for change scale in adolescent programs: Criterion validity. Journal of Substance Abuse Treatment. 2007;33:159–169.
Aarons GA. Measuring provider attitudes toward evidence-based practice: Organizational context and individual differences. Child and Adolescent Psychiatric Clinics of North America. 2005;14:255–271.
Aarons GA, Sawitzky AC. Organizational climate partially mediates the effect of culture on work attitudes and staff turnover in mental health services. Administration and Policy in Mental Health. 2006;33(3):289–301.
Taxman FS, Belenko S. Implementing Evidence-Based Practices in Community Corrections and Addiction Treatment. New York, NY: Springer; 2012.
Williams JR, Dusablon T, Williams WO, et al. Characteristics of community health organizations and decisionmakers considering the adoption of comparative-effectiveness research. Journal of Behavioral Health Services and Research. In review.
Haug NA, Shopshire M, Tajima B, et al. Adoption of evidence-based practices among substance abuse treatment providers. Journal of Drug Education. 2008;38(2):181–192.
Gallo KP, Barlow DH. Factors involved in clinician adoption and nonadoption of evidence-based interventions in mental health. Clinical Psychology: Science and Practice. 2012;19(1):93–106.
Rogers EM. Diffusion of preventive innovations. Addictive Behaviors. 2002;27(6):989–993.
Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Administration and Policy in Mental Health and Mental Health Services Research. 2011;38(1):4–23.
Glasgow RE, Emmons KM. How can we increase translation of research into practice? Types of evidence needed. Annual Review of Public Health. 2007;28:413–33.
Green L, Ottoson J, Hiatt R, et al. Diffusion, dissemination, and implementation of evidence-based public health. Annual Review of Public Health. 2009;30:151–74.
Brownson RC, Fielding JE, Maylahn CM. Evidence-based public health: A fundamental concept for public health practice. Annual Review of Public Health. 2009;30:175–201.
McGovern MP, Carroll KM. Evidence-based practices for substance use disorders. Psychiatric Clinics of North America. 2003;26:991–1010.
Damanpour F, Schneider M. Phases of the adoption of innovation in organizations: Effects of environment, organization and top managers. British Journal of Management. 2006;17:215–236.
Brownson RC, Ballew P, Brown KL, et al. The effect of disseminating evidence-based interventions that promote physical activity to health departments. American Journal of Public Health. 2007a;97:1900–1907.
Williams JR, Tregear SJ, Dusablon T, et al. Increasing adoption of comparative effectiveness research in community behavioral health: methodology. Journal of Behavioral Health Services and Research. Accepted.
Prochaska JO, DiClemente CC. Stages and processes of self-change of smoking: Toward an integrative model of change. Journal of Consulting and Clinical Psychology. 1983;51:390–395.
McGovern MP, Fox TS, Xiew H, et al. A survey of clinical practices and readiness to adopt evidence-based practices: Dissemination research in an addiction treatment system. Journal of Substance Abuse Treatment. 2004;26:305–312.
Williams JR, Blais MP, Banks D, et al. Predictors of the decision to adopt motivational interviewing in community health settings. Journal of Behavioral Health Services and Research. 2013. doi:10.1007/s11414-013-9357-8.
Cheong YF, Fotiu RP, Raudenbush SW. Efficiency and robustness of alternative estimators for 2- and 3-level models: The case of NAEP. Journal of Educational and Behavioral Statistics. 2001;26:411–429.
Hox JJ, Maas CJM. The accuracy of multilevel structural equation modeling with pseudobalanced groups and small samples. Structural Equation Modeling. 2001;8(2):157–174.
Fixsen DL, Naoom SF, Blase KA, et al. Implementation Research: A Synthesis of the Literature. FMHI Publication No. 231. Tampa, FL: University of South Florida, Louis de la Parte Florida Mental Health Institute, The National Implementation Research Network; 2005.
Castro L, Williams JR, Dusablon T, et al. Design, development, and evaluation of printed educational materials on motivational interviewing for community behavioral health providers. Journal of Behavioral Health Services and Research. In review.
Kerner JF, Hall KL. Research dissemination and diffusion: Translation within science and society. Research on Social Work Practice. 2009;19(5):519–530.
Texas Christian University, Institute of Behavioral Research. Survey of Structure and Operations (TCU SSO). 2006. Available at: www.ibr.tcu.edu. Accessed 03 November 2010.
Texas Christian University, Institute of Behavioral Research. The Organizational Readiness for Change: Treatment Director Version (TCU ORC-D). 2002. Available at: www.ibr.tcu.edu. Accessed 03 November 2010.
Texas Christian University, Institute of Behavioral Research. The Organizational Readiness for Change: Treatment Staff Version (TCU ORC-S). 2005. Available at: www.ibr.tcu.edu. Accessed 03 November 2010.
Aarons GA. Mental health provider attitudes toward adoption of evidence-based practice: The Evidence-Based Practice Attitude Scale (EBPAS). Mental Health Services Research. 2004;6(2): 61–74.
Acknowledgments
This study was supported by SAMHSA Contract No. HHSS283200700040I/HHSS28342003T. We appreciate the generous support and participation of the community health organization directors and staff who participated in this study and the assistance provided by the National Council for Community Behavioral Healthcare and the National Association of Community Health Centers in recruitment, retention, and advice with overall study design. Special thanks to all study staff from MANILA Consulting Group, Inc., McLean, VA.
Disclaimers
This report was prepared under contract by MANILA Consulting Group, Inc., for SAMHSA. The content of this publication does not necessarily reflect the views or policies of SAMHSA or the U.S. Department of Health and Human Services.
Conflicts of Interest
There are no conflicts of interest.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Williams, J.R., Williams, W.O., Dusablon, T. et al. Evaluation of a Randomized Intervention to Increase Adoption of Comparative Effectiveness Research by Community Health Organizations. J Behav Health Serv Res 41, 308–323 (2014). https://doi.org/10.1007/s11414-013-9369-4
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11414-013-9369-4