Abstract
Implementation research is a new scientific discipline emerging from the recognition that the public does not derive sufficient or rapid benefit from advances in the health sciences (Berwick DM, JAMA 289:1969–1975, 2003; Lenfant C, N Engl J Med 349:868–874, 2003). One often-quoted estimate claims that it takes an average of 17 years for even well-established clinical knowledge to be fully adopted into routine practice (Kiefe CI, Sales A, J Gen Intern Med 21(Suppl 2):S67–S70, 2006). In this chapter, we will discuss particular barriers to evidence implementation, present tools for implementation research, and provide a framework for designing implementation research studies, emphasizing the randomized trial. The reader is advised that this chapter only provides a basic introduction to several concepts for which new approaches are rapidly emerging. Therefore, our goal is to stimulate interest and promote additional in-depth learning for those who wish to develop new implementation research projects or better understand this exciting field.
I think when people look back at our time, they will be amazed at one thing more than any other. It is this-that we do know more about ourselves now than people did in the past, but that very little if this knowledge has been put into effect.
D Lessing http://daronlarson.blogspot.com/2006/12/prisons-we-choose-to-live-inside.html
© [Springer International Publishing AG]
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
Keywords
- Implementation research tools
- Rogers Diffusion Theory
- Translational barriers
- Academic detailing
- Pay-for-performance
Introduction
Overview and Definition of Implementation Research
Implementation research is an emergent discipline born from the recognition that the public does not derive sufficient or rapid benefit from advances in the health sciences [1, 2]. Implementation research bridges the gap between scientific knowledge and its application to daily practice with the overall purpose of improving the health of individuals and populations. One often-quoted estimate claims that it takes an average of 17 years for even well-established clinical knowledge to be fully adopted into routine practice [3]. In addition, approximately half of trials funded by the National Institutes of Health were published in peer-reviewed publications two and a half years after study completion [4].
For example, in 2000, only one-third of patients with coronary artery disease received aspirin when no contraindications to its use were present; [2] furthermore, a landmark study estimated that the American public was only receiving about 55 % of recommended care [5]. Implementation research definitions are shown in Table 13.1.
A glossary of terms used in implementation research is now available [13, 14]. The definition of implementation research may be expanded to encompass work that promotes patient safety and eliminates racial and ethnic disparities in health care. Health disparities implementation research aims to identify strategies to close gaps in health care through culturally-appropriate interventions for patients, clinicians, health care systems, and populations [15–18]. Under-represented populations make up a significant portion of the U.S. population, shoulder a disproportionate burden of disease, and receive inadequate care [19]. According to the U.S. National Institute of Health (NIH), ‘dissemination and implementation research intends to bridge the gap between public health, clinical research, and everyday practice by building a knowledge base about how health information, interventions, and new clinical practices and policies are transmitted and translated for public health and health care service use in specific settings’ [6].
Gaps in health care may be classified as ‘errors of omission,’ (failure to provide necessary care [20]) and ‘errors of commission,’ such as the delivery of unnecessary or inappropriate care which causes harm. A landmark report from the Institute of Medicine drew attention to patient safety and the concept of preventable injury [21]. Studies of patient safety have focused on ‘medical error resulting in an inappropriate increased risk of iatrogenic adverse event(s) from receiving too much or hazardous treatment (overuse or misuse)’ [20].
For example, inappropriate antibiotic use may promote microbial resistance and cause unnecessary adverse events. Since 1999, public efforts have been underway to promote appropriate prescribing of antibiotics for acute respiratory infections (ARIs) [22]. Based on well-designed studies demonstrating no benefit, guidelines have long recommended against antibiotic use for acute bronchitis; [23, 24] however, physicians continue to prescribe antibiotics for patients diagnosed with ARIs. Although overall antibiotic use for ARIs declined between 1995 and 2002, use of broad-spectrum antibiotic prescriptions for ARIs increased [25]. A more recent implementation research project successfully used a multidimensional intervention in emergency departments to decrease antibiotic prescribing [26].
In response to what may be perceived as overwhelming evidence that thousands of lives are lost each year from errors of omission and commission, there have been strong national calls for health systems, hospitals, and physicians to adopt new approaches for moving evidence into practice, but rigorous supporting evidence is often lacking [27, 28].
As our understanding of implementation science is evolving, local clinicians and health systems must strive to improve the quality of care for every patient. Certain local decisions must be based on combinations of incomplete empiric evidence and personal experience. As with the clinician caring for the individual patient, every decision about local implementation cannot be guided by data from a randomized trial [29, 30]. However, a stronger evidence base is needed to inform wide-spread implementation efforts. Widespread implementation beyond evidence raises concern about unintended consequences and opportunity costs from public resources wrongly expended on ineffective interventions [30].
Implementation researchers use a variety of techniques, ranging from qualitative exploration to the controlled, group-randomized trial. For example, methods used in social, cognitive, and organizational psychology are also applicable to implementation research [31]. Berwick reminds us of the importance of understanding the mechanism and context through which implementation techniques exert their potential effects within complex human systems [32]. Berwick cautioned that important lessons may be lost through aggregation and rigorous scientific experimentation, challenging the implementation research community to reconsider the basic concept of evidence, itself.
Interventions for translating evidence into practice must operate in complex, poorly understood environments with multiple interacting components that may not be easily reducible to a clean, scientific formula. Therefore, we later present situational analysis as a framing device for implementation research. Nonetheless, in keeping with the theme of this book, we mainly focus on the randomized trial as one of the many critical tools for implementation research.
In summary, implementation research is an emerging body of scientific work seeking to close the gap between knowledge generated from the health sciences and routine practice, ultimately improving patient and population health outcomes. Implementation research, which encompasses the patient, clinician, health system, and community, may promote the use of needed services or the avoidance of unneeded services. Implementation research often focuses on patients who are vulnerable because of race/ethnicity or socioeconomic position. By its very nature implementation research is inter-disciplinary.
In this chapter, we discuss barriers to evidence implementation, present tools for implementation research, and provide a framework for designing implementation research studies. The reader is advised that this chapter only provides a basic introduction to several concepts for which new approaches are rapidly emerging. Therefore, our goal is to stimulate interest and promote additional in-depth learning for those who wish to develop new implementation research projects or better understand this exciting field.
Overcoming Barriers to Evidence Implementation
Successful implementation of evidence-based interventions largely depends on their fit with the preferences and priorities of those who shape, deliver, and participate in healthcare [33]. Although the conceptual basis for moving evidence into practice has not been fully developed, a solid grounding in relevant theory may be useful to those designing new implementation research projects [34]. Many conceptual models have been developed in other settings and subsequently adapted for translating evidence into practice [35]. For example, implementation researchers frequently apply Roger’s theory describing innovation diffusion. Rogers proposed three clusters of influence on the rapidity of innovation uptake is influenced by:
-
Perceived advantages of the innovation
-
The classification of new technology users according to rapidity of uptake; and
-
Contextual factors [36].
First, potential users are unlikely to adopt an innovation that is perceived to be complex and inconsistent with their needs and cultural norms. Second, rapidity of innovation uptake often follows a sigmoid-shaped curve, with an initial period of slow uptake led by the ‘innovators.’ Next follows a more rapid period of uptake led by the early adopters, or ‘opinion leaders.’ During the last adoption phase, the rate of diffusion again slows as the few remaining ‘laggards’ or traditionalists adopt the innovation. Finally, contextual or environmental factors such as organizational culture exert a profound impact on innovation adoption, a concept that is explored in more detail in the following sections of this chapter.
Consistent with the model proposed by Rogers, multiple barriers often work synergistically to hinder the translation of evidence into practice [37]. Interventions often require significant time, money, and staffing. Implementation sites may experience difficulties in implementation as a result of limited resources, competing demands, and entrenched practices. For example, the intervention may have been developed and tested under circumstances that differ from those at the planned implementation site. Further, the implementation team may not adequately understand the environmental characteristics proposed by Roger’s diffusion theory as critical to the adoption of innovation. Because of such concerns a thorough environmental analysis is needed prior to widespread implementation efforts [37].
Building upon models proposed by Sung et al. [38] and Rubenstein et al. [8], Figure 13.1 depicts the translational barriers implementation research seeks to overcome. The first translational roadblock lies between basic science knowledge and clinical trial design. The second roadblock involves translation of knowledge gained from clinical trials into meaningful clinical guidance, which often takes the form of evidence-based guidelines.
The third roadblock specific to implementation science occurs between current clinical knowledge and routine practice, carrying important implications for individual practitioners, health care systems, communities, and populations. Given the expansive nature of this third roadblock, a multifaceted armamentarium of tools is required. One tool, industrial-style quality improvement, described below in more detail, operates at the level of the clinical microsystem, the smallest, front-line functional unit that actually delivers care to a patient [39]. Clinical microsystems consist of complex adaptive relationships among patients, providers, support staff, technology, and processes of care. To achieve sustainable success, researchers seeking to overcome this third translational barrier need to be effective advocates for changes in larger macrocosms of the healthcare system including local and governmental health policy. Finally, implementation research may inform clinical trials and basic science.
As an underlying source of challenge for the U.S. healthcare system as a whole, health disparities contribute equally as a barrier to overcome in implementation research. Members of minority populations such as African-Americans and Latinos in particular, as well as individuals of low socioeconomic status disproportionately fall victim to markedly dismal health outcomes as compared to their white counterparts [40]. Despite the advancements in health and life expectancy as a country, population specific black-white gaps continue to persist in areas such as access to care, quality of care, chronic disease risk factors, and disease incidence and related mortality [41–43]. These examples of inequity have seeped into multiple tiers of the healthcare system from which implementation research is not exempt.
The rapidly changing U.S. demographics indicate that this very marginalized minority population will soon comprise the majority of the U.S. population. The 2012 U.S. Census reported that 50.4 % of 1-year olds born nationwide were racial/ethnic minorities and that Latinos are the largest and fastest growing ethnic group, currently comprising 16.7 % of the population [42, 44]. At this rate of growth, it is postulated that the U.S. will be a majority minority society by 2050, if not sooner [44].
As such, health disparities are an added dimension to the barriers to overcome in evidence implementation. Implementation science aims to make evidence-based findings work in real world patients. However systems, providers and patients in the real world are much different than what is encountered in the research process. Therefore, studies must be specially tailored to include these specific differences among communities and cultures. Should the health disparities in minority groups continue to persist, they will inevitably impede upon the success of implementation research and impose upon the well-being of the nation.
Finally, to promote the spectrum of research depicted in Fig. 13.1, the 2003 NIH Roadmap acknowledges translational research as an important discipline [45]. In fact, several branches of the NIH now have open funding opportunities for implementation research. The integration of research findings from the molecular to the population level is a priority. The Roadmap seeks to join communities and interdisciplinary academic research centers to translate new discoveries into improved population health [46].
Implementation Research Tools
The tools used to translate clinical evidence into routine practice are varied, and no single tool or combination of tools has proven sufficient or completely effective. Furthermore, it may not be the tool itself but how it is implemented in a system that drives change [47]. In fact, this lack of complete effectiveness spurs implementation research to develop innovative adaptations or combinations of currently available tools [48].
Below, we provide an overview of available tools, which are intended as basic building blocks for future implementation research projects. Although different classification systems have been proposed [49], we arranged these tools by their focus: on the patient, the community, the provider, and the healthcare organization. We acknowledge that this classification is somewhat arbitrary because several implementation tools overlap multiple categories.
Patient-Based Implementation Tools
A growing body of evidence suggests that patients may be successfully ‘activated’ to improve their own care. For example, a medical assistant may review the medical record with the patient and encourage the patient to ask questions at an upcoming visit with the physician. Patients exposed to such programs had better health outcomes, such as improved glycemic control for those with diabetes [50, 51]. In another study, a health maintenance reminder card presented by the patient to the physician at appointments significantly increased rates of influenza vaccination and cancer screening [52].
Other interventions have taught disease-management and problem solving skills to improve chronic disease outcomes. Teaching patient self-management skills is more effective than passive patient education, and these skills have been shown to improve outcomes and reduce costs for patients with arthritis and asthma [53]. As part of the ‘collaborative model,’ self-management is encouraged through better interactions between the patient, physician, and health care team. The collaborative model includes: (1) identifying problems from the joint perspective of the patient and clinical care team; (2) targeting problems, setting appropriate goals, and developing action plans together; (3) continuing self-management training and support services for patients; (4) active follow up to reinforce the implementation of the care plan [53].
Community-Based Implementation Tools
The Community Health Advisor (CHA) model has been implemented throughout the world to deliver health messages, promote positive health behavior change, and facilitate access to the health care system [54]. Similarly, the Community Health Worker (CHW) model has been used to engage medically underserved communities on a number of different health issues to help individuals overcome financial, social, political, and cultural barriers to health care [55]. Based on the CHA/CHW models, community members, usually without formal education in the health professions, undergo special training and certification in order to carry out an intervention or research protocol in their local community. CHA/CHW interventions have been used to promote prevention and treatment for a large array of conditions, including cancer, asthma, cardiovascular disease, depression, and diabetes. These programs have also been developed to decrease youth violence and risky sexual behavior, and may be especially relevant for underserved populations and those living in rural areas. Although promising, CHA/CHW interventions often rely on volunteer workers who may be vulnerable to stress and burnout from work overload. Also, intense training and oversight is often required to assure the accuracy of the health messages being transmitted. A review by Swider found limited high-quality evidence that CHA interventions actually improve health outcomes, which is postulated to be a result of the poor quality of the studies. As such, Swider also called for additional rigorous research on the efficacy and underlying mechanisms through which CHA/CHW interventions work [56]. A more recent review commissioned by the Robert Wood Johnson Foundation found that specific CHA interventions may reduce health disparities, particularly for patients with hypertension and diabetes [17].
As a vital community-based implementation tool, it is useful to consider the basic principles of interaction with a community as the foundation for research success. In the city of Lawrence, Massachusetts the Lawrence Research Initiative was created in order to promote community-participatory and community-responsive research [57]. A document to closely guide the research process was created that included:
-
The core principles of a partnership approach to research
-
Questions for research partnership agreements for researchers and community groups to review
-
Steps to building successful research partnerships
-
Glossary of research terms in order to develop a common vocabulary that empowers the community to communicate with researchers.
The core principles of partnership as defined in the Lawrence Research Initiative are notions that are applicable to a broad scope of community-based implementation projects [58]. The principles are as follows:
-
Research is helpful to community development
-
True partnerships between the community and academia make better science
-
Researchers and members of the community can and should create good partnerships based on fairness and positive exchanges.
-
Partnerships should be based on fair and equitable distribution of resources
For community-based implementation research to be truly successful, it should be rooted in the foundation of good community partnerships and relationships. This is a principle that will hold true across the use of community health workers, community health advisors, and community based participatory research. Equitable partnerships are key.
Provider-Based Implementation Tools
Clinical Guidelines
Clinical guidelines have been defined as ‘systematically developed statements to assist practitioners’ and patients’ decisions about appropriate health care for specific clinical circumstances’ [59]. Ideally, guideline development involves a complete review of the relevant literature; however, a Canadian group demonstrated literature reviews were less likely to be completed on more recently developed guidelines [60]. During the last 30 years, guideline dissemination efforts may be suboptimal, leading only to modest improvements in care [60, 61]. However, guideline dissemination alone is not sufficient for implementation of best practices [62].
For many clinical situations encountered today, thousands of evidence-based guidelines and practice recommendations have been published. Such sheer volume often precludes the individual practitioner from implementing all recommendations for every patient. As an example, Boyd et al. noted that if one were to treat a hypothetical 79 year old woman with diabetes, chronic obstructive pulmonary disease (COPD), hypertension, osteoporosis, and osteoarthritis, and follow all recommended guidelines for her multiple co-morbidities, the patient would require 12 medications at a cost of $406 per month [63].
Continuing Medical Education
Continuing medical education (CME), a requirement for ongoing medical licensure, has traditionally relied on text-based, didactic methods to affect clinical knowledge, skills, attitudes, practice patterns, and patient outcomes [64]. However, passive, text-based educational materials and formal CME conferences do not lead to measurable improvements in practice patterns [65, 66]. Rather, CME using interactive techniques which actively engage physicians may have small effects improving practice patterns and patient outcomes [67–70]. Physicians who reflect on their own individual performance may identify areas for improvement and seek CME through multifaceted, self-directed learning opportunities. The use of multiple modalities that promote active learning – such as case-based problem solving – has yielded modest improvements in clinical practice [71]. In general, however, more complex behavioral change likely requires practicing skills similar to traditional quality improvement (QI) [72].
With the advantages of being convenient, flexible, and inexpensive, the Internet has become a useful platform to reach a wider audience for interactive CME, while maintaining an effectiveness comparable to traditional approaches [73]. Fordis et al. conducted a randomized controlled trial comparing live, small-group interactive CME workshops with Internet CME [74]. Both groups focused on cholesterol management. All physicians received didactic instruction, interactive cases with feedback, practice tools and resources, and access to expert advice. Knowledge scores for physicians in the Internet CME group increased more than scores for those in the live CME group. Additionally, the online CME group demonstrated a statistically significant improvement in appropriate drug treatment for high-risk patients. Success of the Internet CME may have been partially driven by the participants’ ability to repeatedly return to the website for reinforcement and the ability to structure the learning experience to meet individual needs.
Academic Detailing
Academic detailing relies on site visits to physicians’ offices for intense relationship building and one-on-one information delivery. Important components for successful detailing include: (1) assessment of baseline knowledge and motivations for current behavior; (2) articulating clear objectives for education and behavior; (3) gaining credibility with ties to respected organizations through ongoing relationship building; (4) encouraging physicians to actively participate in educational interventions; (5) using graphic representations for educational materials; (6) focusing on a limited number of ‘take-home’ points; and, (7) supplying positive reinforcement for improved behaviors during follow up [75]. Representatives from pharmaceutical companies have effectively used academic detailing to boost product sales. In a systematic review, academic detailing alone yielded small effects on medication prescribing practices [76].
Opinion Leaders
Several implementation programs have relied on influential colleagues to disseminate evidence-based practices [79]. Opinion leader strategies may include using celebrities, employing people in leadership positions, and asking those doing front-line work to refer ‘up the ladder.’ In a systematic review, opinion leaders may have a positive effect on evidence-based practice uptake when tested in randomized controlled trials [77].
Physician Audit and Feedback
The utility of audit and feedback hinges on developing credible data-driven summaries of how patient populations are being managed. In theory, such reports may prompt clinicians to reflect on their personal clinical practices and motivate subsequent improvement. Performance feedback may focus on outcomes (such as percentage of patients with diabetes who have achieved glycemic control) or process (such as the percentage of patients with diabetes for whom the physician measured glycemic control). The credibility of performance feedback relies on the ability to capture the many clinical nuances that the physician must consider when delivering care to the individual patient. Because the difficulties in capturing these clinical nuances have not yet been completely surmounted, comparisons of performance to a data-driven, peer-based benchmark may be more appropriate than comparison to an arbitrary standard of perfect performance [80]. A systematic review of randomized trials of audit and feedback studies demonstrated small effects on professional performance. The effect varied by which targeted behavior was chosen. Additionally, the analysis suggests that audit and feedback may be more effective when: the baseline performance is low; the information is provided by a manager or colleague multiple times, communicated in verbal and written formats, and when it is linked to specific goals and an action plan [78].
Organization-Based Implementation Tools
Industrial-Style Quality Improvement
This type of improvement activity originated outside of health care and has acquired such labels as Total Quality Management (TQM) and Continuous Quality Improvement (CQI). These approaches make two fundamental assumptions: (a) that poor outcomes are attributable to system failures, rather than lack of individual effort or individual mistakes, and (b) achieving improvement and excellence, even in the absence of system failures, is possible through iterative cycles of planning, acting, and observing the results. In general, complex systems must have built-in redundancy to function well. If an individual makes a mistake at one point in the system, checks and balances built into other parts of the system may prevent an adverse event. However, as described in the example below, patient safety maybe endangered by simultaneous failure of multiple system components, thus defeating built-in redundancy.
As a simple example, multiple mechanisms should be in place to ensure that incompatible blood products are not given to hospitalized patients. Delivery of the wrong blood type to a patient requires failure at multiple points, including preparation of the blood in the blood bank and administration of the blood by the nurse. Taking such a systems approach stands in stark contrast to blaming individuals, thereby avoiding low morale and reluctance to disclose mistakes.
Improvement activity usually proceeds through a series of ‘plan-do-study-act’ cycles. These cycles emphasize measuring the process of clinical care delivery at the level of the clinical microsystem, which has been previously described. Here, small amounts of data guide the initial improvement process. The process emphasizes small, continuous gains through repeated cycles and does not rely on the statistical significance of the measurements. Although many health care institutions have adopted such methodology based on compelling case studies, additional studies with high-quality experimental methods are still needed [81].
Systems Reengineering
Instead of incremental changes to clinical microsystems, major redesign of the entire system may be undertaken. For example, in the 1990s the Veterans’ Health Administration (VHA) undertook a major reengineering of its health care system, focusing on the improved use of information technology (IT), the measurement and reporting of performance, and the integration of services [82]. By 2000, the VHA had made statistically significant improvements in nine areas, including preventive care, outpatient care (diabetes, hypertension, and depression), and inpatient care (acute myocardial infarction and congestive heart failure). Additionally, the VHA performed better than the fee-for-service Medicare system on 12 of 13 quality measures [82]. Because systems engineering requires changes on such a large scale, little evidence exists about its efficiency and effectiveness in yielding more improvements than smaller changes [3].
With the passage of the Affordable Care Act of 2010, U.S. healthcare systems are making necessary changes to how care is delivered and reimbursed in order to increase the quality of care, reduce costs, and improve patient outcomes. The Center for Medicare & Medicaid Innovation (CMMI) [83] has funded hundreds of demonstration projects to test new models of accountable care organizations, primary care transformation, bundled payments of related services, adoption of best practices, and new care and payment models. The aim is for healthcare systems to implement improvements, and demonstrate positive changes to patient outcomes and cost savings, then CMMI will disseminate the knowledge for quick uptake. Further study of how these changes are implemented as well as program evaluation is needed in the coming years.
Computer-Based Systems
Computer-based systems target links in the process of care delivery that are most prone to human error. Such systems may provide clinical decision support by assisting the clinician with making a diagnosis, choosing among alternative treatments, or deciding upon a particular drug dosage. Other functions may include delivery of clinical reminders and computerized provider order entry (CPOE) [84]. A systematic review documented improvements in time to therapeutic goals, decreases in toxic drug levels and adverse reactions, and shorter hospital stays [85]. However, adverse effects of computer-based systems have also been reported, including increased mortality rates, increased rates of adverse drug reactions, delays in medication administration, increased workload, and new types of errors [86–89]. A systematic review of studies reporting the effect of CPOE on inpatient medication errors also demonstrated mixed results [90], with increases as well as decreases in errors after the introduction of computer-based systems. Therefore, computer-based systems should not be implemented without safeguards to prevent unintended consequences. We need more work to better understand how computer-based systems interact with human users and the complex health care environment and how these interactions affect quality, safety, and outcomes.
Public Report Cards
Public reports on the quality of health care delivered by institutions are proliferating. For example, public reports may focus on risk-adjusted mortality after cardiac surgery or quality at long-term care facilities. In addition, such reports will probably be expanded to include physician groups and individual physicians. Public reports are often promoted under the assumption that the public will use them to choose high-quality providers, thus better enabling a competitive ‘medical marketplace.’ Although scant evidence links report cards to improved health care [91], report cards may have profound adverse effects: (1) physicians may avoid sicker patients to improve their ratings; (2) physicians may strive to meet the targeted rates for interventions even in situations where intervention is inappropriate; and, (3) physicians may ignore patient preferences and neglect clinical judgment [92]. Even worse, report cards may actually widen gaps in health disparities [92].
The Centers for Medicare & Medicaid Services (CMS) introduced public reporting of quality measures in 2001 [93]. In the last several years, the amount of information accessible to the public has grown tremendously. On the CMS website, the public can see data from hospitals, inpatient rehabilitation facilities, long term care hospitals, nursing homes, and home health agencies.
Pay-for-Performance (P4P)
Currently, there is mounting pressure to tie reimbursement for health care services to quality measurement. Although allowing market forces to freely operate through P4P reimbursement may seem logical, systematic reviews have not yielded conclusive results. Because not everything that is important is currently measured, linking reimbursement to measured quality may divert attention from important, but unmeasured aspects of care (i.e., ‘spotlight’ effect). As with public reporting, P4P may actually widen health disparities, although empiric data are lacking.
To date, evidence for the effectiveness of P4P in improving the delivery of health care is uncertain [94]. One study found that when implemented in physician practice groups, P4P produced improvements for those with higher baseline performance but had minimal effect on the lowest performers [95]. Glickman et al. found hospitals voluntarily participating in the P4P initiative for myocardial infarction did not show appreciable improvement [96]. A recent study found that hospitals participating in P4P and public reporting programs sponsored by CMS had slightly greater improvements in quality than those only participating in the public reporting program [97]. Several ongoing studies may soon deliver new insights about P4P.
Designing Implementation Research Studies
Because the implementation science base is still emerging, researchers have at their disposal an array of tools that are variously effective, depending upon the patient population and delivery setting. Moving beyond the tools described above, is the need to develop innovative adaptations and approaches to bridge the gap between clinical knowledge and health care practice. It is necessary to test the effectiveness of these new approaches with rigorous scientific methods to avoid adverse consequences from the wide-spread dissemination and adoption of unproven interventions [30]. Therefore, in the remainder of this chapter, we discuss the critical design elements for implementation randomized controlled trials, followed by an example of an implementation research study.
Overview of Implementation Research Study Design
Randomized designs for implementation research, somewhat analogous to the traditional clinical trial, allow causal inference and offer protection from measured and unmeasured confounding (See Chap. 3) [48]. As described below in more detail, such designs include an active intervention, random allocation to a comparison or intervention group, and blinded assessment of objective endpoints. Although many of the same principals are involved in clinical RCTs and implementation RCTs, these will be reviewed with emphasis on some of the differences between the two.
Falling lower in the hierarchy of evidence, implementation studies may use other non-randomized or controlled designs. For example, a research team may observe a single group for changes in health care delivery or patient outcomes before and after intervention implementation. In this case, the observed changes may result from multiple factors not associated with the intervention. Secular trends may produce broad, population-based changes, independent of the intervention under study. Without a comparison group, secular trends may be confused with intervention effects [48]. Interrupted time-series designs, with data collected from multiple points in time before and after the intervention, can better account for secular trends.
In addition to confounding from secular trends, uncontrolled study designs are susceptible to other ‘non-interventional’ aspects of the intervention. For example, an intervention may bestow more attention on patients or clinicians through data collection, leading to self-reported improvement through placebo-like effects. Comparison groups, even without randomization, offer important protection against secular trends and placebo-like effects. Non-randomized allocation to intervention and comparison groups does not assure that both groups are similar in all important characteristics. Matched study designs may balance study groups for a limited number of measured characteristics. In contrast, successfully implemented randomization equalizes recognized and unrecognized confounders across study groups and is, therefore, essential for cause-and-effect inference.
In summary, limitations of study designs without randomization or a comparison group include difficulty establishing causality, confounding, bias, and spurious associations from multiple comparisons [29]. Although such studies are generally considered to be lower within the evidence hierarchy, they may provide useful information when randomized controlled trials (RCTs) are not feasible or generate important hypotheses for subsequent testing with more rigorous study designs. We focus the remainder of this chapter on key RCTs for implementation research, in particular the cluster RCT – where clusters of individuals (groups) are randomized [98] rather than individuals. Due to the complex design, we strongly recommend that investigators obtain expert consultation with methodologists and statisticians early during the planning stages.
Other study designs applicable for implementation research and quality improvement projects are reviewed elsewhere [99, 100]. Proctor et al. reviews funding mechanisms and provides key ingredients for implementation research proposals (Table 13.2) [33]. Competencies for trainees of implementation and dissemination research are also available elsewhere [101].
Implementation Randomized Controlled Trials
Many principles for the design of high-quality, traditional RCTs discussed elsewhere in this book also apply to implementation research. As a discussion guide, our approach parallels the Consolidated Standards of Reporting Trials (CONSORT), which were designed to encourage high-quality clinical randomized trials and promote a uniform reporting style. The CONSORT criteria emphasize the ability to understand the flow of all actual and potential research participants through the experimental design. Although originally designed for the traditional or ‘parallel’ clinical trial [102, 103], the CONSORT criteria were subsequently modified for the cluster RCT [104, 105].
We refer the reader to specific example of an implementation randomized trial illustrating the formative development of one of the outcomes [106], challenges and barriers with recruitment [107], main outcome, [108] and secondary outcomes [109] (ClinicalTrials.gov Identifier: NCT00403091; Available at: http://clinicaltrials.gov).
Participants and Recruitment
In contrast to the randomized clinical trial where patients are the unit of intervention and analysis, implementation randomized trials and interventions have a broader reach. For example, key participants in implementation RCTs may be doctors, patients, clinics, or hospitals, or hospital wards. Because implementation research is conducted in the ‘real world’ and often seeks to engage busy clinicians, systems, and patients who are otherwise overwhelmed with their usual activities, recruitment may be particularly difficult. Therefore, recruitment protocols for implementation research demand careful consideration and may require a dedicated recruitment and retention team that is specific to the target population. Often multiple approaches (e.g., word of mouth, e-mail, phone, fax, personal contacts, or lists from professional organizations) must be pursued, and still the desired number of participants may not be reached. This is a particular challenge as it pertains to recruiting individuals and engaging systems that cater to marginalized minority populations.
African Americans and Latinos continue to bear an unequal burden of disease. Individuals from these populations are underrepresented in implementation research. To reach wide applicability, a diverse pool of participants in research studies is necessary. However, racial and ethnic minorities remain underrepresented in research participation. For example, less than one-third of those enrolled in research studies sponsored by the National Institutes of Health (NIH) are minorities [110, 111],– African Americans comprising 12.6 % and Latinos 7.5 % [111].
Minorities have often been underrepresented in traditional clinical research studies for several reasons. Researchers and participants often do not share common cultural perspectives, which may lead to lack of trust. Moreover, limited resources, such as low levels of income, education, health insurance, social integration, and health literacy, may also preclude participation in research studies studies [112]. In addition, the history of racism in the U.S. and particularly in medical research and clinical care, has contributed to deep suspicion among minority communities about the motives of the medical system [113–115].
Low research participation from communities of color stems directly from these historical inequities and power imbalances that have created a lack of trust between community and academic medical institutions.
Within the past two decades, a series of nationwide mandates for federally funded research have been created in order to directly address the concerns of distrust in these populations including: the NIH Revitalization Act created in 1993 and updated in 2001 mandating the inclusion of women and minorities in clinical trials [41, 116], the 1997 Federal and Drug Administration (FDA) Modernization Act providing strict requirements on the standardization of data collection on racial/ethnic minority groups in clinical trials [116], and the Centers for Medicare and Medicaid Services (CMS) authorization of routine care costs for Medicare beneficiaries who are participants in clinical trials in 2000 [116].
Despite these mandates, challenges in the recruitment of minorities still exist. Chapter 8 of this textbook offers additional insight on broad recruitment strategies for implementation research. Table 13.3 offers some solutions to these commonly faced barriers [117].
Human Subjects
Review and approval of implementation studies by an institutional review board (IRB) is necessary. Often, the research protocol may pose minimal danger to participants and the review may be conducted under an expedited protocol. We refer the reader to more detailed reviews on this topic [118–120]. Randomization, intent to publish study findings, or present at scientific conferences places the work in the research domain. Although usual local quality improvement activities, which are important in health care, do not require IRB approval, the addition of a rigorous design for implementation research does require review.
Investigators designing cluster RCTs must carefully consider the ethical issues that arise when consent occurs at the cluster level with subsequent enrollment of participants within the cluster. If the target of the research is clearly the clinician, informed consent may often be waived for the patient. For studies that focus on the clinician but collect outcomes from medical record review or administrative patient records, the researchers may consider applying for a waiver of informed patient consent. Such waivers are especially reasonable when a large volume of patient records would make patient informed consent impractical. Implementation research usually generates personally identifiable health information, which may be subject to the Health Insurance Portability and Accountability Act (HIPAA). Waiver of HIPAA consent by the patient may often be obtained based on requirements similar to waiver of informed patient consent. Finally, it may be necessary to obtain consent from both patients and providers if the intervention targets both populations.
Investigators should develop detailed plans to protect the security and confidentiality of study data. Data should be housed in physically secured locations with strong logical protection, such as password protection and encrypted files. Access to study data should be only on a ‘need-to-know’ basis. Participant identifiers should be maintained only as necessary for data quality control and linkage. Patients and clinicians should be assured that personal information will not be revealed in publications or presentations. Data integrity should also be protected with detailed protocols for verification and cleaning, which are beyond the scope of this chapter [121].
We agree with the International Committee of Medical Journal Editors (ICMJE) that descriptions of all randomized clinical trials should be deposited in publically available registries before recruitment begins [122]. The ICJME includes interventions focusing on process-of-care within the rubric of clinical trials. Trial registries guard against the well-recognized bias that negative studies are less likely to be published than positive studies. Negative publication bias may significantly limit meta-analytic studies, leading to the false conclusion that ineffective interventions are actually effective. Registries also increase the likelihood that participation in clinical trials will promote the public good, even if the study is negative. Although the template is not customized for implementation research, one such registry may be found at http://clinicaltrials.gov.
Intervention Design
Based on the concepts described earlier in this chapter, the design of the intervention is often guided using a formative-evaluation process [123, 124]. Formative evaluation incorporates input from end users and stakeholders to refine an intervention during the early stages of development. It is critical that investigators carefully explore and understand the need of those who will be affected by the intervention. For example, the design can be guided by focus groups or nominal group technique [125–128]. Glasgow et al. [37] recommend key features to include in the content design:
-
barrier analysis
-
integration of multiple types of evidence
-
adoption of practical trials that address clinician concerns
-
investigation of multiple outcomes, generalizability, and contextual factors
-
design of multilevel programs using systems and social networking models mindful of the integration of the study’s components and levels, and
-
adaptation of program to local needs and ongoing issues.
For example, for an internet-delivered intervention for physicians important features to consider include [129]:
-
needs assessment from office practice data
-
multimodal strategies
-
modular design with multiple parts
-
clinical cases for contextual learning
-
tailoring intervention based on individual responses
-
interactivity with the learner
-
audit and feedback
-
evidence-based content
-
established credibility of organization providing website and funding entity
-
patient education resources
-
high level of usability, and
-
accessibility to the Internet site despite limited bandwidth.
Comparison Group
In behavioral research, it is often appropriate to randomize participants to either an active intervention versus an attention control. The attention control – in contrast to ‘placebo’ or no intervention- accounts for changes in behavior attributable to social exposure when participants receive services and attention from study personnel [130]. Positive social interactions may create expectations for positive outcomes, potentially confounding intervention effects collected through such methods as self-report. However, the precise implementation of attention controls may be difficult [131].
In our experience, clinicians and communities may be reluctant to enter a study with the possibility of being randomized to a group with no apparent benefit. This problem may be compounded by intensive procedures needed for data collection, regardless of the study group. To overcome such barriers, investigators may offer to open the intervention to the comparison group at the close of the study. Alternatively, study design might more formally incorporate a delayed intervention or test two variations of an active intervention.
Blinding
‘Blinding’ is important to decrease bias in outcome ascertainment (similar to randomized clinical trials). Study personnel who perform outcome assessment should be unaware of whether an individual participant has been assigned to the intervention or comparison group. For example, it may be necessary to blind those doing patient examinations, those performing medical record abstraction, or those administering patient, physician, or organizational surveys. When participants are blinded to the allocation arm, the study is single-blinded. If those delivering the intervention and collecting the outcomes are blinded as well, then the study is double-blinded. If the analysts are unaware of the assignments, then the study is triple-blinded. For implementation research, it is often not feasible to conceal study allocation from the research team.
Units of Intervention, Randomization, and Analysis
Investigators planning an implementation randomized trial must carefully consider the units of study assignment for intervention, randomization, and analysis. Examples of units of intervention are patients, physicians, nurses, clinics, hospitals, hospital wards, among others. Within any given study, the unit level may vary across components, meaning that the analysis plan must account for the clustered nature of the outcome data.
For example, consider a study of a patient-based intervention that will be implemented through a group of affiliated multi-physician clinics. ‘Contamination’ could arise from physicians learning about the intervention and then exposing comparison patients to part of the intervention. Therefore, for this particular study, the investigators may choose to randomize at the physician level to avoid contamination. Thus, all patients assigned to a given physician will be allocated to the same condition: intervention or comparison.
In practice, the threat of contamination may be more perceived than real, depending upon the exact nature of the intervention and study setting. When present, contamination decreases the precision with which the intervention effect will be measured and increases the risk of a Type II error. As an alternative to cluster-based randomization to overcome contamination, the sample size could be increased [122].
Measurement and Outcomes
In implementation research, the science of determining an approach to define the measures to obtain and the specific outcomes is rapidly evolving [132]. Concepts of treatment integrity utilized in the traditional randomized controlled trial also apply to implementation research, also known as treatment fidelity. In addition, concepts are also applicable for the assessment of external validity – the applicability of the findings in other settings.
The use of a systematic strategy allows the implementation researcher to plan ahead and define the measures and relevant outcomes. The ultimate goal is to have a strong foundation for formative and summative evaluations utilizing quantitative and qualitative methods. Similar in importance as knowing whether an intervention worked (or did not) is to understanding ‘how’ the intervention worked (or did not).
Research that uses a mixed methods (or multimethods) approach is suitable to understand problems from multiple perspectives and contextualize information [133]. Mixed methods research is defined as “the type of research in which a researcher or team of researchers combines elements of qualitative and quantitative research approaches (e.g., use of qualitative and quantitative viewpoints, data collection, analysis, inference techniques) for the broad purposes of breadth and depth of understanding and corroboration” [134].
Strategies are available to design, evaluate, and report implementation research studies. A systematic review and a book are available elsewhere; [135, 136] in this chapter, we briefly review the following:
-
Reach, Effectiveness, Adoption, Implementation, and Maintenance (RE-AIM)
-
Pragmatic-Explanatory Continuum Indicator Summary (PRECIS)
-
Predisposing, Reinforcing, and Enabling Constructs in Ecosystem Diagnosis and Evaluation (PRECEDE)-Policy, Resourcing, and Organization for Educational and Environmental Development (PROCEDE).
-
Realist Evaluation
The ‘RE-AIM’ approach is particularly helpful to evaluate the public health impact of interventions [37, 137–142]:
-
Reach – the intended target population, study’s reach and representativeness, participants and setting – ‘How many participate?’
-
Effectiveness – the magnitude of intervention effect, adverse outcomes, and costs – ‘Does it work in usual settings?’
-
Adoption – use by the target audience – ‘How many use it?’
-
Implementation – the consistency of use, costs, and adaptations made during delivery – ‘Is it used as intended?’
-
Maintenance – the intervention’s long-term effects, sustainability, and attrition rates - ‘Is it sustained over time?’
The Pragmatic-Explanatory Continuum Indicator Summary (PRECIS) originated from a Canadian and European initiative to promote trials in developing and middle-income countries [143]. Within this context, a pragmatic trial seeks to answer the question, “Does an intervention work in usual settings under usual conditions?;” a pragmatic trial tests the effectiveness of the intervention and informs decision makers [144]. An explanatory trial seeks to answer the question, “Does an intervention work in research settings?;” an explanatory trial tests the efficacy of an intervention [145, 146].
The visual representation of measures among the ten domains described in PRECIS allows scientists and community in general understand the applicability of the interventions. The ten domains include: (1) participant eligibility criteria, (2) experimental intervention flexibility, (3) practitioner expertise (experimental), (4) comparison intervention, (5) practitioner expertise (comparison), (6) follow-up intensity, (7) primary trial outcome, (8) participant compliance, (9) practitioner adherence, and (10) analysis of primary outcome. For a more detailed discussion, the reader is referred elsewhere [143, 147].
First proposed in 1974, the PRECEDE-PROCEED is an approach to assess the effects of health programs and health education applicable for implementation research [148–150]. The realist approach offers a quantitative and qualitative model for synthesis of the effects of complex programs that are intimately related to the contextual factors where the program is developed and evaluated [151–153]. A realist approach addresses some of the short-comings of purely quantitative methods for program evaluation [152].
Reviews are available to guide the design, measurement, and reporting of implementation research studies [132, 142, 144, 154]. The addition of information about the context, protocol implementation, and generalizability – among other characteristics – are enhancements to the CONSORT reporting guidelines for the traditional efficacy study [142, 154].
Approaches to Randomization
Randomization, also described elsewhere in this book, is a procedure to assure that study units are allocated to the study conditions according to chance alone. The specific approach to randomization is described as ‘sequence generation’ and may include matching or stratification [104]. Allocation concealment is a ‘technique used to prevent selection bias by concealing the allocation sequence from those assigning participants to intervention groups, until the moment of assignment, the purpose of which is to prevent researchers from influencing which participants are assigned to a given group [102, 103]. The concealment may be simply based on a coded list of randomly ordered study groups created by a statistician who is not a member of the intervention team. After enrollment, each participant is assigned to a study group based on the sequence in the list.
For cluster-randomized trials, the assignment of individuals to a study group is determined at the level of the cluster, which increases the opportunity for selection bias from failed concealment. For example, consider a cluster RCT where randomization occurs at the physician level with subsequent enrollment of patients with diabetes from the physicians’ practice. Depending upon the nature of the intervention, physicians may be able to determine their randomization group. If the randomized physician also recruits patients for the study, this knowledge of the randomization group may lead to biased patient selection. An ‘attention control’ comparison group described above would also decrease the chances of the physician to discover the assignment.
Successful randomization ensures balanced characteristics at the unit of randomization, and larger numbers of randomized units increase the chance of successful randomization. Investigators should be aware that for cluster RCTs, successful randomization does not ensure balanced characteristics at units below the level of randomization [155]. Again, consider the illustration above where randomization occurs at the physician level. Although this design may produce intervention and comparison groups that are balanced based on physician characteristics, there may be important imbalances in patient characteristics, decreasing the power of randomization. To guard against imbalances of lower-level units in cluster randomized trials, investigators might consider stratifying or matching on a limited number of critical characteristics [156]. Alternatively, imbalances may require statistical adjustment at the point of analysis after the study has been completed. Decisions about matched study designs for cluster randomized trials are complex and beyond the scope of this chapter.
Intent-to-Treat
As with the traditional clinical randomized trial, the primary analysis for an implementation randomized trial should test hypotheses specified a priori and should follow intent-to-treat principles [157]. With the intent-to-treat approach, all units are analyzed with the group to which they were originally randomized, regardless of whether the units are subsequently exposed to the intervention (i.e., cross over). For example, in a randomized trial of an Internet-based continuing medical education (CME) intervention for physicians, outcomes for all physicians randomized to the intervention group must be analyzed as part of the intervention group, regardless of whether the physician visited the Internet site. Intent-to-treat protocols preserve the power of randomization by protecting against bias resulting from differential participation or cross-over among intervention units with a greater or lesser propensity for success.
Unfortunately, participants lost to follow up may generate no data for analysis. As with violation of the intent-to-treat principle, loss to follow up may reduce the power of randomization. Although complete follow up is desirable, it is usually not obtainable. Many scientists hold that for clinical trials, loss to follow up of greater than 20 % introduces severe potential for bias [158]. Therefore, many study designs include run-in phases before randomization. From the perspective of internal validity, it is better to exclude participants before randomization than have participants lost to follow up, cross between study groups, or become non-adherent to intervention protocols after randomization. For example, in the study of Internet-based CME described above, physicians might be required to demonstrate a willingness to engage in Internet learning and submit data for study evaluation before randomization. According to the CONSORT criteria for group randomized trials, investigators must carefully account for all individuals and clusters that were screened or randomized [104].
Retention, Special Populations
Retention of research participants is a challenging issue in research, and can be of particular concern when working with vulnerable and underserved populations. The broad goal of implementation science is to translate evidence-based practice into real world application. Specific to this goal, there is an overarching need to target and tailor implementation to the specific system, providers, and patients that exist in a given community. Methods that to date have been typically implemented at the system and provider level within academic health settings, and included predominately homogenous white patient participation, will not translate well in more diverse, community driven settings.
Lack of retention and loss to follow up can be a barrier in implementation research, especially for projects concerning health disparities in minorities. In the recently conducted Healthy Aging in Neighborhoods of Diversity Across the Life Span (HANDLS) study, Ejogu et al. present a multifaceted approach to recruitment and specifically retention strategies for minority and low socioeconomic status (SES) participants [159]. In this 20-year longitudinal examination of how race and SES influence the development of age-related health disparities, the investigators created a multifactorial recruitment and retention strategy that targeted known barriers and identified those unique to the study’s urban environment [159]. Through this approach, they were able to recruit over 3,700 participants, of whom 59 % were African American with a 75 % baseline completion [159]. The success of the HANDLS investigation relied primarily on the emphasis of the community-based platform to alleviate many of the barriers that might exclude this key population from participation and retention.
Underrepresented minorities are a special population whose participation in implementation research holds promise in revealing methods to reduce health disparities [160–162]. It may be difficult to ascertain the true population benefit or effectiveness of an intervention if a significant proportion of its participants are lost to follow-up. While Chap. 8 of this textbook is solely dedicated to addressing broad strategies for recruitment and retention, this section offers specific insight into retention issues pertaining to the participation of minorities. As an overarching challenge between the health care system and minority communities, the establishment of trust continuously strikes a chord as a key necessity in retaining the attention and participation of this population in implementation research [160, 163]. Yancey et al. suggest some targeted approaches to decreasing participant loss specifically in underrepresented and minority groups [160–162, 164–166]:
-
Intensive follow-up and contact with subjects
-
Retain interviewers, field staff, and study staff over time
-
Involve staff from the targeted community
-
Provide social support and offering accessible locations for study visits and/or data collection
-
Ensure timely incentive payments and accessibility of project staff
-
Encourage study staff’s knowledge of community dynamics and project leadership/staff visibility and involvement in the community.
Statistical Analysis
Statistical analysis for cluster RCTs is a vast, technical topic that falls largely beyond the domain of the basic introduction provided in this book. However, an example will illustrate some important principles. More specifically, consider the previous illustration in which physicians are randomized to an intervention or comparison group, with patients being subsequently enrolled and assigned to the same study condition as their physician. To conduct the analysis at the physician level, the investigators might simply compare the mean post-intervention outcomes for the two study groups. However, this approach leads to loss of statistical power, because the number of physicians randomized will be less than the number of patients included in the study. Alternatively, the investigators could plan a patient-level analysis that appropriately considers the clustering of patients within physicians. The investigators could also collect outcomes for intervention and comparison patients before and after intervention implementation. Generalized estimation equations could then be used to compare the change in study endpoints over time for the intervention versus comparison group. Here, the main study effect will be reflected by a group-time interaction variable included in the multivariable model. This approach uses a marginal, population-averaged model to account for clustered observations and potentially adjust for observed imbalances in the study groups. Alternatively, the analyst may use a cluster-specific (or conditional) approach that directly incorporates random effects. Murray reviewed the evolving science and controversies surrounding the analysis of group-randomized trials [156].
Although the main analysis should follow intent-to-treat principles as described above, most implementation randomized trials include a range of secondary analyses. Such secondary analyses may yield important findings, but they do not carry the power of cause-and-effect inference. ‘Per-protocol’ or ‘compliers only’ analyses may address the impact of the intervention among those who are sufficiently exposed or may examine dose-response relationships between intervention exposure and outcomes. Mediation analysis using a series of staged regression models may investigate mechanisms through which an intervention leads to a positive study effect [167, 168].
Sample Size Calculations
The investigator must determine the number of participants necessary to detect a meaningful difference in study endpoints between the intervention and comparison groups, i.e., the power of the study. Typically, a power of 80 % is considered adequate to decrease the likelihood of a false negative result. If an intervention is sustained over an extended period of time, the investigators may wish to test specifically for effect decay, perhaps with a time-trend analysis. Such a hypothesis of no difference demands a special approach to power calculation. Sample size calculations for traditional randomized trials are discussed elsewhere in this book (see Chap. 15).
The analysis for an implementation randomized trial may be at a lower level than the unit of randomization. Under these circumstances, the power calculations must account for the clustering of participants within upper-level units, such as the clustering of patients within physicians from the example above. Failure to account for the hierarchical data structure may inflate the observed statistical significance and increase the likelihood of a false positive finding [169].
Several approaches to accounting for the clustering of, say, patients within physicians from the above example, rely on the intra-class correlation coefficient (ICC). The ICC is the ratio of the between-cluster variance to the total sample variance (between clusters + within cluster). In this example, the ICC would be a measure of how ‘alike’ patient outcomes were within the physician clusters. If the ICC is 1, the outcomes for all patients clustered within a given physician are identical. If the ICC is 0, clustering within physicians is not related to patient outcomes [170]. In other words, with an ICC of 1, adding additional patients provides no additional information. Therefore, as the ICC increases, one must increase the sample size to retain the same power. For 0 < ICC < 1, increasing the number of patients will increase study power less than increasing the number of physicians. Typical values for ICCs range from 0.01 to 0.50 [171].
Although the topic of power calculations for group randomized trials is vast and largely beyond the scope of this book, Donner provides a straight-forward framework for simple situations [169]. Taking this approach, the analyst first calculates an unadjusted sample size (Nun) using approaches identical to those described elsewhere in this book for the traditional randomized clinical trial. Next, the analyst calculates a sample inflation factor (IF) that is used to derive a cluster-adjusted sample size (Nadj). Then:
where m is the number of study units per cluster, and ρ is the ICC.
Situational Analysis and External Validity
Because implementation randomized trials occur in a ‘real-word’ setting, we place special emphasis on understanding and reporting of context. In contrast to the traditional randomized clinical trial, the study setting for the implementation trial is an integral part of the study design. To address the importance of context in implementation research, Davidoff and Batalden promote the concept of situational analysis for quality improvement studies [81]. We believe that many of these principles are relevant to the implementation randomized trial. For example, published reports for implementation research should include specific details about the clinic setting, patient population, prior experience with system change, and how the context contributed to understanding the problem for which the study was designed. In addition, specialized approaches to economic evaluation provide additional important context for interpreting the results from implementation trials [172].
Because implementation research often focuses on dissemination to large populations, external validity, or generalizability, acquires special importance. One must consider how study findings are applicable to other patients, doctors, clinics, or geographic locations.
Summary
Implementation research bridges the gap between scientific knowledge and its application to daily practice with the overall purpose of improving the health of individuals and populations. To advance the science of implementation research, the Institute of Medicine published findings from the Forum on the Science of Health Care Quality Improvement and Implementation in 2007 [173] and the Veterans’ Health Administration sponsored a state-of-the-art (SOTA) conference in 2004 [3]. Together, these documents summarized current knowledge, identified barriers to implementation research, and defined strategies to overcome these barriers. Given the well-documented quality and safety problems of our health care system despite the vast resources invested in the biomedical sciences, we need to promote interest in implementation research, an emerging scientific discipline focused on improving health care for all, regardless of geography, socioeconomic status, race, or ethnicity.
References
Berwick DM. Disseminating innovations in health care. JAMA. 2003;289:1969–75.
Lenfant C. Shattuck lecture – clinical research to clinical practice – lost in translation? N Engl J Med. 2003;349:868–74.
Kiefe CI, Sales A. A state-of-the-art conference on implementing evidence in health care. Reasons and recommendations. J Gen Intern Med. 2006;21 Suppl 2:S67–70.
Ross JS, Tse T, Zarin DA, Xu H, Zhou L, Krumholz HM. Publication of NIH funded trials registered in ClinicalTrials.gov: cross sectional analysis. BMJ. 2012;344:d7292. doi:10.1136/bmj.d7292.
McGlynn EA, Asch SM, Adams J, et al. The quality of health care delivered to adults in the United States. N Engl J Med. 2003;348:2635–45.
National Institute of Health (NIH). (2013, January 9). Dissemination and implementation research in Health (R01). Available at: http://grants.nih.gov/grants/guide/pa-files/PAR-13-055.html. Accessed 9 May 2013.
Kiefe CI, Safford M, Allison JJ. Forces influencing the care of complex patients: a framework. In: Academy health annual meeting 2007. Orlando; 2007.
Rubenstein LV, Pugh J. Strategies for promoting organizational and practice change by advancing implementation research. J Gen Intern Med. 2006;21 Suppl 2:S58–64.
Canadian Institutes of Health Research. More about Knowledge Translation at CIHR. Canadian Institutes of Health Research; 2013. Available at: http://www.cihr-irsc.gc.ca/e/39033.html. Accessed 9 May 2013.
Menear M, Grindrod K, Clouston K, Norton P, Legare F. Advancing knowledge translation in primary care. Can Fam Physician. 2012;58:623–7, e302-7.
Eccles MP, Armstrong D, Baker R, et al. An implementation research agenda. Implement Sci. 2009;4:18.
Centers for Disease Control and Prevention (CDC). (2007, February 16). Improving public health practice through translation research (R18). Available at: http://grants.nih.gov/grants/guide/rfa-files/rfa-cd-07-005.html. Accessed 9 May 2013.
Rabin BA, Brownson RC, Haire-Joshu D, Kreuter MW, Weaver NL. A glossary for dissemination and implementation research in health. J Public Health Manag Pract. 2008;14:117–23.
Knowledge Translation (KT) Clearinghouse. Canadian Institute of Health Research (CIHR). (2013). Available at: http://ktclearinghouse.ca/knowledgebase/glossary. Accessed 9 May 2013.
Smedley BD, Stith AY, Nelson AR. Unequal treatment: confronting racial and ethnic disparities in health care. Washington, DC: The National Academies Press; 2003.
Allison JJ. Health disparity: causes, consequences, and change. Med Care Res Rev. 2007;64:5S–6.
Chin MH, Walters AE, Cook SC, Huang ES. Interventions to reduce racial and ethnic disparities in health care. Med Care Res Rev. 2007;64:7S–28.
Kilbourne AM, Switzer G, Hyman K, Crowley-Matoka M, Fine MJ. Advancing health disparities research within the health care system: a conceptual framework. Am J Public Health. 2006;96:2113–21.
Smedley BD, Stith AY, Nelson AR. Unequal treatment: confronting racial and ethnic disparities in health care. Washington, DC: Institute of Medicine; 2003.
Hayward RA, Asch SM, Hogan MM, Hofer TP, Kerr EA. Sins of omission: getting too little medical care may be the greatest threat to patient safety. J Gen Intern Med. 2005;20:686–91.
Kohn LT, Corrigan JM, Donaldson MS. To err is human: building a safer health system. Washington, DC: Institute of Medicine; 1999.
Interagency Task Force on Antimicrobial Resistance. Co-Chairs: Centers for Disease Control and Prevention, Food and Drug Administration, National Institutes of Health. Public health action plan to combat antimicrobial resistance centers for disease control and prevention; 2011. http://www.cdc.gov/drugresistance/pdf/public-health-action-plan-combat-antimicrobial-resistance.pdf. Accessed 4 April 2014.
Snow V, Mottur-Pilson C, Gonzales R. Principles of appropriate antibiotic use for treatment of acute bronchitis in adults. Ann Intern Med. 2001;134:518–20.
Wenzel RP, Fowler 3rd AA. Clinical practice. Acute bronchitis. N Engl J Med. 2006;355:2125–30.
Roumie CL, Halasa NB, Grijalva CG, et al. Trends in antibiotic prescribing for adults in the United States – 1995 to 2002. J Gen Intern Med. 2005;20:697–702.
Metlay JP, Camargo Jr CA, MacKenzie T, et al. Cluster-randomized trial to improve antibiotic use for adults with acute respiratory infections treated in emergency departments. Ann Emerg Med. 2007;50:221–30.
Committee on Quality of Health Care in America. Crossing the quality chasm: a new health system for the 21st century. Washington, DC: Institute of Medicine; 2001.
Berwick DM, Calkins DR, McCannon CJ, Hackbarth AD. The 100,000 lives campaign: setting a goal and a deadline for improving health care quality. JAMA. 2006;295:324–7.
Smith GC, Pell JP. Parachute use to prevent death and major trauma related to gravitational challenge: systematic review of randomised controlled trials. BMJ. 2003;327:1459–61.
Auerbach AD, Landefeld CS, Shojania KG. The tension between needing to improve care and knowing how to do it. N Engl J Med. 2007;357:608–13.
Brennan TA, Gawande A, Thomas E, Studdert D. Accidental deaths, saved lives, and improved quality. N Engl J Med. 2005;353:1405–9.
Berwick D. The stories beneath. Med Care. 2007;45:1123–5.
Proctor EK, Powell BJ, Baumann AA, Hamilton AM, Santens RL. Writing implementation research grant proposals: ten key ingredients. Implement Sci. 2012;7:96.
Bhattacharyya O, Reeves S, Garfinkel S, Zwarenstein M. Designing theoretically-informed implementation interventions: fine in theory, but evidence of effectiveness in practice is needed. Implement Sci. 2006;1(Feb 23):5.
Effective Health Care: Getting evidence into practice. National Health Service Center for Reviews and Dissemination, Royal Society of Medicine Press. 1999;5(1). http://www.york.ac.uk/inst/crd/ehc51.pdf. Accessed Nov 2007.
Rogers EM. Diffusion of innovations. 5th ed. New York: Free Press; 2003.
Glasgow RE, Emmons KM. How can we increase translation of research into practice? Types of evidence needed. Annu Rev Public Health. 2007;28:413–33.
Sung NS, Crowley Jr WF, Genel M, et al. Central challenges facing the national clinical research enterprise. JAMA. 2003;289:1278–87.
Nelson EC, Batalden PB, Huber TP, et al. Microsystems in health care: Part 1. Learning from high-performing front-line clinical units. Jt Comm J Qual Improv. 2002;28:472–93.
Pinsky PF, Ford M, Gamito E, et al. Enrollment of racial and ethnic minorities in the Prostate, Lung, Colorectal and Ovarian Cancer Screening Trial. J Natl Med Assoc. 2008;100:291–8.
Ford ME, Siminoff LA, Pickelsimer E, et al. Unequal burden of disease, unequal participation in clinical trials: solutions from African American and Latino community members. Health Soc Work. 2013;38:29–38.
Bleich SN, Jarlenski MP, Bell CN, LaVeist TA. Health inequalities: trends, progress, and policy. Annu Rev Public Health. 2012;33:7–40.
The Commonwealth Fund Commission on a High Performance Health System, Why not the best? Results from the National Scorecard on U.S. Health System Performance, 2011, The Commonwealth Fund, October 2011.
Census Bureau. Statistical abstract of the United States: 2012: the national data book. Washington, DC: Census Bureau; 2012.
Zerhouni EA. Medicine. The NIH roadmap. Science. 2003;302:63–72.
Zerhouni EA. US biomedical research: basic, translational, and clinical sciences. JAMA. 2005;294:1352–8.
Chao SR. The state of quality improvement and implementation research: expert views – workshop summary. Washington, DC: The National Academies Press; 2007.
Shojania KG, Grimshaw JM. Evidence-based quality improvement: the state of the science. Health Aff (Millwood). 2005;24:138–50.
Shojania KG, McDonald KM, Wachter RM, Owens DK, editors. Closing the quality gap: a critical analysis of quality improvement strategies, Series overview and methodology, vol. 1. Rockville: Agency for Healthcare Research and Quality (US); 2004. PubMed PMID: 20734525.
Williams GC, Deci EL. Activating patients for smoking cessation through physician autonomy support. Med Care. 2001;39:813–23.
Williams GC, McGregor H, Zeldman A, Freedman ZR, Deci EL, Elder D. Promoting glycemic control through diabetes self-management: evaluating a patient activation intervention. Patient Educ Couns. 2005;56:28–34.
Turner RC, Waivers LE, O’Brien K. The effect of patient-carried reminder cards on the performance of health maintenance measures. Arch Intern Med. 1990;150:645–7.
Bodenheimer T, Lorig K, Holman H, Grumbach K. Patient self-management of chronic disease in primary care. JAMA. 2002;288:2469–75.
Eng E, Parker E, Harlan C. Lay health advisor intervention strategies: a continuum from natural helping to paraprofessional helping. Health Educ Behav. 1997;24:413–7.
Cherrington A, Ayala GX, Amick H, Scarinci I, Allison J, Corbie-Smith G. Applying the community health worker model to diabetes management: using mixed methods to assess implementation and effectiveness. J Health Care Poor Underserved. 2008;19:1044–59.
Swider SM. Outcome effectiveness of community health workers: an integrative literature review. Public Health Nurs. 2002;19:11–20.
Silka L, Cleghorn GD, Grullon M, Tellez T. Creating community-based participatory research in a diverse community: a case study. J Empir Res Hum Res Ethics. 2008;3:5–16.
Lawrence Mayor’s Task Force. Tools for research partnerships in Lawrence, MA. Lawrence: Lawrence Mayor’s Task Force; 2006. Available from: http://www.tuftsctsi.org/About-Us/CTSI-Components/Community-Engagement/~/media/B35A1D1535DB422D90E1A47544743E4E.ashx. Accessed 27 June 2013.
Institute of Medicine. Clinical practice guidelines: directions for a new program. Washington, DC: National Academy Press; 1990.
Kryworuchko J, Stacey D, Bai N, Graham ID. Twelve years of clinical practice guideline development, dissemination and evaluation in Canada (1994 to 2005). Implement Sci. 2009;4:49.
Grimshaw J, Eccles M, Thomas R, et al. Toward evidence-based quality improvement. Evidence (and its limitations) of the effectiveness of guideline dissemination and implementation strategies 1966–1998. J Gen Intern Med. 2006;21 Suppl 2:S14–20.
Cabana MD, Rand CS, Powe NR, et al. Why don’t physicians follow clinical practice guidelines? A framework for improvement. JAMA. 1999;282:1458–65.
Boyd CM, Darer J, Boult C, Fried LP, Boult L, Wu AW. Clinical practice guidelines and quality of care for older patients with multiple comorbid diseases: implications for pay for performance. JAMA. 2005;294:716–24.
Marinopoulos SS, Dorman T, Ratanawongsa N, et al. Effectiveness of continuing medical education. Evid Rep Technol Assess (Full Rep). 2007;149:1–69.
Davis D, O’Brien MA, Freemantle N, Wolf FM, Mazmanian P, Taylor-Vaisey A. Impact of formal continuing medical education: do conferences, workshops, rounds, and other traditional continuing education activities change physician behavior or health care outcomes? JAMA. 1999;282:867–74.
Davis DA, Thomson MA, Oxman AD, Haynes RB. Changing physician performance. A systematic review of the effect of continuing medical education strategies. JAMA. 1995;274:700–5.
Forsetlund L, Bjorndal A, Rashidian A, et al. Continuing education meetings and workshops: effects on professional practice and health care outcomes. Cochrane Database Syst Rev. 2009;2:CD003030.
Mansouri M, Lockyer J. A meta-analysis of continuing medical education effectiveness. J Contin Educ Health Prof. 2007;27:6–15.
Grimshaw JM, Eccles MP, Walker AE, Thomas RE. Changing physicians’ behavior: what works and thoughts on getting more things to work. J Contin Educ Health Prof. 2002;22:237–43.
Mazmanian PE, Davis DA. Continuing medical education and the physician as a learner: guide to the evidence. JAMA. 2002;288:1057–60.
Centor R, Casebeer L, Klapow J. Using a combined CME course to improve physicians’ skills in eliciting patient adherence. Acad Med. 1998;73:609–10.
Shojania KG, Silver I, Levinson W. Continuing medical education and quality improvement: a match made in heaven? Ann Intern Med. 2012;156:305–8.
Cook DA, Levinson AJ, Garside S, Dupras DM, Erwin PJ, Montori VM. Internet-based learning in the health professions: a meta-analysis. JAMA. 2008;300:1181–96.
Fordis M, King JE, Ballantyne CM, et al. Comparison of the instructional efficacy of Internet-based CME with live interactive CME workshops: a randomized controlled trial. JAMA. 2005;294:1043–51.
Soumerai SB, Avorn J. Principles of educational outreach (‘academic detailing’) to improve clinical decision making. JAMA. 1990;263:549–56.
O’Brien MA, Rogers S, Jamtvedt G, et al. Educational outreach visits: effects on professional practice and health care outcomes. Cochrane Database Syst Rev. 2007;4:CD000409.
Flodgren G, Parmelli E, Doumit G, et al. Local opinion leaders: effects on professional practice and health care outcomes. Cochrane Database Syst Rev. 2011;8:CD000125.
Ivers N, Jamtvedt G, Flottorp S, et al. Audit and feedback: effects on professional practice and healthcare outcomes. Cochrane Database Syst Rev. 2012;6:CD000259.
Valente TW, Pumpuang P. Identifying opinion leaders to promote behavior change. Health Educ Behav. 2007;34:881–96.
Kiefe CI, Allison JJ, Williams OD, Person SD, Weaver MT, Weissman NW. Improving quality improvement using achievable benchmarks for physician feedback: a randomized controlled trial. JAMA. 2001;285:2871–9.
Davidoff F, Batalden P. Toward stronger evidence on quality improvement. Draft publication guidelines: the beginning of a consensus project. Qual Saf Health Care. 2005;14:319–25.
Jha AK, Perlin JB, Kizer KW, Dudley RA. Effect of the transformation of the Veterans Affairs Health Care System on the quality of care. N Engl J Med. 2003;348:2218–27.
Centers for Medicare & Medicaid Services. About the CMS Innovation Center. Available at: http://innovation.cms.gov/About/index.html. Accessed 26 June 2013.
Payne TH. Computer decision support systems. Chest. 2000;118:47S–52.
Walton RT, Harvey E, Dovey S, Freemantle N. Computerised advice on drug dosage to improve prescribing practice. Cochrane Database Syst Rev. 2001;1:CD002894.
Han YY, Carcillo JA, Venkataraman ST, et al. Unexpected increased mortality after implementation of a commercially sold computerized physician order entry system. Pediatrics. 2005;116:1506–12.
Nebeker JR, Hoffman JM, Weir CR, Bennett CL, Hurdle JF. High rates of adverse drug events in a highly computerized hospital. Arch Intern Med. 2005;165:1111–6.
Scalise D. Technology. CPOE: are you really ready? Hosp Health Netw. 2006;80:14, 6.
Ash JS, Sittig DF, Poon EG, Guappone K, Campbell E, Dykstra RH. The extent and importance of unintended consequences related to computerized provider order entry. J Am Med Inform Assoc. 2007;14:415–23.
Reckmann MH, Westbrook JI, Koh Y, Lo C, Day RO. Does computerized provider order entry reduce prescribing errors for hospital inpatients? A systematic review. J Am Med Inform Assoc. 2009;16:613–23.
Fung CH, Lim YW, Mattke S, Damberg C, Shekelle PG. Systematic review: the evidence that publishing patient care performance data improves quality of care. Ann Intern Med. 2008;148:111–23.
Werner RM, Asch DA, Polsky D. Racial profiling: the unintended consequences of coronary artery bypass graft report cards. Circulation. 2005;111:1257–63.
Centers for Medicare & Medicaid Services. Quality initiatives – general information. Available at: http://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/QualityInitiativesGenInfo/index.html. Accessed 26 June 2013.
Houle SK, McAlister FA, Jackevicius CA, Chuck AW, Tsuyuki RT. Does performance-based remuneration for individual health care practitioners affect patient care?: a systematic review. Ann Intern Med. 2012;157:889–99.
Rosenthal MB, Frank RG, Li Z, Epstein AM. Early experience with pay-for-performance: from concept to practice. JAMA. 2005;294:1788–93.
Glickman SW, Ou FS, DeLong ER, et al. Pay for performance, quality of care, and outcomes in acute myocardial infarction. JAMA. 2007;297:2373–80.
Lindenauer PK, Remus D, Roman S, et al. Public reporting and pay for performance in hospital quality improvement. N Engl J Med. 2007;356:486–96.
Murray DM. Design and analysis of group-randomized trials. New York: Oxford University Press; 1998.
Cable G. Enhancing causal interpretations of quality improvement interventions. Qual Health Care. 2001;10:179–86.
Eccles M, Grimshaw J, Campbell M, Ramsay C. Research designs for studies evaluating the effectiveness of change and improvement strategies. Qual Saf Health Care. 2003;12:47–52.
Gonzales R, Handley MA, Ackerman S, O’Sullivan PS. A framework for training health professionals in implementation and dissemination science. Acad Med. 2012;87:271–8.
Begg C, Cho M, Eastwood S, et al. Improving the quality of reporting of randomized controlled trials. The CONSORT statement. JAMA. 1996;276:637–9.
Moher D, Schulz KF, Altman D. The CONSORT statement: revised recommendations for improving the quality of reports of parallel-group randomized trials. JAMA. 2001;285:1987–91.
Campbell MK, Elbourne DR, Altman DG. CONSORT statement: extension to cluster randomised trials. BMJ. 2004;328:702–8.
Elbourne DR, Campbell MK. Extending the CONSORT statement to cluster randomized trials: for discussion. Stat Med. 2001;20:489–96.
Safford MM, Shewchuk R, Qu H, et al. Reasons for not intensifying medications: differentiating “clinical inertia” from appropriate care. J Gen Intern Med. 2007;22:1648–55.
Foster PP, Williams JH, Estrada CA, et al. Recruitment of rural physicians in a diabetes internet intervention study: overcoming challenges and barriers. J Natl Med Assoc. 2010;102:101–7.
Estrada CA, Safford MM, Salanitro AH, et al. A web-based diabetes intervention for physician: a cluster-randomized effectiveness trial. Int J Qual Health Care. 2011;23:682–9.
Billue KL, Safford MM, Salanitro AH, et al. Medication intensification in diabetes in rural primary care: a cluster-randomised effectiveness trial. BMJ Open. 2012;2:e000959.
Fouad MN, Partridge E, Green BL, et al. Minority recruitment in clinical trials: a conference at Tuskegee, researchers and the community. Ann Epidemiol. 2000;10:S35–40.
Department of Health and Human Services National Institutes of Health. “Monitoring adherence to the NIH policy on the inclusion of women and minorities as subjects in clinical research” comprehensive report: tracking of human subjects research as reported in fiscal year 2008 and fiscal year 2009. http://orwh.od.nih.gov/research/inclusion/pdf/Inclusion-ComprehensiveReport-FY-2008-2009.pdf. Accessed 13 May 2013.
Flaskerud JH, Nyamathi AM. Attaining gender and ethnic diversity in health intervention research: cultural responsiveness versus resource provision. ANS Adv Nurs Sci. 2000;22:1–15.
LaVeist TA, Nickerson KJ, Bowie JV. Attitudes about racism, medical mistrust, and satisfaction with care among African American and white cardiac patients. Med Care Res Rev. 2000;57 Suppl 1:146–61.
Randall VR. Slavery, segregation and racism: trusting the health care system ain’t always easy! An African American perspective on bioethics. St Louis Univ Public Law Rev. 1996;15:191–235.
Charatz-Litt C. A chronicle of racism: the effects of the white medical community on black health. J Natl Med Assoc. 1992;84:717–25.
Fouad MN. Enrollment of minorities in clinical trials: did we overcome the barriers? Contemp Clin Trials. 2009;30:103–4.
Wendler D, Kington R, Madans J, et al. Are racial and ethnic minorities less willing to participate in health research? PLoS Med. 2006;3:e19.
Casarett D, Karlawish JH, Sugarman J. Determining when quality improvement initiatives should be considered research: proposed criteria and potential implications. JAMA. 2000;283:2275–80.
Emanuel EJ, Wendler D, Grady C. What makes clinical research ethical? JAMA. 2000;283:2701–11.
Lynn J, Baily MA, Bottrell M, et al. The ethics of using quality improvement methods in health care. Ann Intern Med. 2007;146:666–73.
Van den Broeck J, Cunningham SA, Eeckels R, Herbst K. Data cleaning: detecting, diagnosing, and editing data abnormalities. PLoS Med. 2005;2:e267.
Torgerson DJ. Contamination in trials: is cluster randomisation the answer? BMJ. 2001;322:355–7.
Scriven M. Beyond formative and summative evaluation. In: McLaughlin MW, Phillips DC, editors. Evaluation and education: 90th yearbook of the National Society for the Study of Education. Chicago: University of Chicago Press; 1991. p. 18–64.
Weston CB, McAlpine L, Bordonaro T. A model for understanding formative evaluation in instructional design. Educ Technol Res Dev. 1995;43:29–49.
Delbecq AL, Van de Ven AH, Gustafson DH. Group techniques for program planning: a guide to nominal group and Delphi processes. Glenview: Scott Foresman; 1975.
Krueger RA, Casey MA. Focus groups: a practical guide for applied research. 3rd ed. Thousand Oaks: Sage Publications; 2000.
Nielsen J, Mack R. Usability inspection methods. New York: Wiley; 1994.
Strauss A, Corbin J. Basics of qualitative research: grounded theory, procedures, and techniques. Newbury Park: Sage Publications; 1990.
Casebeer LL, Strasser SM, Spettell CM, et al. Designing tailored Web-based instruction to improve practicing physicians’ preventive practices. J Med Internet Res. 2003;5:e20.
Bootzin RR. The role of expectancy in behavior change. In: White L, Turskey B, Schwartz G, editors. Placebo: theory, research, and mechanisms. New York: Guilford Press; 1985. p. 196–210.
Gross D. On the merits of attention-control groups. Res Nurs Health. 2005;28:93–4.
Proctor E, Silmere H, Raghavan R, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38:65–76.
Creswell JW, Klassen AC, Plano Clark VL, Smith KC for the Office of Behavioral and Social Sciences Research. Best practices for mixed methods research in the health sciences. August 2011. National Institutes of Health. Available at: http://obssr.od.nih.gov/mixed_methods_research. Accessed 1 June 2013.
Johnson RB, Onwuegbuzie AJ, Turner LA. Toward a definition of mixed methods research. J Mix Methods Res. 2007;1:112–33.
Tabak RG, Khoong EC, Chambers DA, Brownson RC. Bridging research and practice: models for dissemination and implementation research. Am J Prev Med. 2012;43:337–50.
Brownson RC, Colditz GA, Proctor EK, editors. Dissemination and implementation research in health: translating science to practice. Oxford: Oxford Scholarship Online; Oxford University Press; 2012.
Reach Effectiveness Adoption Implementation Maintenance (RE-AIM). Virginia Tech University. Available at: http://www.re-aim.org/. Accessed 19 May 2013.
Kessler RS, Purcell EP, Glasgow RE, Klesges LM, Benkeser RM, Peek CJ. What does it mean to “employ” the RE-AIM model? Eval Health Prof. 2013;36:44–66.
Glasgow RE. Implementation science models (and related metrics) to help reduce health disparities. National Cancer Institute. Available at: http://cancercontrol.cancer.gov/IS/presentations/12-18-2012_Disparities%20Conference_Glasgow_508compliant.pdf. Accessed 19 May 2013.
Glasgow RE, Klesges LM, Dzewaltowski DA, Estabrooks PA, Vogt TM. Evaluating the impact of health promotion programs: using the RE-AIM framework to form summary measures for decision making involving complex issues. Health Educ Res. 2006;21:688–94.
Glasgow RE, Vogt TM, Boles SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89:1322–7.
Glasgow RE, Magid DJ, Beck A, Ritzwoller D, Estabrooks PA. Practical clinical trials for translating research to practice: design and measurement recommendations. Med Care. 2005;43:551–7.
Thorpe KE, Zwarenstein M, Oxman AD, et al. A pragmatic-explanatory continuum indicator summary (PRECIS): a tool to help trial designers. CMAJ. 2009;180:E47–57.
Tunis SR, Stryer DB, Clancy CM. Practical clinical trials: increasing the value of clinical research for decision making in clinical and health policy. JAMA. 2003;290:1624–32.
Schwartz D, Lellouch J. Explanatory and pragmatic attitudes in therapeutical trials. J Chronic Dis. 1967;20:637–48.
Schwartz D, Lellouch J. Explanatory and pragmatic attitudes in therapeutical trials. J Clin Epidemiol. 2009;62:499–505.
Selby P, Brosky G, Oh PI, Raymond V, Ranger S. How pragmatic or explanatory is the randomized, controlled trial? The application and enhancement of the PRECIS tool to the evaluation of a smoking cessation trial. BMC Med Res Methodol. 2012;12:101.
Green LW. What can we generalize from research on patient education and clinical health promotion to physician counseling on diet? Eur J Clin Nutr. 1999;53 Suppl 2:S9–18.
Green LW. The precede-proceed model of health program planning & evaluation. Available at: http://lgreen.net/index.html. Accessed 19 May 2013.
Bakken S, Lantigua RA, Busacca LV, Bigger JT. Barriers, enablers, and incentives for research participation: a report from the Ambulatory Care Research Network (ACRN). J Am Board Fam Med. 2009;22:436–45.
Pawson R, Greenhalgh T, Harvey G, Walshe K. Realist review–a new method of systematic review designed for complex policy interventions. J Health Serv Res Policy. 2005;10 Suppl 1:21–34.
Connelly JB. Evaluating complex public health interventions: theory, methods and scope of realist enquiry. J Eval Clin Pract. 2007;13:935–41.
Rycroft-Malone J, McCormack B, Hutchinson AM, et al. Realist synthesis: illustrating the method for implementation research. Implement Sci. 2012;7:33.
Zwarenstein M, Treweek S, Gagnier JJ, et al. Improving the reporting of pragmatic trials: an extension of the CONSORT statement. BMJ. 2008;337:a2390.
Puffer S, Torgerson D, Watson J. Evidence for risk of bias in cluster randomised trials: review of recent trials published in three general medical journals. BMJ. 2003;327:785–9.
Murray DM, Varnell SP, Blitstein JL. Design and analysis of group-randomized trials: a review of recent methodological developments. Am J Public Health. 2004;94:423–32.
Lachin JM. Statistical considerations in the intent-to-treat principle. Control Clin Trials. 2000;21:167–89.
Schulz KF, Grimes DA. Sample size slippages in randomised trials: exclusions and the lost and wayward. Lancet. 2002;359:781–5.
Ejiogu N, Norbeck JH, Mason MA, Cromwell BC, Zonderman AB, Evans MK. Recruitment and retention strategies for minority or poor clinical research participants: lessons from the Healthy Aging in Neighborhoods of Diversity across the Life Span study. Gerontologist. 2011;51 Suppl 1:S33–45.
Yancey AK, Ortega AN, Kumanyika SK. Effective recruitment and retention of minority research participants. Annu Rev Public Health. 2006;27:1–28.
Janson SL, Alioto ME, Boushey HA. Attrition and retention of ethnically diverse subjects in a multicenter randomized controlled research trial. Control Clin Trials. 2001;22:236S–43.
Warren-Findlow J, Prohaska TR, Freedman D. Challenges and opportunities in recruiting and retaining underrepresented populations into health promotion research. Gerontologist. 2003;43(Spec No 1):37–46.
Arean PA, Alvidrez J, Nery R, Estes C, Linkins K. Recruitment and retention of older minorities in mental health services research. Gerontologist. 2003;43:36–44.
Ashing-Giwa K, Ganz PA. Effect of timed incentives on subject participation in a study of long-term breast cancer survivors: are there ethnic differences? J Natl Med Assoc. 2000;92:528–32.
Gauthier MA, Clarke WP. Gaining and sustaining minority participation in longitudinal research projects. Alzheimer Dis Assoc Disord. 1999;13 Suppl 1:S29–33.
Parra-Medina D, D’Antonio A, Smith SM, Levin S, Kirkner G, Mayer-Davis E. Successful recruitment and retention strategies for a randomized weight management trial for people with diabetes living in rural, medically underserved counties of South Carolina: the POWER study. J Am Diet Assoc. 2004;104:70–5.
Baron RM, Kenny DA. The moderator-mediator variable distinction in social psychological research: conceptual, strategic, and statistical considerations. J Pers Soc Psychol. 1986;51:1173–82.
Preacher KJ, Hayes AF. SPSS and SAS procedures for estimating indirect effects in simple mediation models. Behav Res Methods Instrum Comput. 2004;36:717–31.
Donner A, Klar N. Pitfalls of and controversies in cluster randomization trials. Am J Public Health. 2004;94:416–22.
Beach ML. Primer on group randomized trials. Eff Clin Pract. 2001;4:42–3.
Campbell MK, Fayers PM, Grimshaw JM. Determinants of the intracluster correlation coefficient in cluster randomized trials: the case of implementation research. Clin Trials. 2005;2:99–107.
Sculpher M. Evaluating the cost-effectiveness of interventions designed to increase the utilization of evidence-based guidelines. Fam Pract. 2000;17 Suppl 1:S26–31.
Advancing quality improvement research: challenges and opportunities – workshop summary. Institute of Medicine. The National Academies Press; 2007. www.nap.edu/catalog/11884.html. Accessed 14 July 2013.
Acknowledgement
We thank Winter Williams, Kierstin Leslie, and Natalie Wilson for critically reviewing a prior version of this chapter.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Resources
Resources
Selected Journals That Publish Implementation Research
-
Annals of Internal Medicine
-
BMJ Quality and Safety in Health Care
-
Implementation Science
-
JAMA
-
Journal of General Internal Medicine
-
Journal of Hospital Medicine
-
Medical Care
-
Pediatrics
-
The Joint Commission Journal on Quality and Patient Safety
Selected Checklists and Reporting Guidelines
-
Standards for Quality Improvement Reporting Excellence (SQUIRE)
-
SQUIRE are guidelines for publishing quality improvement interventions.
-
The guidelines provide specific details to be addressed in each section of manuscripts that report quality improvement interventions.
-
Davidoff F, Batalden P. Toward stronger evidence on quality improvement. Draft publication guidelines: the beginning of a consensus project. Qual Saf Health Care 2005;14:319–25.
-
-
Enhancing the Quality and Transparency of Health Research (EQUATOR)
-
EQUATOR is an international initiative that seeks to improve the quality of scientific reporting.
-
This initiative includes statements about reporting for a range of experimental and observational study types, including randomized trials, group randomized trials, behavioral trials, and quality interventions. It also provides education and training on the use of reporting guidelines.
-
-
Consolidated Standards of Reporting Trials (CONSORT)
-
This initiative focuses on design and reporting standards for randomized controlled trials (RCTs) in health care.
-
Although originally designed for the traditional ‘parallel’ randomized clinical trial, the CONSORT criteria have been extended to include cluster RCTs and behavioral RCTs.
-
-
Workgroup for Intervention Development and Evaluation Research (WIDER)
-
This checklist is useful in reporting the quality of behavioral change intervention studies.
-
http://interventiondesign.co.uk/wp-content/uploads/2009/02/wider-recommendations.pdf
-
Selected Resources for Intervention Design
-
Agency for Healthcare Research and Quality (AHRQ) website for clinicians and providers
-
The Effective Health Care (EHC) Program invites clinicians to join networks that promote patient-centered outcomes research. http://www.ahrq.gov/professionals/clinicians-providers/
-
Closing the Quality Gap: Revisiting the State of the Science Quality Improvement Interventions to Address Health Disparities: http://www.ahrq.gov/legacy/clinic/tp/gapdisptp.htm
-
Morbidity & Mortality Reviews on the Web. Education site with cases, commentaries, and reviews. http://webmm.ahrq.gov.
-
-
Patient-Centered Outcomes Research Institute: http://www.pcori.org/ Patient Reported Outcomes Measurement Information System (PROMIS): a National Institutes of Health (NIH) funded system of patient–reported assessment tools to health status. http://www.nihpromis.org/default
-
The National Guideline Clearinghouse is a database of evidence-based practice guidelines available to the public. http://www.guideline.gov
-
Veterans’ Administration Quality Enhancement Research Initiative (QUERI) Implementation Guides
-
The QUERI Implementation Guide is a three-part series focusing on practical issues for designing and conducing implementation research.
-
The guide includes material on conceptual models, diagnosing performance gaps, developing interventions, evaluating implementation research, lessons learned from prior QUERI projects, tools and toolkits, as well as many resources.
-
-
Finding Answers
-
This program is sponsored by the Robert Wood Johnson Foundation to develop interventions for eliminating racial/ethnic disparities in health care.
-
The Finding Answers Intervention Research (FAIR) database includes 388 summaries of journal articles from 11 systematic reviews of interventions to decrease racial/ethnic disparities for many commonly encountered diseases, such as diabetes and hypertension. Interventions based on cultural leverage and performance-based reimbursement are also included.
-
-
National Center for Cultural Competence
-
This center is sponsored by Georgetown University and offers several implementation tools, manuscripts, and policy statements for organizations, clinicians, and consumers.
-
The Internet site has a section describing ‘promising practices’ which may be particularly useful in designing new interventions.
-
-
Clinical Microsystems
-
The Dartmouth Institute for Health Policy and Clinical Practice maintains this website that offers tools for improving clinical microsystems.
-
Most tools are generally available to the public at no cost.
-
The Clinical Microsystems Action Guide (under the materials, workbooks tab) may be particularly useful for designing new interventions.
-
-
Institute for Healthcare Improvement
-
This not-for-profit organization maintains an Internet site that contains several tools for improving the quality, safety, and efficiency of health care. Many tools are publically available at no cost.
-
White papers describing the ‘Breakthrough Series’ may be particularly useful for those developing new interventions.
-
Selected Resources for Implementing and Disseminating Quality Improvement
-
Splaine, M. E., Dolansky, M. A., Patrician, P. A., Estrada, C. A. Editors. Oakbrook: Joint Commission Resources. Practice-based Learning and Improvement: A Clinical Improvement Action Guide, 3rd Edition. 2012.
-
Authors explain proven methods for integrating the core competency of practice-based learning and improvement (PBLI) into daily clinical work. Practical tools are described for health professionals working on quality improvement.
-
-
Ogrinc GS, Headrick LA, Moore SM, Barton AJ, Dolansky MA, Madigosky WS. Oakbrook: Joint Commission Resources. Fundamentals of Health Care Improvement: A Guide to Improving Your Patients’ Care. 2nd Edition. 2012.
-
The book provides a single source for nursing students, medical students, and resident physicians to learn and practice the basics of QI.
-
-
Brownson RC, Colditz GA, Proctor EK. Dissemination and Implementation Research in Health: Translating Science to Practice. Oxford Scholarship. 2012.
-
The authors provide a comprehensive roadmap for implementation research.
-
Selected Training Programs
-
Veterans Affairs Quality Scholar Fellowship Program
-
A two-year inter-professional education program that offer scholars opportunities to become leaders by applying knowledge and methods of health care improvement to the care of veteran, innovate and continually improve health care, teach health professionals about health care improvement, perform research and develop new knowledge for the ongoing improvement of the quality and value of health care services.
-
-
Quality and Safety Education for Nurses (QSEN) Institute
-
The Institute offers comprehensive, competency based resources to empower nurses with knowledge, skills, and attitudes to improve quality and safety across the healthcare system.
-
-
Training in Dissemination and Implementation Research in Health (TIDIRH).
-
This training is sponsored by the NIH’s Office of Behavioral and Social Sciences Research and the U.S. Department of Veterans Affairs.
-
See Meissner et al. Implement Science 2013 Jan 24;8:12.
-
http://conferences.thehillgroup.com/OBSSRinstitutes/TIDIRH2013/
-
-
VA Enhancing Implementation Science in VA Cyber Seminar
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland (outside the USA) 2014
About this chapter
Cite this chapter
Mixon, A.S., Powell, L., Estrada, C.A. (2014). Implementation Research: Beyond the Traditional Randomized Controlled Trial. In: Glasser, S. (eds) Essentials of Clinical Research. Springer, Cham. https://doi.org/10.1007/978-3-319-05470-4_13
Download citation
DOI: https://doi.org/10.1007/978-3-319-05470-4_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-05469-8
Online ISBN: 978-3-319-05470-4
eBook Packages: MedicineMedicine (R0)