Hostname: page-component-586b7cd67f-dlnhk Total loading time: 0 Render date: 2024-11-28T01:19:01.252Z Has data issue: false hasContentIssue false

Leveraging implementation science to advance antibiotic stewardship practice and research

Published online by Cambridge University Press:  02 December 2021

Daniel J. Livorsi*
Affiliation:
Iowa City Veterans’ Affairs Health Care System, Iowa City, Iowa Department of Internal Medicine, University of Iowa Carver College of Medicine, Iowa City, Iowa
Mari-Lynn Drainoni
Affiliation:
Section of Infectious Diseases, Department of Medicine, Boston University School of Medicine, Boston, Massachusetts Department of Health Law, Policy and Management, Boston University School of Public Health, Boston, Massachusetts
Heather Schacht Reisinger
Affiliation:
Iowa City Veterans’ Affairs Health Care System, Iowa City, Iowa Department of Internal Medicine, University of Iowa Carver College of Medicine, Iowa City, Iowa
Neha Nanda
Affiliation:
Department of Medicine, Keck School of Medicine, University of Southern California, California
Jessina C. McGregor
Affiliation:
Department of Pharmacy Practice, College of Pharmacy, Oregon State University, Portland, Oregon
Tamar F. Barlam
Affiliation:
Section of Infectious Diseases, Department of Medicine, Boston University School of Medicine, Boston, Massachusetts
Andrew M. Morris
Affiliation:
Department of Medicine, Division of Infectious Diseases, Sinai Health, University Health Network, and University of Toronto, Toronto, Canada
Julia E. Szymczak*
Affiliation:
Department of Biostatistics, Epidemiology and Informatics, University of Pennsylvania Perelman School of Medicine, Philadelphia, Pennsylvania The Leonard Davis Institute of Health Economics, University of Pennsylvania, Philadelphia, Pennsylvania
*
Author for correspondence: Daniel J. Livorsi, E-mail: [email protected]. Or Julia E. Szymczak, E-mail: [email protected]
Author for correspondence: Daniel J. Livorsi, E-mail: [email protected]. Or Julia E. Szymczak, E-mail: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Type
SHEA White Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2021. Published by Cambridge University Press on behalf of The Society for Healthcare Epidemiology of America

Antibiotics are among the most commonly prescribed medications, and there is evidence to guide the optimal use of these agents for most situations encountered in clinical medicine, including for both treatment and prophylaxis. Nevertheless, clinicians routinely prescribe antibiotics in ways that diverge from this evidence, such as prescribing them when not indicated, for durations longer than necessary, or selecting broad-spectrum antibiotics when a narrower-spectrum agent would suffice. Reference Magill, O’Leary and Ray1,Reference Fleming-Dutra, Hersh and Shapiro2 This overuse of antibiotics contributes to the public health crisis of antibiotic resistance while exposing patients to potential antibiotic-related harms.

Antibiotic stewardship refers to the coordinated effort to improve how antibiotics are prescribed by clinicians. To achieve this goal, antibiotic stewardship leverages different practices, such as prospective audit-and-feedback, prior authorization, education, clinical decision support, and rapid diagnostics. Reference Barlam, Cosgrove and Abbo3 Implementing antibiotic stewardship can be challenging because it involves changing knowledge, deeply held attitudes, cultural norms, and the emotionally influenced behaviors of clinicians and patients toward antibiotic prescribing and use. The implementation process may also vary based on the healthcare setting, the antibiotic prescribing practice that needs to be improved, and the stakeholders involved. Even though those who participate in antibiotic stewardship know this to be true from experience, a gap exists in our understanding of the way each of these factors can be addressed to increase the likelihood of implementation success. Reference Morris, Calderwood and Fridkin4

Implementation science can help address this research gap. Implementation science is “the scientific study of methods to promote the systematic uptake of proven clinical treatments, practices, organizational and management interventions into routine practice, and hence to improve health.” 5 The field originated in response to the growing recognition of how difficult it is to translate research into routine use. Reference Nilsen, Stahl, Roback and Cairney6 Implementation science is unique from quality improvement, although the 2 fields share much in common (Table 1). In this white paper, we discuss how to apply key implementation science principles and methods to studies of antibiotic stewardship. This paper is meant to be a primer on conducting implementation research about antibiotic stewardship while demonstrating how implementation science principles can inform local antibiotic stewardship efforts.

Table 1. Comparison of Quality Improvement and Implementation Science

Note. CFIR, The Consolidated Framework for Implementation Research; RE-AIM, Reach Effectiveness Adoption Implementation Maintenance Framework.

Here, we take the reader through the steps involved in the design and conduct of an implementation research study, while introducing and defining key implementation science principles. We describe the importance of establishing the evidence–practice gap to be addressed in a study. We then introduce 2 bedrock concepts within the field of implementation science and how to think about them in relation to antibiotic stewardship: frameworks and implementation strategies. We argue that what has been historically defined as antibiotic stewardship strategies, such as prospective audit and feedback and prior authorization, can also be described as implementation strategies. We discuss important preimplementation activities, including stakeholder engagement, understanding the reasons for the evidence-practice gap and selecting implementation strategies. Finally, we describe how to evaluate the implementation process. The order of the sections follows the general order of events in planning and executing a study; however, implementation research is rarely linear. Prior decisions often need to be revisited based on new information that is gained in the course of research.

Identifying the evidence–practice gap

Implementation science is the systematic study of methods to promote the uptake of proven approaches to improve health outcomes; thus, implementation research projects must clearly identify the evidence-based clinical treatment, practice, organizational, or management intervention that is the focus of the implementation effort. Within antibiotic stewardship, the evidence-based practice will typically be some aspect of antibiotic prescribing (eg, the decision to initiate antibiotic therapy, antibiotic selection, or antibiotic duration).

Implementation science focuses on generating knowledge about how to improve health outcomes by reducing the gap between what we know works to promote health (ie, an evidence-based practice) and how we actually make it work (ie, how we deliver evidence-based practice in routine, real-world settings). Reference Brownson, Colditz and Proctor7,Reference Eccles and Mittman8 Therefore, it is critical to establish whether your project addresses a gap between evidence-based optimal practice and the current status quo. Reference Kitson and Straus9 Establishing that such an evidence gap exists requires data from either the local setting or from the extant research literature. Reference Crable, Biancarelli, Walkey, Allen, Proctor and Drainoni10

Identifying this gap in antibiotic stewardship should begin with documenting whether antibiotic prescribing in a clinical setting is consistent with the best available evidence. Although clear and reproducible definitions of unnecessary or suboptimal antibiotic prescribing are needed, a helpful framework to consider comprises the “5 D’s” of optimal antibiotic therapy: diagnosis, drug, dose, duration, and de-escalation. Reference Morris, Calderwood and Fridkin4,Reference Doron and Davidson11,Reference Joseph and Rodvold12 Gaps between what we know about how antibiotics should be used and how they are actually used can occur in any of these moments along the prescribing pathway. Reference Tamma, Miller and Cosgrove13

Data demonstrating these gaps could include (1) documentation of clinician- or site-level use of antibiotics for conditions that do not require them, (2) excessive duration of antibiotic therapy for uncomplicated conditions in which short-course therapy has been demonstrated to be safe, (3) high rates of using broad-spectrum antibiotics when narrow-spectrum drugs would be equally effective, or (4) composite proportions of suboptimal antibiotic-prescribing based on evidence-based guidelines for common clinical conditions.

Implementation strategies defined

The primary focus of implementation science is the rigorous evaluation of “implementation strategies.” An implementation strategy is a method or technique used to enhance the adoption, implementation, and sustainability of an evidence-based practice by adapting it to fit the local context. 14 Implementation strategies can take several different forms. A review by Waltz et al Reference Waltz, Powell and Matthieu15 classified 73 discrete implementation strategies into 9 broad categories (Table 2). Others have categorized implementation strategies broadly into “top down–bottom up,” “push–pull,” or “carrot–stick” approaches. Reference Proctor, Powell and McMillen16

Table 2. Implementation Strategies

a Examples of audit-and-feedback include “handshake stewardship,” a form of postprescription audit and review, and leveraging peer-to-peer comparisons to give retrospective feedback.

b An antibiogram is one example of capturing and sharing local knowledge, as it aggregates local antibiotic resistance data into a single tool.

c Institutional guidelines on antibiotic prescribing are often developed by a local, multidisciplinary group of experts, which includes the stewardship team. Antibiotic stewardship committees often draw upon a variety of disciplines to garner broad support for stewardship activities.

d Clinicians can be reminded in a number of ways, such as antibiotic time-outs or order sets that prompt clinicians to choose a guideline-concordant antibiotic regimen.

e For example, antibiotic stewardship personnel may call clinicians with the results of rapid diagnostic tests and use this conversation as an opportunity to provide real-time antibiotic treatment recommendations.

f Hospitals frequently mandate that certain antibiotics receive prior authorization before initiation.

g A hospital’s antibiotic stewardship expertise can be centralized in a stewardship team, eg, an infectious disease–trained pharmacist and physician.

h In ambulatory care, patients and families are often educated about when antibiotics are unnecessary. Shared decision making is another way to involve patients in the antibiotic-prescribing process; this has been used to decrease unnecessary antibiotic treatment for acute respiratory tract infections.

In simplistic terms, the evidence-based practice is “the thing” that is known to be effective in improving health or minimizing harm, based on prior studies. Reference Curran17 The implementation strategy is what is done “to try to help people and places ‘do the thing.’” If the evidence-based practice is some aspect of antibiotic prescribing, then the implementation strategy is what helps clinicians adopt this evidence-based practice into their routine prescribing behavior. Many commonly used and proven antibiotic stewardship practices, such as audit and feedback, education, commitment posters, electronic medical record-based prompts, or the development of local consensus guidelines, are examples of implementation strategies that can be used alone or in combination to encourage clinicians to prescribe antibiotics in an evidence-based manner.

Choosing an implementation science framework

All research on implementing evidence-based antibiotic stewardship practices should make use of an implementation science framework. Implementation frameworks are grounded in theory and are sometimes referred to as “models.” In-depth discussions of the technical differences between theories, models, and frameworks can be found elsewhere. Reference Nilsen18,Reference Damschroder19

Frameworks, which are typically summarized in a single diagram, provide a systematic approach to developing, managing, and evaluating the implementation process. These frameworks clarify assumptions about the multiple interrelated factors that influence implementation (eg, individuals, units, and organizations), and they focus attention on essential change processes. By standardizing concepts, relationships, and definitions, implementation frameworks speak a common language, and studies that use established frameworks more efficiently contribute to generalized knowledge. Reference Leeman, Birken, Powell, Rohweder and Shea20

Choosing an implementation framework should be informed by the research question, the implementation strategies that a given project plans to use, and the stage of program development and evaluation. Consulting with someone who has a working knowledge of implementation science frameworks and methods is highly recommended. In general, implementation frameworks can be categorized into 1 of 3 categories: process, determinant, or evaluation. Reference Nilsen18,Reference Damschroder19 A resource for identifying an appropriate framework can be found online at https://dissemination-implementation.org; this web tool was developed through a collaborative effort across 3 United States universities. Reference Tabak, Khoong, Chambers and Brownson21

Process frameworks provide a how-to guide on planning and executing implementation initiatives. These frameworks are meant to be practical and often include a stepwise sequence of implementation stages. Three examples of process frameworks include the Knowledge to Action framework, the Iowa model and Pronovost’s 4E process theory. Reference Graham, Logan and Harrison22Reference Pronovost, Berenholtz and Needham24

  1. 1. The Knowledge to Action framework describes how new knowledge is created and how this knowledge is applied and adapted to the local context. Reference Graham, Logan and Harrison22 In the country of Cameroon, the Knowledge to Action framework was used to guide the successful national implementation of a surgical safety checklist. Reference White, Daya and Karel25

  2. 2. The Iowa model, developed 25 years ago by nurses and revised in 2017, provides users with a pragmatic flow diagram depicting steps and decision points in the process of moving from the realization that a change in practice is needed to the integration and sustainment of that change. Reference Titler, Kleiber and Steelman23,Reference Collaborative, Buckwalter and Cullen26

  3. 3. Pronovost’s 4E process theory moves through the stages of summarizing the available evidence, identifying implementation barriers, measuring current performance, and targeting key stakeholders with the 4Es of implementation: engage, educate, execute, and evaluate. Reference Pronovost, Berenholtz and Needham24

Determinant frameworks outline different factors that potentially impede or enable the implementation process. These factors are often referred to as “barriers” and “facilitators.” Leveraging determinant frameworks can be helpful during the preimplementation process to inform the optimal approach to implementation. Determinant frameworks can also be used to understand in hindsight why the implementation process either succeeded or failed. Three examples of determinant frameworks include the Consolidated Framework for Implementation Research (CFIR), the integrated Promoting Action on Research Implementation in Health Services (i-PARIHS) framework, and the Theoretical Domains framework (TDF). Reference Damschroder, Aron, Keith, Kirsh, Alexander and Lowery27Reference Michie, Johnston and Abraham29

  1. 1. The CFIR is made up of 5 domains that potentially influence the implementation of interventions: intervention characteristics, outer setting, inner setting, characteristics of individuals, and process. Reference Damschroder, Aron, Keith, Kirsh, Alexander and Lowery27 The CFIR model has been used to study perceptions of diagnosing infections in nursing home residents and the perceptions of antibiotic stewardship personnel regarding why their programs were successful. Reference Barlam, Childs and Zieminski30,Reference Kuil, Schneeberger, van Leth, de Jong and Harting31

  2. 2. i-PARIHS, an updated version of PARiHS, suggests that the success of implementation is dependent on the features of the innovation, the context in which implementation occurs, the people who are affected by and influence implementation, and the manner in which the implementation process is facilitated. Reference Harvey and Kitson28 Researchers have used the PARiHS framework to classify organizational factors that may facilitate ASP design, development, and implementation. Reference Chou, Graber and Zhang32

  3. 3. The TDF outlines 14 potential determinants of healthcare worker behavior, including knowledge, skills, beliefs, environmental context, and resources. Reference Michie, Johnston and Abraham29,Reference Cane, O’Connor and Michie33 Researchers used the TDF framework as part of a cluster randomized trial to understand barriers and facilitators to dentists using local measures instead of systemic antibiotics to manage dental infections. Reference Newlands, Duncan and Prior34

The third type of implementation framework is the evaluation framework. These frameworks help choose implementation outcomes to measure and, in turn, can be useful to incorporate into quality improvement initiatives or clinical trials. Two examples of evaluation frameworks include RE-AIM and a framework described by Proctor et al. Reference Glasgow, Vogt and Boles35,Reference Proctor, Landsverk, Aarons, Chambers, Glisson and Mittman36

  1. 1. RE-AIM sets forth 5 dimensions to evaluate health behavior interventions: reach, effectiveness, adoption, implementation, and maintenance. Reference Glasgow, Vogt and Boles35 These dimensions are evaluated at multiple levels (eg, the healthcare system, clinicians, and patients). In rural China, the RE-AIM framework was used to evaluate a community-based program for disposing households’ expired, unwanted or unused antibiotics. Reference Lin, Wang, Wang, Zhou and Hargreaves37

  2. 2. Proctor et al Reference Proctor, Landsverk, Aarons, Chambers, Glisson and Mittman36 proposes a framework in which improvements in outcomes are dependent not only on the evidence-based practice (the “what”) that is implemented but also on the implementation strategies (the “how”) used to implement the practice. The Proctor framework distinguishes between the evidence-based practice and different levels of implementation strategies (the approach for putting the practice into place). It incorporates 3 levels of outcomes: implementation, service, and client.

Implementation scientists sometimes apply 2 or more framework types simultaneously (eg, CFIR and RE-AIM). Such hybrid frameworks can be useful for projects with multifaceted goals or for identifying clear outcomes with frameworks that are more conceptual.

Establishing stakeholder engagement

In biomedical research, it is traditional for investigators to examine a scientific phenomenon in controlled conditions, removed from the people whose lives or work are affected by the study topic. However, addressing the gap between evidence and practice requires a close engagement between researchers and the “end users” of scientific knowledge. A major contribution of implementation science is the finding that successful implementation is a function of how well an evidence-based practice fits with the preferences and priorities of stakeholders (ie, those who will ultimately be affected by the practice) and the context in which they are embedded. Reference Brownson, Colditz and Proctor7 As such, a key component of an implementation research project is stakeholder engagement.

Stakeholders can include patients and their families, clinicians, health system administrators, community groups, insurers, guideline developers, researchers, professional organizations, and regulatory bodies. Antibiotic stewardship stakeholders will vary depending on the evidence-practice gap and setting. For example, an implementation research project in ambulatory surgical centers might target the underuse of β-lactam antibiotics in patients with reported (but inaccurate) penicillin allergies through the use of a nurse-driven, preoperative, penicillin-allergy delabelling program. A wide range of stakeholders could be involved in such a project, including patients, nurses, surgeons, allergists, pharmacists, health system information technology specialists, surgical center administrators, and health system leaders. Reference Jani, Williams and Krishna38

Once stakeholders are identified, they must be engaged. Although the level of engagement may vary depending on the project, the goal is to ensure that stakeholder needs, preferences, priorities, concerns and beliefs about the evidence-based practice proposed for implementation have been identified and used to inform the study. It is also helpful to understand the relationships among the stakeholders, the roles they might play in facilitating or hindering implementation, which stakeholders have access to “levers” of change in the setting, and stakeholder levels of interest in and commitment to the proposed project. Reference Bernstein, Weiss and Curry39

Stakeholder analysis methods can be used to systematically elicit this information using a combination of research techniques. Reference Brugha and Varvasovszky40,Reference Aarons, Wells, Zagursky, Fettes and Palinkas41 Ideally, stakeholders are engaged from the planning process through the end of the project, and they weigh in on all aspects of study design and conduct, including implementation strategies and outcomes. Reference Crable, Biancarelli, Walkey, Allen, Proctor and Drainoni10 A good example of a systematic approach to stakeholder engagement in stewardship is the development and optimization of the Antibiotic Review Kit (ARK), a multifaceted intervention aimed at reducing excess duration of antibiotic use in hospital inpatients. Reference Santillo, Sivyer and Krusche42,43

Considering context: Identifying barriers to implementation and establishing readiness to change

Implementation science approaches to change in healthcare have highlighted the influence of context on the success of integrating an evidence-based practice into routine use. 44 Context is “a multidimensional construct encompassing micro-, meso-, and macro-level determinants that are pre-existing, dynamic and emergent throughout the implementation process.” Reference Rogers, De Brun and McAuliffe45 These micro, meso, and macro determinants include concepts such as individual stakeholder perceptions and attitudes, teamwork, leadership, organizational climate and culture, readiness to change, resources, health system characteristics, and the political environment. Implementation science frameworks often include contextual domains in their conceptualizations of how successful implementation is achieved, for example, inner–outer setting in CFIR, inner–outer context in i-PARIHS, environmental context, and resources–social influences in TDF.

In the planning phase of an implementation research study, methods should be used to identify and assess contextual determinants likely to influence implementation. Such inquiry can be directed toward understanding the reasons for the evidence–practice gap, establishing a setting’s readiness to change, identifying potential barriers to and facilitators of implementation, and identifying which implementation strategy (or combination of strategies) will be needed to encourage sustainable change.

Mixed methods, or the combination of quantitative and qualitative approaches, are most commonly used to understand context in implementation science. Reference Rogers, De Brun and McAuliffe45,46 Survey methodology can be used to measure attitudes and perceptions of individuals or groups. A number of scales with established validity and reliability can be used at this stage, such as the Implementation Climate Scale (ICS), Organizational Readiness to Change Assessment (ORCA), and the Determinants of Implementation Behavior Questionnaire (DIBQ). Reference Ehrhart, Aarons and Farahnak47Reference Huijg, Gebhardt and Dusseldorp49 Qualitative methodologies, such as focus groups, semistructured interviews, and rapid ethnographic observations, are commonly used to gather information about stakeholder beliefs, existing workflows or processes, social norms, and group dynamics related to the evidence–practice gap. Reference Palinkas, Aarons, Horwitz, Chamberlain, Hurlburt and Landsverk50

The selection of methods should be informed by the research question, what is already known about barriers and facilitators to the evidence-based practice, time and resources available to conduct the work, the implementation framework informing the project, and the availability of collaborators with appropriate expertise. Certain evidence-based practices in stewardship have a robust extant literature identifying reasons for the evidence–practice gap, stakeholder response to interventions and barriers to implementation, such as interventions directed at reducing unnecessary use of antibiotics for viral acute respiratory infections (ARIs) in outpatient primary-care settings. Reference Anthierens, Tonkin-Crine and Cals51Reference Jeffs, McIsaac and Zahradnik56 An implementation research project focused on this gap will likely only need to examine local contextual determinants related to the setting’s readiness to change, the need for adaptations to approaches described in the literature and tailored to the local context, and stakeholder engagement. Less is known about the reasons for other evidence–practice gaps in stewardship, such as antibiotic overuse at hospital discharge. An implementation research project on this topic could include rapid formative evaluation as a component of the preimplementation inquiry. Reference Stetler, Legro and Wallace57Reference Sharara, Arbaje and Cosgrove60

Choosing an implementation strategy or a bundle of strategies

Preimplementation planning informs the choice of an implementation strategy. Almost all stewardship interventions are multifaceted and will leverage >1 implementation strategy. The MITIGATE tool kit, for example, aims to deimplement unnecessary antibiotic use in emergency departments and urgent care clinics. Reference May, Yadav, Gaona, Mistry, Stahmer, Meeker, Doctor and Fleischman61 In this case, the evidence-based practice is not prescribing antibiotics for viral ARIs. The tool kit sets forth multiple implementation strategies to enhance the adoption of this evidence-based practice, including developing stakeholder relationships (eg, identifying champions, obtaining commitment letters), training and educating stakeholders, using evaluative strategies (eg, audit and feedback), and engaging patients. Reference May, Yadav, Gaona, Mistry, Stahmer, Meeker, Doctor and Fleischman61

When choosing an implementation strategy or a bundle of strategies for any given project, several questions should be considered. First, what is the evidence-based practice being implemented? Does the practice need to be adapted or tailored to the local context? There is an inevitable tension between fidelity to the practice and adaptation at the local level. Second, what are the known barriers and facilitators to adopting the evidence-based practice? How would any specific implementation strategy address a known barrier or leverage an established facilitator? Third, are the implementation strategies feasible given local resources? Fourth, would pilot testing the implementation strategy on a small scale provide valuable information before the strategy is evaluated more broadly?

Implementation strategies should be rigorously reported so that they can be scientifically retested and employed in other healthcare settings. This aspect can be challenging, as implementation strategies are complex and are often addressing multiple barriers to change. Supplementary Table 1 (online) provides guidance on the level of detail with which an implementation strategy should be described. Such reports describe the actor, the action, the action targets, temporality, dose, a theoretical justification, and which implementation outcomes are addressed. Consensus criteria for reporting implementation studies have also been published. Reference Proctor, Powell and McMillen16,Reference Pinnock, Barwick and Carpenter62

Evaluating the implementation process

Any type of clinical study design can be used to evaluate implementation. However, a key difference between an implementation trial and an effectiveness trial is the measurement of implementation outcomes.

Implementation outcomes are the effects of deliberate actions to implement new practices. These outcomes are often measured from the perspective of clinicians or patients, and they serve as indicators of whether implementation strategies were successful. When evidence-based practices do not lead to the desired outcome, implementation outcomes can help distinguish the reasons for that failure. Such reasons may include a practice that is a poor fit for the setting or an implementation process that was incomplete. Table 3 provides the definitions of common implementation outcomes and examples of how they have been measured in the antibiotic stewardship literature.

Table 3. Implementation Outcomes and Their Measurement in the Antibiotic Stewardship Literature a

Certain implementation outcomes are best measured during specific phases of the implementation process, as shown in Table 3. The findings from these evaluations can be used in real time to refine implementation efforts, or they may provide a summative assessment once the implementation process has concluded. In addition, some outcomes can be measured before implementation even begins.

Implementation outcomes can be measured using both quantitative and qualitative approaches. Some implementation outcomes can be ascertained via survey tools, like the Acceptability of Intervention Measure (AIM), the Intervention Appropriateness Measure (IAM) and the Feasibility of Intervention Measure (FIM). Reference Weiner, Lewis and Stanick63 However, not all of these measures have been empirically tested for content validity, reliability, and usability, so many investigators create study-specific measures. Implementation science experts have called for more rigorous measure development moving forward. Reference Dorsey, Damschroder and Lewis64 In addition, implementation outcomes can be ascertained through qualitative methods, especially semistructured interviews. In this case, qualitative data collection and analysis are typically informed by the implementation framework that is being used. Administrative data can also be used to measure changes in process utilization. Guidance on using these different tools has been summarized previously. Reference Safdar, Abbo, Knobloch and Seo65,Reference Patton66

Within any implementation initiative, clinical effectiveness should continue to be measured. For antibiotic stewardship, clinical effectiveness outcomes may include volume-based metrics of antibiotic use, length of hospital stay, and Clostridioides difficile infections. If the initiative has potential unintended consequences, these should also be measured. Both the RE-AIM and the Proctor implementation frameworks call for measuring outcomes of both clinical effectiveness and implementation effectiveness when testing implementation strategies. Likewise, implementation-effectiveness hybrid trials focus on both clinical effectiveness and implementation outcomes. Reference Curran, Bauer, Mittman, Pyne and Stetler67 These hybrid trials often make use of study designs commonly used in efficacy studies (eg, randomized allocation) while also evaluating at least 1 implementation strategy.

Putting it all together

In Supplementary Tables 24, we show how different implementation frameworks could be applied to an intervention to reduce antibiotic prescribing for viral ARIs in an emergency department network. Although several different frameworks could be useful for this type of project, we have chosen to highlight a process framework (Provonost’s 4E Process Theory), a determinant framework (CFIR), and an evaluation framework (RE-AIM). Supplementary Table 1 (online) shows how the implementation strategies for such a project could be conceptualized and reported.

In conclusion, antibiotic prescribing is complex and often context dependent, but a large body of evidence exists to guide the optimal use of these agents. Efforts to synthesize and evaluate the quality of this evidence are frequently being updated. Reference Davey, Marwick and Scott6870 In addition, clinical research will continue to define the optimal use of antibiotics for many infection types.

Antibiotic stewardship works to implement this evidence into routine medical practice. From a practical perspective, the discipline of implementation science can be beneficial to antibiotic stewards by encouraging thoughtful reflection on the preimplementation phase and can provide a variety of implementation strategies that can be leveraged to enhance the adoption of evidence-based prescribing practices.

We have also shown how implementation science methods can enhance research on stewardship implementation. In this white paper, we have presented different frameworks and reviewed how implementation outcomes can be measured. We have argued that what has been traditionally described as antibiotic stewardship strategies (eg, prospective audit and feedback, prior authorization) can also be conceptualized as implementation strategies. By reframing stewardship strategies in this manner, we have tried to show how the field of antibiotic stewardship intersects with the discipline of implementation science. In addition, through this reframing, we are encouraging researchers to engage a broader range of literature to examine the full extent of implementation strategies in various clinical contexts, as well as to describe the implementation of stewardship strategies with the same rigor that implementation strategies are reported. This should enable future studies to more effectively contribute to generalizable knowledge on stewardship implementation.

In conclusion, antibiotic resistance is a major public health challenge, and practicing antibiotic stewardship is an important solution to this problem. Finding areas of synergy between the disciplines of antibiotic stewardship and implementation science has the potential to accelerate the pace at which evidence-based antibiotic-prescribing practices are incorporated into routine care.

Supplementary material

For supplementary material accompanying this paper visit https://doi.org/10.1017/ice.2021.480

Acknowledgments

We thank Kristy Weinshel for her support in the preparation of this manuscript and for the feedback we received from the SHEA Antimicrobial Stewardship Committee.

Financial support

No financial support was provided relevant to this article.

Conflict of interest

All authors report no conflicts of interest relevant to this article.

References

Magill, SS, O’Leary, E, Ray, SM, et al. Assessment of the appropriateness of antimicrobial use in US hospitals. JAMA Netw Open 1 2021;4:e212007.CrossRefGoogle Scholar
Fleming-Dutra, KE, Hersh, AL, Shapiro, DJ, et al. Prevalence of inappropriate antibiotic prescriptions among US ambulatory care visits, 2010–2011. JAMA 2016; 315: 18641873.CrossRefGoogle Scholar
Barlam, TF, Cosgrove, SE, Abbo, LM, et al. Implementing an antibiotic stewardship program: guidelines by the Infectious Diseases Society of America and the Society for Healthcare Epidemiology of America. Clin Infect Dis 2016; 62: e51e77.10.1093/cid/ciw118CrossRefGoogle Scholar
Morris, AM, Calderwood, MS, Fridkin, SK, et al. Research needs in antibiotic stewardship. Infect Control Hosp Epidemiol 2019; 40: 13341343.10.1017/ice.2019.276CrossRefGoogle ScholarPubMed
Aims and scope. Implementation Science website. https://implementationscience.biomedcentral.com/about. Published 2021. Accessed June 30, 2021.Google Scholar
Nilsen, P, Stahl, C, Roback, K, Cairney, P. Never the twain shall meet? A comparison of implementation science and policy implementation research. Implement Sci 2013;8:63.CrossRefGoogle ScholarPubMed
Brownson, R, Colditz, G, Proctor, E. Dissemination and Implementation Research in Health: Translating Science to Practice. New York: Oxford University Press; 2018.Google Scholar
Eccles, M, Mittman, B. Welcome to implementation science. Implement Sci 2006;1:1.CrossRefGoogle Scholar
Kitson, A, Straus, SE. The knowledge-to-action cycle: identifying the gaps. CMAJ 2010; 182: E73E77.CrossRefGoogle Scholar
Crable, EL, Biancarelli, D, Walkey, AJ, Allen, CG, Proctor, EK, Drainoni, ML. Standardizing an approach to the evaluation of implementation science proposals. Implement Sci 2018;13:71.CrossRefGoogle Scholar
Doron, S, Davidson, LE. Antimicrobial stewardship. Mayo Clin Proc 2011; 86: 11131123.CrossRefGoogle ScholarPubMed
Joseph, J, Rodvold, KA. The role of carbapenems in the treatment of severe nosocomial respiratory tract infections. Expert Opin Pharmacother 2008; 9: 561575.CrossRefGoogle ScholarPubMed
Tamma, PD, Miller, MA, Cosgrove, SE. Rethinking how antibiotics are prescribed: incorporating the 4 moments of antibiotic decision making into clinical practice. JAMA 2019; 321: 139140.CrossRefGoogle ScholarPubMed
The QUERI implementation roadmap. US Department of Veterans’ Affairs Quality Enhancement Research Initiative website. https://www.queri.research.va.gov/tools/roadmap.cfm. Published 2021. Accessed June 30, 2021.Google Scholar
Waltz, TJ, Powell, BJ, Matthieu, MM, et al. Use of concept mapping to characterize relationships among implementation strategies and assess their feasibility and importance: results from the Expert Recommendations for Implementing Change (ERIC) study. Implement Sci 2015;10:109.CrossRefGoogle ScholarPubMed
Proctor, EK, Powell, BJ, McMillen, JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci 2013;8:139.CrossRefGoogle ScholarPubMed
Curran, GM. Implementation science made too simple: a teaching tool. Implement Sci Commun 2020;1:27.CrossRefGoogle ScholarPubMed
Nilsen, P. Making sense of implementation theories, models and frameworks. Implement Sci 2015;10:53.CrossRefGoogle ScholarPubMed
Damschroder, LJ. Clarity out of chaos: use of theory in implementation research. Psychiatry Res 2020;283:112461.CrossRefGoogle ScholarPubMed
Leeman, J, Birken, SA, Powell, BJ, Rohweder, C, Shea, CM. Beyond “implementation strategies”: classifying the full range of strategies used in implementation science and practice. Implement Sci 2017;12:125.CrossRefGoogle ScholarPubMed
Tabak, RG, Khoong, EC, Chambers, DA, Brownson, RC. Bridging research and practice: models for dissemination and implementation research. Am J Prev Med 2012; 43: 337350.CrossRefGoogle Scholar
Graham, ID, Logan, J, Harrison, MB, et al. Lost in knowledge translation: time for a map? J Contin Educ Health Prof 2006; 26: 1324.CrossRefGoogle ScholarPubMed
Titler, MG, Kleiber, C, Steelman, VJ, et al. The Iowa model of evidence-based practice to promote quality care. Crit Care Nurs Clin N Am Dec 2001; 13: 497509.CrossRefGoogle ScholarPubMed
Pronovost, PJ, Berenholtz, SM, Needham, DM. Translating evidence into practice: a model for large scale knowledge translation. BMJ 2008;337:a1714.CrossRefGoogle Scholar
White, MC, Daya, L, Karel, FKB, et al. Using the knowledge to action framework to describe a nationwide implementation of the WHO surgical safety checklist in Cameroon. Anesth Analg 2020; 130: 14251434.CrossRefGoogle ScholarPubMed
Collaborative, Iowa Model, Buckwalter, KC, Cullen, L, et al. Iowa model of evidence-based practice: revisions and validation. Worldviews Evid Based Nurs 2017; 14: 175182.CrossRefGoogle Scholar
Damschroder, LJ, Aron, DC, Keith, RE, Kirsh, SR, Alexander, JA, Lowery, JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci 2009;4:50.CrossRefGoogle ScholarPubMed
Harvey, G, Kitson, A. PARIHS revisited: from heuristic to integrated framework for the successful implementation of knowledge into practice. Implement Sci 2016;11:33.CrossRefGoogle Scholar
Michie, S, Johnston, M, Abraham, C, et al. Making psychological theory useful for implementing evidence based practice: a consensus approach. Qual Saf Health Care 2005; 14: 2633.CrossRefGoogle ScholarPubMed
Barlam, TF, Childs, E, Zieminski, SA, et al. Perspectives of physician and pharmacist stewards on successful antibiotic stewardship program implementation: a qualitative study. Open Forum Infect Dis 2020;7:ofaa229.CrossRefGoogle ScholarPubMed
Kuil, SD, Schneeberger, C, van Leth, F, de Jong, MD, Harting, J. “A false sense of confidence,” the perceived role of inflammatory point-of-care testing in managing urinary tract infections in Dutch nursing homes: a qualitative study. BMC Geriatr 2020;20:450.CrossRefGoogle Scholar
Chou, AF, Graber, CJ, Zhang, Y, et al. Specifying an implementation framework for Veterans’ Affairs antimicrobial stewardship programmes: using a factor analysis approach. J Antimicrob Chemother 2018; 73: 25592566.CrossRefGoogle Scholar
Cane, J, O’Connor, D, Michie, S. Validation of the theoretical domains framework for use in behaviour change and implementation research. Implement Sci 2012;7:37.CrossRefGoogle ScholarPubMed
Newlands, R, Duncan, EM, Prior, M, et al. Barriers and facilitators of evidence-based management of patients with bacterial infections among general dental practitioners: a theory-informed interview study. Implement Sci 2016;11:11.CrossRefGoogle ScholarPubMed
Glasgow, RE, Vogt, TM, Boles, SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health 1999; 89: 13221327.CrossRefGoogle ScholarPubMed
Proctor, EK, Landsverk, J, Aarons, G, Chambers, D, Glisson, C, Mittman, B. Implementation research in mental health services: an emerging science with conceptual, methodological, and training challenges. Adm Policy Ment Health 2009; 36: 2434.CrossRefGoogle ScholarPubMed
Lin, L, Wang, X, Wang, W, Zhou, X, Hargreaves, JR. Cleaning up China’s medical cabinet-an antibiotic take-back programme to reduce household antibiotic storage for unsupervised use in rural China: a mixed-methods feasibility study. Antibiotics (Basel) 2020;9:212.CrossRefGoogle ScholarPubMed
Jani, YH, Williams, I, Krishna, MT. Sustaining and spreading penicillin allergy delabelling: a narrative review of the challenges for service delivery and patient safety. Br J Clin Pharmacol 2020; 86: 548559.CrossRefGoogle ScholarPubMed
Bernstein, SL, Weiss, J, Curry, L. Visualizing implementation: contextual and organizational support mapping of stakeholders (COSMOS). Implement Sci Commun 2020;1:48.CrossRefGoogle Scholar
Brugha, R, Varvasovszky, Z. Stakeholder analysis: a review. Health Policy Plan 2000; 15: 239246.CrossRefGoogle ScholarPubMed
Aarons, GA, Wells, RS, Zagursky, K, Fettes, DL, Palinkas, LA. Implementing evidence-based practice in community mental health agencies: a multiple stakeholder analysis. Am J Public Health Nov 2009; 99: 20872095.CrossRefGoogle ScholarPubMed
Santillo, M, Sivyer, K, Krusche, A, et al. Intervention planning for Antibiotic Review Kit (ARK): a digital and behavioural intervention to safely review and reduce antibiotic prescriptions in acute and general medicine. J Antimicrob Chemother 2019; 74: 33623370.CrossRefGoogle ScholarPubMed
Antibiotic Review Kit website. http://www.arkstudy.ox.ac.uk/ Accessed June 30, 2021.Google Scholar
May CR, Johnson M, Finch T. Implementation, context and complexity. Implement Sci 2016;11:141.CrossRefGoogle Scholar
Rogers, L, De Brun, A, McAuliffe, E. Defining and assessing context in healthcare implementation studies: a systematic review. BMC Health Serv Res 2020;20:591.CrossRefGoogle ScholarPubMed
Qualitative methods in implementation science. National Cancer Institute website. https://cancercontrol.cancer.gov/sites/default/files/2020-09/nci-dccps-implementationscience-whitepaper.pdf. Published 2020. Accessed June 30, 2021.Google Scholar
Ehrhart, MG, Aarons, GA, Farahnak, LR. Assessing the organizational context for EBP implementation: the development and validity testing of the Implementation Climate Scale (ICS). Implement Sci 2014;9:157.CrossRefGoogle Scholar
Helfrich, CD, Li, YF, Sharp, ND, Sales, AE. Organizational readiness to change assessment (ORCA): development of an instrument based on the Promoting Action on Research in Health Services (PARIHS) framework. Implement Sci 2009;4:38.CrossRefGoogle ScholarPubMed
Huijg, JM, Gebhardt, WA, Dusseldorp, E, et al. Measuring determinants of implementation behavior: psychometric properties of a questionnaire based on the theoretical domains framework. Implement Sci 2014;9:33.CrossRefGoogle ScholarPubMed
Palinkas, LA, Aarons, GA, Horwitz, S, Chamberlain, P, Hurlburt, M, Landsverk, J. Mixed method designs in implementation research. Adm Policy Ment Health 2011; 38: 4453.CrossRefGoogle ScholarPubMed
Anthierens, S, Tonkin-Crine, S, Cals, JW, et al. Clinicians’ views and experiences of interventions to enhance the quality of antibiotic prescribing for acute respiratory tract infections. J Gen Intern Med 2015; 30: 408416.CrossRefGoogle ScholarPubMed
Cals, JW, van Leeuwen, ME, Chappin, FH, de Bont, EG, Dinant, GJ, Butler, CC. “How do you feel about antibiotics for this?” a qualitative study of physician attitudes towards a context-rich communication skills method. Antibiotics (Basel) 2013; 2: 439449.CrossRefGoogle Scholar
Tonkin-Crine, S, Anthierens, S, Francis, NA, et al. Exploring patients’ views of primary care consultations with contrasting interventions for acute cough: a six-country European qualitative study. NPJ Prim Care Respir Med 2014;24:14026.CrossRefGoogle ScholarPubMed
Szymczak, JE, Feemster, KA, Zaoutis, TE, Gerber, JS. Pediatrician perceptions of an outpatient antimicrobial stewardship intervention. Infect Control Hosp Epidemiol 2014;35 suppl 3:S69S78.CrossRefGoogle ScholarPubMed
Zetts, RM, Stoesz, A, Garcia, AM, et al. Primary care physicians’ attitudes and perceptions towards antibiotic resistance and outpatient antibiotic stewardship in the USA: a qualitative study. BMJ Open 2020;10:e034983.CrossRefGoogle ScholarPubMed
Jeffs, L, McIsaac, W, Zahradnik, M, et al. Barriers and facilitators to the uptake of an antimicrobial stewardship program in primary care: a qualitative study. PLoS One 2020;15:e0223822.CrossRefGoogle Scholar
Stetler, CB, Legro, MW, Wallace, CM, et al. The role of formative evaluation in implementation research and the QUERI experience. J Gen Intern Med 2006;21 suppl 2:S1S8.CrossRefGoogle ScholarPubMed
Vaughn, VM, Flanders, SA, Snyder, A, et al. Excess antibiotic treatment duration and adverse events in patients hospitalized with pneumonia: a multihospital cohort study. Ann Intern Med 2019; 171: 153163.CrossRefGoogle ScholarPubMed
Vaughn, VM, Gandhi, TN, Chopra, V, et al. Antibiotic overuse after hospital discharge: a multi-hospital cohort study. Clin Infect Dis 2020;ciaa1372.Google Scholar
Sharara, SL, Arbaje, AI, Cosgrove, SE, et al. A healthcare worker and patient-informed approach to oral antibiotic decision making during the hospital-to-home transition. Infect Control Hosp Epidemiol 2021; 42: 12661271.CrossRefGoogle ScholarPubMed
May, L, Yadav, K, Gaona, SD, Mistry, R, Stahmer, A, Meeker, D, Doctor, J, Fleischman, R. MITIGATE Antimicrobial Stewardship Toolkit: a guide for practical implementation in adult and pediatric emergency department and urgent care settings. Society for Healthcare Epidemiologists of America website. http://shea-online.org/images/priority-topics/MITIGATE_TOOLKIT_final.pdf. Published 2018. Accessed June 30, 2021.Google Scholar
Pinnock, H, Barwick, M, Carpenter, CR, et al. Standards for Reporting Implementation Studies (StaRI) statement. BMJ 2017;356:i6795.CrossRefGoogle ScholarPubMed
Weiner, BJ, Lewis, CC, Stanick, C, et al. Psychometric assessment of three newly developed implementation outcome measures. Implement Sci 2017;12:108.CrossRefGoogle ScholarPubMed
Dorsey, CN, KD; M, AJ; P, Damschroder, LJ, Lewis, CC. A systematic review of measures of implementation players and processes: summarizing the dearth of psychometric evidence. Implement Res Pract 2021; 2: 127.Google Scholar
Safdar, N, Abbo, LM, Knobloch, MJ, Seo, SK. Research methods in healthcare epidemiology: survey and qualitative research. Infect Control Hosp Epidemiol 2016; 37: 12721277.CrossRefGoogle ScholarPubMed
Patton, MQ. Qualitative Research and Evaluation Methods: Integrating Theory and Practice 4th ed. Thousand Oaks, CA: Sage; 2014.Google Scholar
Curran, GM, Bauer, M, Mittman, B, Pyne, JM, Stetler, C. Effectiveness-implementation hybrid designs: combining elements of clinical effectiveness and implementation research to enhance public health impact. Med Care 2012; 50: 217226.CrossRefGoogle ScholarPubMed
Davey, P, Marwick, CA, Scott, CL, et al. Interventions to improve antibiotic prescribing practices for hospital inpatients. Cochrane Database Syst Rev 2017;2:CD003543.Google ScholarPubMed
Core elements of hospital antibiotic stewardship programs. Centers for Disease Control and Prevention website. https://www.cdc.gov/antibiotic-use/core-elements/hospital.html. Published 2019. Accessed June 30, 2021.Google Scholar
Antimicrobial stewardship: systems and processes for effective antimicrobial medicine use. National Institute for Health and Care Excellence website. https://www.nice.org.uk/Guidance/NG15. Published 2015. Accessed June 30, 2021.Google Scholar
Szymczak, JE, Kitt, E, Hayes, M, et al. Threatened efficiency not autonomy: prescriber perceptions of an established pediatric antimicrobial stewardship program. Infect Control Hosp Epidemiol 2019; 40: 522527.CrossRefGoogle Scholar
Huang, J, Kassamali Escobar, Z, Bouchard, TS, et al. Finding the path of least resistance: locally adapting the MITIGATE toolkit in emergency departments and urgent care centers. Infect Control Hosp Epidemiol 2021; 42: 13761378.CrossRefGoogle ScholarPubMed
O’Leary, EN, van Santen, KL, Webb, AK, Pollock, DA, Edwards, JR, Srinivasan, A. Uptake of antibiotic stewardship programs in US acute-care hospitals: findings from the 2015 National Healthcare Safety Network annual hospital survey. Clin Infect Dis 2017; 65: 17481750.CrossRefGoogle ScholarPubMed
Hruza, HR, Velasquez, T, Madaras-Kelly, KJ, Fleming-Dutra, KE, Samore, MH, Butler, JM. Evaluation of clinicians’ knowledge, attitudes, and planned behaviors related to an intervention to improve acute respiratory infection management. Infect Control Hosp Epidemiol 2020; 41: 672679.CrossRefGoogle Scholar
MacBrayne, CE, Williams, MC, Levek, C, et al. Sustainability of handshake stewardship: extending a hand is effective years later. Clin Infect Dis 2020; 70: 23252332.CrossRefGoogle Scholar
Anderson, DJ, Watson, S, Moehring, RW, et al. Feasibility of core antimicrobial stewardship interventions in community hospitals. JAMA Netw Open 2019;2:e199369.CrossRefGoogle ScholarPubMed
Stevenson, LD, Banks, RE, Stryczek, KC, et al. A pilot study using telehealth to implement antimicrobial stewardship at two rural Veterans’ Affairs medical centers. Infect Control Hosp Epidemiol 2018; 39: 11631169.CrossRefGoogle Scholar
Kelly, AA, Jones, MM, Echevarria, KL, et al. A report of the efforts of the Veterans’ Health Administration National Antimicrobial Stewardship Initiative. Infect Control Hosp Epidemiol 2017; 38: 513520.CrossRefGoogle Scholar
Sehgal, P, Elligsen, M, Lo, J, et al. Long-term sustainability and acceptance of antimicrobial stewardship in intensive care: a retrospective cohort study. Crit Care Med 2021; 49: 1926.CrossRefGoogle ScholarPubMed
Figure 0

Table 1. Comparison of Quality Improvement and Implementation Science

Figure 1

Table 2. Implementation Strategies

Figure 2

Table 3. Implementation Outcomes and Their Measurement in the Antibiotic Stewardship Literaturea

Supplementary material: File

Livorsi et al. supplementary material

Livorsi et al. supplementary material

Download Livorsi et al. supplementary material(File)
File 38.6 KB