Skip to main content

Optimizing evidence-based practice implementation: a case study on simulated patient protocols in long-term opioid therapy

Abstract

Background

Substantial work has been done to update or create evidence-based practices (EBPs) in the changing health care landscape. However, the success of these EBPs is limited by low levels of clinician implementation.

Objective

The goal of this study is to describe the use of standardized/simulated patient/person (SP) methodology as a framework to develop implementation bundles to increase the effectiveness, sustainability, and reproducibility of EBPs across health care clinicians.

Design

We observed 12 clinicians’ first-time experiences with six unique decision-making algorithms, developed previously using rigorous Delphi methods, for use with patients exhibiting concerning behaviors associated with long-term opioid therapy (LTOT) for chronic pain. Clinicians were paired with two SPs trained to portray individuals with one of the concerning behaviors addressed by the algorithms in a telehealth environment. The SP evaluations were followed by individual interviews, guided by the Consolidated Framework for Implementation Research (CFIR), with each of the clinician participants.

Participants

Twelve primary care clinicians and 24 SPs in Western Pennsylvania.

Main measurement

The primary outcome was identifying likely facilitators for the successful implementation of the EBP using the SP methodology. Our secondary outcome was to assess the feasibility of using SPs to illuminate likely implementation barriers and facilitators.

Results

The SP portrayal illuminated factors that were pertinent to address in the implementation bundle. SPs were realistic in their portrayal of patients with concerning behaviors associated with LTOT for chronic pain, but clinicians also noted that their patients in practice may have been more aggressive about their treatment plan.

Conclusions

SP simulation provides unique opportunities for obtaining crucial feedback to identify best practices in the adoption of new EBPs for high-risk patients.

Setting

Zoom simulated patient evaluations.

Peer Review reports

Introduction

Timely adoption of current evidence-based practices (EBPs) is key to ensuring high-quality care in our changing health care environment. Creating EBPs alone is insufficient to ensure their implementation. Without well-designed implementation strategies, the adoption of these practices can take decades [1]. This is because clinicians often face barriers to implementing EBPs, including limited awareness, resistance to change, and resource constraints. Organizational culture, patient factors, and the complexity of implementation further contribute to the challenges. Evaluation of implementation strategies outside of an active practice setting can address these barriers and increase the likelihood of dissemination, long-term adoption, and appropriate use of EBPs by providing a controlled environment for assessment, feedback, and identification of facilitators for a successful implementation [2,3,4]. We argue that the standardized/simulated patient/person (SP) methodology serves as a valuable tool for formulating implementation strategies for EBPs before their application in practice.

SPs are people trained to portray complex behaviors and react as an actual patient would to a clinician in real time creating a fully interactive patient-clinician experience outside of a real-world practice [5]. SPs can be trained to consistently exhibit specific emotions (e.g., anger [6]), desires (e.g., prescriptions), and/or patient needs (e.g., language barriers [7]) across clinicians. The flexible nature of simulation can be leveraged to reflect either a single patient encounter or multiple patient visits portraying the passage of time depending on the application (e.g., teach providers how to perform a physical exam or re-evaluate patients after a new prescription). While SP methodology is commonly used to train and test clinicians on new techniques [5, 8,9,10,11,12,13,14,15], its application to the planning phases of implementation science remains limited. Our work specifically leverages SP methodology within the planning phases of an implementation bundle for an EBP - a novel approach that has been underutilized in existing literature.

There are several advantages to using SP methodology as a part of implementation strategy. First, the consistent portrayal of a patient case can help identify gaps in EBP implementation and facilitate targeted solutions for future implementation. Second, recruiting clinicians from multiple and diverse practices to use the EBPs with SPs can provide insight into how the EBP would be best implemented in their unique practice setting after the provider has first-hand experience with the EBP. This can provide richer and more diverse insight for implementation scientists relative to feedback from directly implementing an EBP into a singular practice that may not generalize to other clinics.

Likewise, evaluating an EBP outside of the daily activities of a typical clinical practice provides clinicians with immediate and protected time for debriefing. Without dedicated time for good feedback, it is difficult to identify areas of improvement for implementation. Also, developing implementation strategies for EBPs in practice can be high risk for patients. The use of SPs provides a safe environment to develop implementation strategies and gain active experience with EBPs without putting patients at risk [16]. Lastly, SPs can provide insight into events that may be uncommon or take a long time to occur in practice, which can expedite necessary adaptation of implementation strategies for EBPs. Overall, SPs may provide a critical step in increasing the likelihood of a successful adoption of an EBP by identifying the barriers and facilitators prior to implementation in the field.

For these reasons, we adopted the SP methodology for a research project implementing an evidence-based approach to addressing concerning behaviors in patients on long-term opioid therapy (LTOT), such as diversion, use of other substances, or non-adherence to pain therapy. Although the evidence for the effectiveness of LTOT is limited [17,18,19], there are millions of Americans prescribed opioid analgesics yearly, with more than 17% of Americans receiving an opioid prescription in 2017, with an average of 3.4 opioid prescriptions dispensed per patient [20]. Multiple efforts to improve opioid prescribing have occurred on the broader policy level (e.g., prescription drug monitoring programs), the insurance level (limits on doses or length of time), and through education (the RDA risk evaluation and mitigation strategy program [21] and most recently, the drug enforcement agency requirement for training on addiction and opioids) [22]. While opioid prescribing has decreased overall [23], none of these broader measures address concerning behaviors among patients taking LTOT. To augment non-specific recommendations in the CDC guide to prescribing opioids (“weigh the risks and benefits” [24]) and other broader prescribing policy, our team previously developed a set of evidence-based clinical decision-making algorithms using Delphi process to address concerning behaviors among patients prescribed opioids. The lack of uptake of most clinical guidelines [25, 26] led the team to look for effective ways to implement these EBP. Because the concerning behaviors of patients on LTOT may occur sporadically among primary care physicians (PCPs), using the SP methodology would allow for rapid feedback, making it attractive for developing and testing potential implementation methods of the EBP.

In this article, we describe the SP methodology for developing an implementation bundle for a new EBP to address concerning behaviors among patients on LTOT. In conjunction with the SP methodology, we used observation and discussion from one-on-one structured interviews to develop an implementation bundle to increase the likelihood of effective, sustainable, and reproducible adoption in practice. Our approach was guided by the Consolidated Framework for Implementation Research (CFIR), a commonly used tool to guide qualitative inquiry about how clinicians would implement EBPs in practice [27].

Methods

We demonstrate the important and practical use of the SP methodology for developing implementation strategies for a new EBP: 6 treatment algorithms designed to address common and challenging behaviors associated with long-term opioid therapy (LTOT) developed by Merlin and colleagues and published in 2016 [28]. As previously described, these algorithms were developed using a modified Delphi process [29, 30], a rigorous methodology that uses several rounds of questionnaires sent to a panel of experts to find consensus on how to respond to behaviors such as missing appointments with clinicians prescribing the opioid, taking more opioid than prescribed, and substance use. One of the algorithms is included as an example of the new EBP in Fig. 1. In the present study, we conducted SP sessions with providers using 6 SP cases, one for each algorithm. These SP sessions were followed by one-on-one structured interviews with questions mapping onto domains from the CFIR to assist in the development of an implementation bundle for the new EBP.

Fig. 1
figure 1

SEQ figure \* ARABIC 1: “Other Substance Use” Algorithm

Case development

We developed 6 SP cases. Each case simulated a patient exhibiting a unique concerning behavior addressed by the algorithms (see Table 1 outlining the behaviors portrayed). The SP cases were written with unfolding steps to represent three visits with a provider, because the algorithms guide decision points that would normally occur in subsequent follow-up visits in real-life practice (Fig. 1). The unfolding nature of the scenarios was piloted early in the SP case development process to ensure feasibility.

Table 1 SP cases

SP cases were next reviewed by a Patient-Provider Advisory Board (PPAB) consisting of 3 patients with lived experience with opioids, 4 researchers (among whom are PCPs familiar with caring for patients with opioid misuse disorder), and a primary care provider with familiarity with providing care for patients with opioid misuse. SP cases were edited based on feedback from the PPAB. In concert with the review of the 6 cases, the PPAB reviewed the instructions which provided context, expectations for SP-clinician interactions, and training on the algorithms (see Appendix). Finally, cases and instructions were piloted with an SP and a provider outside of the panel. During this pilot, a physician with topical expertise was recruited to interact with SPs portraying two SP cases over three subsequent visits on a remote/telehealth platform (Zoom). This pilot helped to further develop the other five SP cases in structuring how clinicians would be oriented, updated, and guided through the simulations.

Training and description of organization for SPs

Four experienced SPs were recruited from the University of Pittsburgh SP program to portray the patients exhibiting misuse behaviors. The SPs in the University of Pittsburgh School of Medicine SP Program received foundational training in case portrayal, providing feedback, supported physical exam training, and checklist scoring. This 16-h onboarding combines both active training and also guided observation of SP activities. It prepares SPs to identify, recognize, and reward learner skill in portrayal, and to record it faithfully in assessments.

To allow rotation, redundancy and information sharing, the SPs worked in pairs for each case, alternating the role of moderator and patient. When not portraying the patient, the SP acted as a moderator by providing clinicians with inter-visit updates in accordance with what the clinicians ordered in the first session and noted the passage of time between visits. A fifth experienced SP was recruited to proctor the event—orienting the clinicians as they arrived, running the Zoom sessions, and serving as a backup should one of the other SPs not be able to participate. They also were given an overview of case content, portrayal, and event structure. SPs were provided with case materials a week in advance of the portrayal date, were able to ask questions over email, and completed a case-specific training to align portrayal with parameters provided in the inter-visit updates with SP staff in the 45 min preceding the simulation. The SP program follows the Association for Standardized Patient Educators (ASPE) Standards of Best Practice, which “were written to ensure the growth, integrity, and safe application of SP-based education practices.” [9]

Description of session for clinicians

Clinicians were emailed information and instructions about the event prior to participating in the session (see Appendix). All sessions were held virtually via the Zoom interface due to the COVID pandemic. During the sessions, there was a brief orientation for participants. The orientation included (1) a brief training in how to use the algorithms; (2) an overview of how to approach the simulated interaction (i.e., as close to real practice as possible); and (3) an overview of the one-on-one interview that would follow to discuss the approaches to implement the management algorithms.

Clinicians then moved into Zoom breakout rooms to begin their patient encounters. Clinicians were given up to 60 min to have their 3 distinct visits per patient. There was a 15-min break, and then another 60 min for the second patient scenario.

For each of the 60-min SP scenarios, clinicians were told that they were about to see a patient who was being seen by one of their partners (Dr. Williams) who recently left the practice. Dr. Williams had started the patient on opioid therapy and had an opioid agreement with the patient. Participants were given a copy of Dr. Williams’ last progress note and the opioid agreement prior to meeting the patient. After reviewing this information, the clinicians joined a Zoom breakout room with the SP portraying their patient. Once the provider ended the first encounter, the portraying SP turned off their camera, and, to reflect the passage of time between visits, the moderator gave the clinicians the results of any testing they ordered and any information about the patient that had changed between the last and next visit. The provider indicated when they were ready to start the next encounter. This process was repeated between the second and third encounter.

Data collection: semi-structured interviews

Immediately after they interacted with the SPs, each participant completed a one-on-one interview to reflect on and assess the experience, as well as to provide feedback on how the algorithms should ultimately be integrated into practices like theirs. Interviews were conducted by three experienced qualitative data specialists who work at Qualitative, Evaluation and Stakeholder Engagement Research Services (Qual EASE) at the University of Pittsburgh. Multiple interviewers conducted the interviews, because multiple interviews needed to be conducted at the same time following each SP session. Interviewers used a semi-structured interview guide developed by the research team that covered the following domains: (1) Assessment of their orientation to the algorithms, including training; (2) Assessment of their interaction with the SPs; (3) Assessment of and opinions on the algorithms; and (4) Description of how they thought the algorithms would operate in their practices, and how they could best be implemented there. Interviews were conducted on Zoom and recorded.

Questions and further probing were used to best assess how the algorithms could be implemented in their practices, which map onto several CFIR domains and constructs as shown in Table 2.

Table 2 Mapping of CFIR domains and constructs in interview guide

Within one week of their completion, the qualitative methodologist associated with the project wrote a summary of each interview, which was forwarded to the study team so that they could begin to plan for implementation. Following that initial summary, interviews were transcribed verbatim with identifying details redacted. Under the supervision of the qualitative methodologist, experienced analysts at Qual EASE inductively developed a codebook reflecting the content of the interviews, with coding categories reflecting the four areas of the interview guide mentioned above. Use of the codebook was practiced on two transcripts by 2 Qual EASE coders, following which they both applied the codebook to the remaining 10 transcripts. Cohen’s Kappa statistics were used to assess intercoder reliability; the average kappa score was 0.8565, indicating “near perfect” agreement. The primary coder for the project then conducted a conventional content [31] and thematic analysis [32, 33], which was reviewed by the qualitative methodologist, and shared with the study team to better facilitate implementation planning.

Data collection: development of implementation bundle

The final step to developing the implementation bundle—which included materials for initial training, an online algorithm interface, e-consultation support, and electronic health record (EHR) integration for the 6 algorithms—was to review notes from the structured interviews. The bundle was then drafted and reviewed by the PPABs and co-Is.

Recruitment and study sample

Recruitment emails were sent to Community Medical Inc. (CMI). CMI is a network of 400 primary care and specialty physicians who practice throughout western and central Pennsylvania and provide care for over 495,000 patients. The practices cover a large geographic area; however, the network is predominantly in Allegheny County. Participants were required to be primary care clinicians at CMI practices and at least 18 years of age. Each of the clinicians were recruited to participate in two virtual patient evaluations followed by one-on-one interviews. The experience lasted approximately 4 h and clinicians were paid $1000 for their participation. We ultimately recruited 12 PCPs to participate in the virtual experience, which provided two perspectives for each of the 6 SP cases.

Results

Table 3 summarizes the demographic characteristics of the clinicians participating in our study. All of our participants (100%) were trained as physicians with 33% specializing in Internal Medicine, while 66% specialized in Family Medicine during their residency. There was a prevalence of urban practitioners (58%), followed by those in suburban areas (42%), with an absence of participants from rural locales. We had 42% male and 58% female participants. The racial and ethnic composition of our study cohort is diverse, with White participants comprising the majority at 50%, followed by 33% of participants identifying as Asian. Additional categories encompass Hispanic, Latino, or Spanish origin of any race (17%), and two or more races (17%), with a nuanced representation of other racial and ethnic identities.

Table 3 Participant characteristics

Implementation support strategies

When asked about how algorithms should be implemented in practices like theirs, clinicians indicated that the orientation they had received to the algorithms would be a useful implementation support strategy. Other themes illustrating helpful implementation support strategies included (1) the importance of having the algorithm use endorsed by practice leadership, and of having a local “champion” who promoted their use; (2) integration of the algorithm workflow into practice EHRs; (3) practice and location-specific inputs into the algorithms, such that a suggestion to refer to a specialist come with a list of who, specifically, to refer to, or a suggestion to call security provide the practice-specific number for security; (4) access to specialists who could help interpret unclear or difficult-to read drug screens or suggest a particular course of action with a tricky patient.

Representative quotes supporting these themes, as well as the CFIR domains that they map to, are provided in Table 4. These findings were integrated into an implementation toolkit that included an initial training session followed by a suite of supports, including EHR integration, algorithm guidance hosted on a separate website with links to useful tools, and support for clinician participants via e-consultation.

Table 4 Quotes from semi-structured interviews

Simulation feedback from clinicians

We identified two themes related to the physicians’ encounters with the SPs: (1) clinicians found it useful to practice the algorithms with the SPs; (2) while clinicians applauded the skill of the SPs, they noted that not all actual patient counters go so smoothly. Each is presented in more detail below.

Clinicians found it useful to practice the algorithms with the SPs

Clinicians interviewed found it useful to practice the algorithms with the SPs. As will be discussed below, not all clinicians found the scenarios or SP reactions to be fully realistic. However, they did find practicing the algorithms in this way to be a useful way of learning the algorithms. As one provider put it:

It was a good chance to sort of get to look through the algorithm while I’m talking to them and sort of follow along. So, that was good to get familiar with the algorithm itself in a situation where you don’t feel like you’re with a real patient who you’re, like, ignoring to read through the algorithm.

Another provider similarly reflected:

So, that was really helpful, because this is sort of cut and dry of the way it’s written. And not until you’re in an actual patient scenario do you see some of the gray nuances. For example, one of the cases, the patient was having trouble sleeping secondary to pain. So, she was using her oxycodone in the evening to help with sleep, but it was related to pain. So, it wasn’t this clear-cut ‘I’m just using this to fall asleep at night.’ It was ‘I’m using this because at night my pain is worse which is affecting my sleep, so that’s why I’m using it.’ Which is a gray space. So, having the algorithm to sort of follow through and use as a guide let me make sure I’m asking all the right questions, let me make sure I’m offering all the other alternative things, was definitely beneficial.

While clinicians applauded the skill of the SPs, they noted that not all actual patient counters go so smoothly

Many clinicians described the practice session with SPs as being realistic or very similar to encounters with real patients. One provider described themselves as “shocked” at how realistic the SPs were, adding that “I felt very engaged in each of the scenarios. Like, they knew their background, they kind of were living the patient. I was really impressed... the scenarios were spot-on.” Other clinicians described the scenarios as “realistic situations that you can see in the office every day,” and “totally realistic.”

However, some clinicians described pointed differences with real life patient visits. For example, the following provider described that some of their actual patients would simply never agree to the treatment plans presented in the algorithms:

In the back of my mind I’m thinking of my actual patients who I’ve run into these instances and how this would go, and I don’t think it would’ve – it won’t go the way that it went with the SPs. Because it sometimes doesn’t matter how good your rapport is, they just aren’t gonna do what’s suggested... I think I run into much harder stops with some of my real non-SP patients.

Another clinician echoed this description, noting that:

My experience is that patients don’t normally accept what you say so easily. […] The interactions that I have with my patients are not anything like these, ‘cause these were very calm, very reasonable, willing to listen to you; they seemed to have a health literacy level that is well beyond a lot of the patients I deal with.

While these concerns were not voiced by every clinician, they were voiced by clinicians who experienced different scenarios with the SPs, indicating that patients may not always be agreeable to the actions suggested in the algorithm—and that that lack of agreement would be something that would need to be managed in an ongoing patient relationship, rather than disappearing at the end of the role play with the SP.

Discussion

In this study, we used the SP methodology in combination with one-on-one interviews guided by CFIR to develop an implementation bundle for 6 algorithms designed to address common and challenging behaviors associated with LTOT. We found the use of the SP methodology to be a valuable tool for highlighting important components of an implementation bundle. Specifically, we found that an implementation bundle addressing (1) the importance of having the algorithm use endorsed by practice leadership, and of having a local “champion” who promoted their use; (2) integration of the algorithm workflow into practice EHRs; and (3) practice and location-specific inputs into the algorithms would be most effective in promoting the successful adoption and implementation of the EPBs for the LTOT algorithms. We also found that the SPs were realistic in their portray of patients with LTOT; however, it was noted that patients of the clinicians that participated in the simulations were likely to be more resistant to the adoption of the recommendations outlined by the algorithms than the SP portrayal. SPs are trained to recognize and reward participant skill, which may account for this observation.

Of methodological note in the realm of qualitative research: completing the interviews just after the SP interactions set an excellent stage for collecting qualitative data, likely because clinicians had just had a novel experience that was fresh in their minds. They could also talk about the details of the SP cases without concern for inappropriately describing actual patient cases in too much detail and contrast the SPs with their patients in general. This made for highly engaging interviews in which rapport building between interviewer and interviewee was more easily built. Additionally, interviews were conducted by qualitative research specialists who were not personally invested in the development of the algorithms or orientation to the algorithms, setting the stage for open and honest feedback.

Discussion

In this study, we used the SP methodology in combination with one-on-one interviews guided by CFIR to develop an implementation bundle for 6 algorithms designed to address common and challenging behaviors associated with LTOT. Our findings underscore the value of the SP methodology in elucidating essential components of the implementation bundle. Specifically, we found that an implementation bundle addressing (1) the importance of having the algorithm use endorsed by practice leadership, and of having a local “champion” who promoted their use; (2) integration of the algorithm workflow into practice EHRs; and (3) practice and location-specific inputs into the algorithms would be most effective in promoting the successful adoption and implementation of the EPBs for the LTOT algorithms. We also found that the SPs were realistic in their portrayal of patients with LTOT; however, it was noted that patients of the clinicians that participated in the simulations were likely to be more resistant to the adoption of the recommendations outlined by the algorithms than the SP portrayal.

Of methodological note in the realm of qualitative research: completing the interviews just after the SP interactions set an excellent stage for collecting qualitative data, likely because the experience was fresh in their minds. They could also talk about the details of the SP cases without concern for inappropriately describing actual patient cases in too much detail and contrast the SPs with their patients in general. This made for highly engaging interviews in which rapport building between interviewer and interviewee was more easily built. Additionally, interviews were conducted by qualitative research specialists who were not personally invested in the development of the algorithms or orientation to the algorithms, setting the stage for open and honest feedback.

Despite the merits of the SP approach in examining EBP implementation, several limitations warrant consideration. The applicability of SP methodology to diverse practices and various points in care management raises questions about its universal relevance. The effectiveness or practicality of SPs for EBP training and adoption may vary across different clinical settings, requiring careful consideration when extrapolating findings to practices with distinct characteristics or specific care management points.

The selection of long-term opioid therapy (LTOT)-related care as a case study introduces a contextual limitation. While SP methodology effectively addresses concerns within LTOT-related care, the transferability of findings to other healthcare scenarios might be constrained. The unique nature of LTOT-related care may not fully capture challenges present in different medical specialties or care contexts.

Additionally, while the goal of this manuscript is to illuminate the SP methodology, our study's findings may not be universally generalizable, considering factors such as regional variations in healthcare practices and differing levels of familiarity with EBP implementation. The dynamic nature of clinical practice introduces a limitation in capturing all potential scenarios through SP methodology. Clinician encounters with patients can vary widely, and SPs may not fully replicate the complexity of real-world situations.

Overall, this study demonstrates the potential of using the SP methodology guided by the CFIR framework to develop effective implementation strategies for improving care in real-world healthcare settings. The use of SPs allowed the research team to observe the EBP in practice with feedback from end-users with experience from different health care clinics. The CFIR framework provided a comprehensive approach to guiding the development of an implementation bundle that addressed the multiple factors that influence EBP implementation. The study’s success prompts further exploration of whether the developed implementation bundle correlates with increased EBP adoption levels to further validate the use of SPs for this purpose.

Availability of data and materials

The dataset supporting the conclusions of this article is available from the corresponding author on reasonable request.

Abbreviations

CFIR:

Consolidated framework for implementation research

CMI:

Community Medical Inc.

Co-I:

Co-investigator

EBP:

Evidence-based practice

EHR:

Electronic health record

LTOT:

Long-term opioid therapy

PCP:

Primary care physician

PPAB:

Patient-provider advisory board

SP:

Simulated patient

References

  1. Luciano M, Aloia T, Brett J. 4 Ways to make evidence-based practice the norm in Health Care. Harv Bus Rev. 2019;4.

  2. Rapport F, Clay-Williams R, Churruca K, et al. The struggle of translating science into action: Foundational concepts of implementation science. J Eval Clin Pract. 2018;24(1):117–26. https://0-doi-org.brum.beds.ac.uk/10.1111/jep.12741.

    Article  PubMed  Google Scholar 

  3. Brownson RC, Jacobs JA, Tabak RG, et al. Designing for dissemination among public health researchers: findings from a national survey in the United States. Am J Public Health. 2013;103(9):1693–9. https://0-doi-org.brum.beds.ac.uk/10.2105/AJPH.2012.301165.

    Article  PubMed  PubMed Central  Google Scholar 

  4. Powell BJ, Waltz TJ, Chinman MJ, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10(1):1–14.

    Article  Google Scholar 

  5. Barrows HS. An overview of the uses of standardized patients for teaching and evaluating clinical skills. AAMC Acad Med. 1993;68(6):443–51.

    Article  CAS  PubMed  Google Scholar 

  6. Liao C-S, Hsieh M-C. Standardized Patient Training: Using ANGER to quickly evoke anger in standardized patients. Med Teach. 2015;37(9):883–883.

    Article  PubMed  Google Scholar 

  7. Zabar S, Hanley K, Kachur E, et al. “Oh! She doesn’t speak English!” Assessing resident competence in managing linguistic and cultural barriers. J Gen Intern Med. 2006;21:510–3.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Taylor LJ, Adkins S, Hoel AW, et al. Using Implementation Science to Adapt a Training Program to Assist Surgeons with High-Stakes Communication. J Surg Educ. 2019;76(1):165–73. https://0-doi-org.brum.beds.ac.uk/10.1016/j.jsurg.2018.05.015.

    Article  PubMed  Google Scholar 

  9. Bryant KA, Wesley GC, Wood JA, Hines C, Marshall GS. Use of standardized patients to examine physicians’ communication strategies when addressing vaccine refusal: a pilot study. Vaccine. 2009;27(27):3616–9. https://0-doi-org.brum.beds.ac.uk/10.1016/j.vaccine.2009.03.048.

    Article  PubMed  Google Scholar 

  10. Lewis KL, Bohnert CA, Gammon WL, et al. The association of standardized patient educators (ASPE) standards of best practice (SOBP). Adv Simul. 2017;2(1):1–8.

    Article  Google Scholar 

  11. Wilson L, CHSE C-A, Wittmann-Price RA. Review manual for the certified healthcare simulation educator exam. Springer Publishing Company; 2018.

    Book  Google Scholar 

  12. Jeffries P. Clinical simulations in nursing education: Advanced concepts, trends, and opportunities. Lippincott Williams & Wilkins; 2022.

    Google Scholar 

  13. Ross JG, Burrell SA. Standardized patient simulation to facilitate learning in evidence-based oncology symptom management. J Nurs Educ. 2018;57(4):250–3.

    Article  PubMed  Google Scholar 

  14. Rethans JJ. Needs assessment in continuing medical education through standardized patients. J Contin Educ Health Prof. 1998;18(3):172–8.

    Article  Google Scholar 

  15. Ainsworth MA, Rogers LP, Markus JF, Dorsey NK, Blackwell TA, Petrusa ER. Standardized patient encounters: a method for teaching and evaluation. JAMA. 1991;266(10):1390–6.

    Article  CAS  PubMed  Google Scholar 

  16. Oxman AD, Flottorp SIGNE. An overview of strategies to promote implementation of evidence-based health care. Evid-Based Pract Prim Care. 2001;2:101–19.

    Google Scholar 

  17. Busse JW, Wang L, Kamaleldin M, et al. Opiods for Chronic Noncancer Pain: A Systematic Review and Meta-analysis. JAMA. 2018;320(23):2448–60.

    Article  PubMed  PubMed Central  Google Scholar 

  18. Krebs EE, Gravely A, Nugent S, et al. Effect of opioid vs nonopioid medications on pain-related function in patients with chronic back pain or hip or knee osteoarthritis pain: the space randomized clinical trial. JAMA. 2018;319(9):872–82.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  19. Chou R, Turner JA, Devine EB, et al. The effectiveness and risks of long-term opioid therapy for chronic pain: a systematic review for a National Institutes of Health Pathways to Prevention Workshop. Ann Intern Med. 2015;162(4):276–86.

    Article  PubMed  Google Scholar 

  20. Prescribing Practices. Centers for Disease Control and Prevention. Updated August 13, 2019. Accessed December 6, 2023. https://www.cdc.gov/drugoverdose/deaths/prescription/practices.html.

  21. Opioid Analgesic Risk Evaluation and Mitigation Strategy (REMS). U.S. Food and Drug Administration. Updated November 14, 2023. Accessed January 31, 2024. https://www.fda.gov/drugs/information-drug-class/opioid-analgesic-risk-evaluation-and-mitigation-strategy-rems.

  22. U.S. Department of Justice, Drug Enforcement Administration. Letter to DEA Registered Practitioners. https://deadiversion.usdoj.gov/pubs/docs/MATE_Training_Letter_Final.pdf. Accessed January 31, 2024.

  23. United States Dispensing Rate Maps. Centers for Disease Control and Prevention. Updated December 11, 2023. Accessed January 31, 2024. https://www.cdc.gov/drugoverdose/rxrate-maps/index.html.

  24. Dowell D, Ragan KR, Jones CM, et al. CDC Clinical Practice Guideline for Prescribing Opioids for Pain — United States, 2022. MMWR Recomm Rep. 2022;71:1–95 No. RR-3.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Fischer F, Lange K, Klose K, et al. Barriers and Strategies in Guideline Implementation – A Scoping Review. Healthcare (Basel). 2016;4(3):36.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Jiang V, Brooks EM, Tong ST, et al. Factors Influencing Uptake of Changes to Clinical Preventitive Guidelines. J Am Board Fam Med. 2020;33(2):271–8.

    Article  PubMed  PubMed Central  Google Scholar 

  27. Breimaier HE, Heckemann B, Halfens RJ, Lohrmann C. The Consolidated Framework for Implementation Research (CFIR): a useful theoretical framework for guiding and evaluating a guideline implementation process in a hospital-based nursing practice. BMC Nursing. 2015;14:1.

    Article  Google Scholar 

  28. Merlin JS, Young SR, Azari S, et al. Management of problematic behaviours among individuals on long-term opioid therapy: protocol for a Delphi study. BMJ Open. 2016;6(5):e011619.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Dalkey N, Helmer O. An experimental application of the Delphi method to the use of experts. Manag Sci. 1963;9(3):458–67.

    Article  Google Scholar 

  30. Keeney S, McKenna HA, Hasson F. The Delphi technique in nursing and health research. John Wiley & Sons; 2011.

    Book  Google Scholar 

  31. Hsieh HF, Shannon SE. Three approaches to qualitative content analysis. Qual Health Res. 2005;15:1277–88.

    Article  PubMed  Google Scholar 

  32. Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3:77–101.

    Article  Google Scholar 

  33. Braun V, Clarke V. Thematic Analysis: A Practical Guide. Thousand Oaks, CA: Sage Publications; 2021.

    Google Scholar 

Download references

Acknowledgements

Not applicable.

Checklist

A completed SRQR checklist has been completed for this paper.

Funding

We gratefully acknowledge funding from NIDA Agency for our publication through an R34 grant mechanism entitled “Consensus-based algorithms to address opioid misuse behaviors among individuals prescribed long-term opioid therapy: developing implementation strategies and pilot testing.” Project Number: 5R34DA050004-03. Jessica Merlin is supported by the following grant from the NIH: K24DA05683701A1.

Author information

Authors and Affiliations

Authors

Contributions

EG contributed to the development of the standardized patient protocol and was a major contributor to writing the manuscript. MH conducted the interviews as well as analyzed and interpreted the data. CG and RVD developed the standardized patient protocol and conducted the simulations. JDW and JML contributed to the development of the standardized patient cases and interpretation of the data to ensure it aligned with primary care practice. JM provided oversight to the entire study.

Corresponding author

Correspondence to Ellen Green.

Ethics declarations

Ethics approval

This study was conducted at School of Medicine, University of Pittsburgh, Pittsburgh PA between June and July of 2021. The University of Pittsburgh IRB determined that the study was considered an exempt-level research project (STUDY20030189).

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix: Instructions for participants

Appendix: Instructions for participants

Dear Participant,

Thank you for participating in our study of opioid misuse in primary care.

In this exercise, you will encounter 2 different simulated patients played by standardized patients (SPs) of the SP Program of the University of Pittsburgh School of Medicine. The purpose of these visits is to help us study clinical algorithms for managing opioids.

Therefore, please be aware of the following expectations:

  • For this simulated scenario, each of these patients were started on opioid treatment by one of your partners who recently left your practice (Dr. Kia Williams). You may not have started opioids if it were up to you, but they have already been started and have an opioid agreement with this practice. Therefore, please focus your time on the algorithms and not on whether the patient should/should not have been started on opioids.

  • You will see each patient in 3 separate “telemedicine” visits via Zoom. Therefore, you do NOT need to perform a physical examination for these visits.

  • The “visits” will occur in break out rooms on the Zoom platform. The 1st visit will be to establish care with you after Dr. Williams has left the practice. The next 2 visits will be follow-up visits.

  • For each scenario, there will be a “moderator” in the breakout room with you and the SP. The moderator’s camera will be off. This person will be helping with timing of the visits, and they will post updates about the patient’s case before each visit in the chat section.

  • Therefore, please enable the chat on your screen.

  • Also, please “hide nonvideo participants”, so the presence of the moderator is not a distraction for you as you conduct the visits. (If you need help in how to do this, please ask, so a team member can walk you through the steps)

  • In the interest of transparency, the moderators are also SPs. They are not clinicians.

  • After you are done with the visits, you will meet with researchers from the study to debrief your experience.

Timing of the whole activity:

  • Orientation: 30 min

  • Encounter with 1st patient: 60 min

  • Break (including time to prepare for 2nd patient): 15 min

  • Encounter with 2nd patient: 60 min

  • Debrief with researchers: 75 min

Timing of your patient visits: You have 1 h for each session, which includes 3 distinct visits with the same patient. You will see timing banners at 15-min increments, and a 5-min warning. How you divide the time between the three visits is up to you.

If it would help you communicate with the patients in the simulation, here is some information about Dr. Kia Williams:

  • Dr. Williams recently left your practice to be closer to her family in South Carolina. Her father’s dementia has been worsening, and she wanted to be closer to her family in this time. As your partner, she was well liked by your colleagues, the staff, and her patients. She was an excellent doctor and a friend.

  • Prior to meeting each patient, we will share Dr. Williams’ last progress note with you. You will have time to review that information before starting the first visit. This note will have information about what work up has been done and what pain treatments have been tried.

  • The patients are aware of what has been tried, and they can answer these questions, but for the sake of this study, you do not need to explore this in great detail given the limited timeframe of each visit.

  • Dr. Williams had an opioid agreement with each of the patients you will meet today. This will also be shared with you in case you need to reference it during the sessions.

Thank you

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Green, E., Hamm, M., Gowl, C. et al. Optimizing evidence-based practice implementation: a case study on simulated patient protocols in long-term opioid therapy. Implement Sci Commun 5, 44 (2024). https://0-doi-org.brum.beds.ac.uk/10.1186/s43058-024-00575-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/s43058-024-00575-y