Skip to main content

Simulation-based team training in time-critical clinical presentations in emergency medicine and critical care: a review of the literature

Abstract

Background

The use of simulation-based team training has increased over the past decades. Simulation-based team training within emergency medicine and critical care contexts is best known for its use by trauma teams and teams involved in cardiac arrest. In the domain of emergency medicine, simulation-based team training is also used for other typical time-critical clinical presentations. We aimed to review the existing literature and current state of evidence pertaining to non-technical skills obtained via simulation-based team training in emergency medicine and critical care contexts, excluding trauma and cardiac arrest contexts.

Methods

This systematic review was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) statement. Before the initiation of the study, the protocol was registered in the International Prospective Register of Systematic Reviews (PROSPERO) database. We conducted a systematic literature search of 10 years of publications, up to December 17, 2019, in the following databases: PubMed/MEDLINE, EMBASE, Cochrane Library, and CINAHL. Two authors independently reviewed all the studies and extracted data.

Results

Of the 456 studies screened, 29 trials were subjected to full-text review, and 13 studies were included in the final review. None of the studies was randomized controlled trials, and no studies compared simulation training to different modalities of training. Studies were heterogeneous; they applied simulation-training concepts of different durations and intensities and used different outcome measures for non-technical skills. Two studies reached Kirkpatrick level 3. Out of the remaining 11 studies, nine reached Kirkpatrick level 2, and two reached Kirkpatrick level 1.

Conclusions

The literature on simulation-based team training in emergency medicine is heterogeneous and sparse, but somewhat supports the hypothesis that simulation-based team training is beneficial to teams’ knowledge and attitudes toward non-technical skills (Kirkpatrick level 2). Randomized trials are called for to clarify the effect of simulation compared to other modalities of team training. Future research should focus on the transfer of skills and investigate improvements in patient outcomes (Kirkpatrick level 4).

Background

The use of simulation-based team training as an educational tool has increased over the past decades [1, 2]. Healthcare professionals most often welcome simulation training, and team-based simulation has been shown to improve knowledge, skills, and behaviors [3, 4]. In emergency medicine (EM), simulation-based team training is best known for its use in trauma and cardiac arrest team training. However, many other patients require time-critical management, including patients in shock, suffering from hypoglycemia, or experiencing adrenal crisis, where a team-based approach is desirable [4].

The purpose of simulation-based training is ultimately to increase the quality of patient treatment and patient safety by increasing adherence to the principles of a team-based approach to resuscitation [5]. If the team adheres to these principles, this will in theory optimize time-critical diagnostics and interventions [5]. If critical diagnostics and time-critical interventions are optimized, it has been speculated that this will improve important patient outcomes, such as morbidity and mortality [6].

The Kirkpatrick Model is well-known as a tool for analyzing and evaluating the results of education and training. It presents a hierarchy that stratifies the level of impact of training results according to four levels: clinical parameters and patient outcomes (level 4), changes in the participants’ behavior after training (level 3), individual learning (level 2), and participants’ initial reactions toward training (level 1) [7].

The focus and results of simulation-based training consist of technical and non-technical skills. Technical skills are defined as the “adequacy of the actions taken from a medical and technical perspective,” while non-technical skills are defined as the “decision-making and team interaction processes used during the team’s management of a situation” [8, 9]. Per definition, training in non-technical skills requires a team effort, whereas technical skills can be performed and measured at either the individual or the team level [10, 11].

Simulation-based training can be performed either in situ or in a simulation center. In situ training takes place in the clinical environment where patients are usually received, and the participants are the actual staff on call; this is the antithesis of training that occurs in a facility away from the clinical setting, in which participants practice team training on a course or in a simulation training program [12]. Simulation mannequins or actors can be used in both settings, and the simulation mannequins can have a variety of patient-like features, including voices and reacting pupils. Depending on the complexity of the set-up, simulations can be denominated either high fidelity or low fidelity [13].

Original research on and reviews of simulation-based team training with trauma or cardiac-arrest teams is manifold [14,15,16]. It has been concluded that simulation-based team training contributed to a significant effect on learning within trauma training [16] and that simulation-based team training contributes to increase in survival to discharge when training in cardiac arrest was implemented [17]. However, research into simulation-based training in other time-critical clinical presentations is not gaining the same attention. It has been proposed to view the obtained experience from simulation-based team training as a library to draw upon when specific scenarios are encountered [18]. It is important to acknowledge that many time-critical diseases other than traumas and cardiac arrest are present in everyday emergency medicine and critical care. These are also taught in simulation-based team training. Hence, broadening simulation training beyond cardiac arrest and trauma is important. Cardiac arrest and trauma care consists of multiple case-dependent tasks (i.e., compressions of the chest), often involve larger and multispecialty teams, and are founded on uniform training prerequisites. Therefore, emphasis on simulation training outside these realms is important. However, the necessary evidence must exist before venturing into costly simulation-based team training. No recent reviews have focused solely on simulation-based team training in EM and critical care outside the trauma team and cardiac arrest contexts.

Since simulation-based team in training cardiac arrest and trauma are covered elsewhere and much training is beyond these areas, we aim to review the existing literature on simulation-based team training in EM and critical care in time-critical patient presentations. The research question for this review is as follows: What literature exists on simulation-based team training within emergency medicine and critical care outside cardiac arrest and trauma? Secondarily, we ask what kinds of training have been researched within this delimitation.

Methods

Review questions

This review aimed to determine what evidence exists to support the usefulness of simulation-based team training in EM and critical care regarding the following:

  • Improvement of attitudes toward simulation training (Kirkpatrick Model level 1)

  • Improvement of team skills in simulation settings (Kirkpatrick Model level 2)

  • Improvement of team skills in clinical practice (Kirkpatrick Model level 3)

  • Improvement of clinical parameters and patient outcomes (Kirkpatrick Model level 4)

The Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) statement was used for the reporting in this review.

Protocol and registration

The details of the protocol for this systematic review were registered in the International Prospective Register of Systematic Reviews (PROSPERO) before title and abstract screening began (Record ID: 161941).

Eligibility criteria

Study inclusion adhered to the following population, intervention, comparison, and outcomes (PICO) criteria:

  • Population: We aimed to identify studies working with populations of health care providers (nurses, doctors, technicians, paramedics, and so on) with clinical responsibilities in EM and critical care.

  • Intervention: We aimed to identify studies examining simulation-based team training in EM and critical care (in situ or in a simulation center).

  • Comparison: We aimed to identify studies comparing participant performance before (control) versus after (intervention) simulation-based team training. The control group was permitted to have been training or teaching as usual or to have undergone no training at all. Participant evaluations occurred over time. We also accepted comparisons between pre- and post-intervention outcomes.

  • Outcomes: We aimed to identify studies investigating the following: (1) improvement of clinical parameters and patient outcomes; (2) improvement of non-technical skills in clinical practice; (3) improvement of non-technical skills in the simulation setting; and (4) improvement in attitudes toward simulation training.

We included all published studies within 10 years prior to the search dates (December 16 and 17, 2019). Manuscripts in English, Danish, Swedish, and Norwegian were included. All study designs were included. We excluded the following types of studies: (1) studies on undergraduate team training; (2) studies focusing exclusively on technical skills; (3) teams with less than three people or without at least one physician on the team; (4) studies focused exclusively on cardiac arrest and/or trauma teams; (5) studies focused exclusively on pediatric patients; and (6) studies focused exclusively on surgical emergencies.

Information sources

We searched the following databases: PubMed/MEDLINE, EMBASE, Cochrane Library, and CINAHL. Citations from included manuscripts were screened, and relevant studies were included in the review.

Literature search

Search terms included the following: Crisis Intervention, Crisis Resource Management, Advanced Life Support, Emergency Medicine, Critical Care, Patient Care Team, Interprofessional Relations, Interdisciplinary Team, Medical Emergency Team OR Medical Emergency Response Team, Simulation Training, Simulation Based. The full search protocol for all databases is available in Supplement 1. An experienced librarian performed the search.

Study selection

MN and JW independently performed the title and abstract screening for eligibility. MN and JW subsequently screened eligible studies via full-text reading for inclusion. The study inclusion process was performed using the Covidence software platform (Covidence.org). In cases of disagreement, a third reviewer (MI) was consulted to obtain consensus.

Data collection process and data items

MN, JW, and SH extracted pre-specified data from each included study. The data included first author, year, country/countries, study design, aim, intervention, control group, type of participant(s), in situ/simulation center, fidelity level, re-test, outcome measure, main results, and Kirkpatrick level.

MN, SH, and JW assessed the risk of bias for the included studies. According to the study design, the following risk of bias tools were used:

  • Non-randomized trials: Risk of Bias in Non-Randomized Studies of Interventions (ROBINS-I) [19].

  • Qualitative studies: Critical Appraisal Skills Programme (CASP) Score [20].

Due to the heterogeneous outcome measures, no meta-analysis was performed. Facing the inability to conduct meta-analysis, we conducted a narrative synthesis of the results from the included studies. We performed the synthesis following the guidelines proposed by Popay et al. [21].

Results

Study selection

The search identified 725 studies and yielded 456 unique studies when duplicates were removed. After title screening and abstract screening, respectively, 414 and 29 studies were identified for full-text screening. A total of 13 full-text studies were included for data extraction and synthesis in this review. Figure 1 shows the PRISMA flow diagram.

Fig. 1
figure1

PRISMA flow diagram of the included studies

Study characteristics

Further details on the included study designs, participants, and Kirkpatrick levels are presented in Table 1. Further details on the study aims, interventions, simulation characteristics, outcomes, and results are presented later in Table 2. None of the studies were randomized controlled trials.

Table 1 Included studies and sample size/types of participants and the Kirkpatrick level investigated in each study
Table 2 Study aims, interventions, simulation characteristics, outcomes, and main results

Participants

According to the inclusion criteria, all studies involved physicians and teams. The included number of participants ranged from 14 to 151 with a median of 57. The number of included physicians ranged from 9 to 51 (median 19). Besides physicians, 10 studies included nurses, ranging from 3 to 75 (median 27), 5 studies included technicians or nurse assistants, ranging from 1 to 50 (median 10), 1 study included 1 paramedic, and 2 studies included advanced practice providers, with 1 study numbering 6 and the other an unknown number less than 9 [22].

Interventions

The studies used heterogeneous interventions. However, all studies included team-based simulation training. The length of training ranged from 15 min to 4 weeks (median 1 day); 1 study did not report the time of training. The use of a combination of didactic sessions and simulation-based training was reported in three studies. In situ simulation was used in 7 out of 13 studies, and eight studies used high-fidelity mannequins in the simulation training.

Comparison

The majority (n = 9) of studies used a pre-post design without a control group [22,23,24, 27, 29,30,31, 33]. Only one study included a retention test [33].

Outcomes

All studies used outcomes that could be stratified according to the Kirkpatrick Model. No studies investigated the impact of training on patient outcomes, such as length of hospital stay or mortality, and hence, the transfer of learning to patient outcomes has not been investigated in any study. Two studies measured behavioral changes in professional settings, and hence reached Kirkpatrick level 3 [27, 28]. To assess participants’ changes in behavior, two studies used the validated Ottawa Global Rating Scale [36]. Out of the remaining 11 studies, 9 reached Kirkpatrick level 2, and only 2 reached level 1. To assess changes in attitudes toward patient safety, two studies used the validated Safety Attitudes Questionnaire [29, 30].

Main results

Below is a short summary of the studies revealed in this review. Overall, 10 quantitative studies and 3 qualitative studies were identified.

Kirkpatrick level 3

The two studies that reached the highest Kirkpatrick level were Hicks et al. [27] and Marker et al. [28]. Hicks et al. reported pre- and post-test results on levels 1, 2, and 3 [27]. The intervention was a 1-day course with 14 participating residents, out of which 10 residents participated in pre- and post-course simulations. Results on Kirkpatrick level 1 were collected via a post-course survey, in which residents agreed or disagreed with predefined statements. Interdisciplinary team training was endorsed by all participants, and agreement was reached regarding the positive impact of the training. Kirkpatrick level 2 was assessed using the Human Factors Attitude Survey (HFAS), which can assess attitudinal shifts regarding team behaviors [37]. The participants filled out the HFAS before and after the intervention. Only 1 out of 23 statements had a statistically significant positive change. To assess Kirkpatrick level 3, the Ottawa Crisis Resource Management Global Rating Scale (Ottawa GRS) was used by two independent reviewers. The reviews were performed using DVD film clips of the pre- and post-course simulations. The Ottawa GRS is a behavioral assessment tool that has been proven valid and has a high interrater reliability [36, 38]. Hicks et al. found a tendency toward better performance in the post-intervention simulation. However, no statistically significant changes were reported between the pre-test and post-test scores. The authors concluded that the lack of significance could be a result of underpower.

Marker et al. [28] reported improvement in preparedness, communication, and teamwork after an intervention when compared to before the intervention. Their results were based on post-intervention interviews with participating physicians. Hence, their results reflected the subjective evaluations of the physicians. The participating physicians described incidents in which the simulation course had changed their behaviors in their everyday clinical lives in a positive direction. No objective scales were used in this trial, which was purely qualitative.

Kirkpatrick levels 1 and 2

On the two lowest Kirkpatrick levels, there were multiple studies using both quantitative and qualitative designs.

One of the studies used a control group. In this study by Meurling et al. [29], a secondary outcome was a reduction in sick leave when a simulation-training program was introduced in an intensive care unit (ICU). To measure the outcome, a different ICU from the same hospital was used as a control. The result was positive and found a decrease in sick leave in the intervention group and an increase in sick leave in the control group during the study period. Meurling et al. [29] also reported changes in self-efficacy identifiable between questionnaires given before and after the training. On a 1–7-point Likert scale, they found a significant rise from 5.6 (SD = 0.9) to 5.9 (SD = 0.7); p < 0.0001.

Similar to Hicks et al. [27], video-recorded simulations were used by Parsons et al. [31] and Truta et al. [33]. Both studies used recordings from a test simulation of all participants before and after the intervention. In both studies, two observers reviewed the recordings, and the observations were rated on a global rating scale. Parsons et al. evaluated a cohort of 14 EM interns on leadership, problem solving, situational awareness, resource utilization, and communication, as well as providing an overall performance score, using the Ottawa CRM Global Rating Scale (GRS) [36]. The authors found no statistically significant improvements resulting from the intervention and concluded that their study was underpowered. Truta et al. [33] had two blinded observers assess the skills of 30 physicians and found statistically significant improvement in the following measured modalities: management and supervision, teamwork and cooperation, decision-making, and situational awareness.

Bond et al. [22] aimed to investigate the benefit of tele-health communication assistance provided by an ICU nurse during simulation. The introduction of a bidirectional video cart made it possible to consult an experienced ICU nurse during an in situ simulation of septic shock. The outcome was self-reported self-confidence before and after the intervention. The results revealed a rise in self-confidence in managing sepsis from a mean of 7.1 (SD = 2.5) to 8.9 (SD = 1.1); p < 0.05.

Brewster et al. [23] also used questionnaires before and after an intervention. The intervention consisted of pre-course learning and 4 h of lectures and simulation. The results were based on questionnaires that were supplemented by a third questionnaire 4 months after the intervention, hence reaching Kirkpatrick level 2. The study revealed a statistically significant rise in satisfaction among nurses and physicians before and after the intervention.

One study found a different approach to simulation training. Dagnone et al. [26] presented their results after conducting “simulation Olympics” (i.e., an intervention in which 11 teams competed against each other in simulated resuscitation scenarios). The study investigated the participants’ responses to an evaluation of the Olympics and found that all participants but one “strongly agreed” or “agreed” to statements expressing satisfaction with the event.

Chan et al. [24] investigated simulation training and relied on questionnaires. They found an increase in scores on a multiple-choice questionnaire for questions regarding resuscitation. The improvement between mean pre- and post-test scores was 11.5%. Couto et al. [25] investigated the use of simulation to improve participants’ ability to identify latent safety threats (LST). An observer filled out a checklist during debriefing and found a higher proportion of equipment-related LST (p < 0.01) after the intervention. These two studies were assessed to reach Kirkpatrick level 2.

Three studies had study designs differing from the aforementioned studies. The first study, by Wong et al. [34], used a mixed-methods approach to investigate the usefulness of interprofessional standardized patient simulation for emergency-department agitation management. An actor who received instructions through an earpiece during the simulation played a standardized patient. All participants were interviewed in interprofessional and uniprofessional focus groups after the intervention. The interviews revealed a “need and desire for more interprofessional training in agitated patient care.”

In a mixed-methods study by Paltved et al. [30], results were presented from a thematic analysis, a needs analysis, and an evaluation of a simulation-training program. The authors argued that a needs analysis is required in order to tailor training for emergency departments. The needs analysis revealed that the handover between shifts could be frustrating and that a common language is lacking. A result of this was that the “staff valued clear and structured communication and communication strategies such as thinking aloud in order to enhance patient safety. These communication skills improved shared understanding, but interruptions impaired communication.” Furthermore, the study included a validated questionnaire (the Safety Attitudes Questionnaire [SAQ]) filled out before and after simulation training, which revealed an increase in the safety climate from the pre-SAQ score (mean = 25.74, SD = 4.41) to the post-SAQ score (mean = 26.59, SD = 4.23); p < 0.001.

Lastly, Rasmussen et al. [32] conducted 17 semi-structured telephone interviews with personnel who had participated in an Advanced Life Support (ALS) course. They analyzed the data using a constructivist grounded theory approach and found challenges transferring the skills from the course to the clinical setting. This was mostly due to the fact that other personnel in the clinical setting did not have the same course background. They concluded that the course was insufficient in training for the development of transferable skills.

Risk of bias within studies

All studies in this review were associated with a serious risk of bias. The reporting of important aspects of the studies was incomplete, leaving items unclear. Figure 2 shows the risk of bias assessment for the non-randomized trials using the ROBINS-I tool, and Table 3 shows the risk of bias assessment for the two qualitative studies using CASP [20].

Fig. 2
figure2

Risk of bias summary in quantitative studies

Table 3 Risk of bias summary in qualitative studies

Discussion

This systematic review revealed 13 studies investigating the impact of simulation training within EM and critical care. All studies focused on simulation-based team training outside the realms of trauma and cardiac arrest.

The study conducted by Hicks et al. [27] stands out in terms of its thoroughness and rigidity. Even though the researchers did not find statistically significant results in favor of simulation, the scaffolding of the study could serve as an inspiration for future work in the field of simulation training. We agree with the authors that the study was underpowered, and the participants were not randomly assigned. We encourage future researchers indulging in team-based simulation to follow the framework set forth by this group. Adding a control group to the study would have strengthened the study, as would random allocation of the participants.

The risk of bias was serious in all studies included. Many studies have been performed on existing courses, leaving the authors with little to no influence on the intervention design or the choice of participants. A solution to this is the more complicated and less feasible approach wherein the investigators themselves design the study and ensure the inclusion of participants in a manner that involves little to no risk of inclusion bias. This will in many cases be a more costly approach [39]. However, this investment will lead to the production of more valuable evidence regarding the actual benefit of simulation-based team training.

In general, studies used pre- and post-simulation questionnaires to report the effect of training. Three studies used video recordings to objectify the outcomes. Only a minority of the studies included validated questionnaires like Ottawa GRS and SAQ; this makes a comparison of the studies difficult. We recommend that future studies use objective outcomes like video recordings where possible. When using self-reported outcomes, we recommend the use of validated questionnaires.

The studies included in this review measured outcomes at low Kirkpatrick levels, and no studies reported results on the highest Kirkpatrick levels. This makes it impossible to synthesize measures of transfer of learning to outcomes directly affecting important patient outcomes. Only one study, by Rasmussen et al., described transferability to clinical practice. The course consisted of a pre-course reading preparation and a 2-day mixed course with simulation as well as didactic sessions. The qualitative interview study found the course insufficient for the transfer of skills to a clinical setting [32]. This finding was contrary to what we expected. It remains uncertain whether conducting team-based simulations within EM settings provides competences that can be transferred to clinical settings and thus potentially benefit patients.

The rationale behind the choice of simulation-based intervention is often scarcely described in the studies included in this review. One study, by Paltved et al., conducted a needs analysis to investigate the proper aim of simulation [30]. The studies generally fail to answer the question of whether the same result would have been reached if the training had been longer or shorter or if a different modality of teaching had been chosen. We encourage future studies to thoroughly establish the rationale behind the design of any intervention.

We propose two main directions for future investigations in simulation-based training within EM and critical care. First, randomized trials with outcomes higher than Kirkpatrick 3 are needed. It is paramount for continuing investment in simulation that the actual patient important outcomes, such as morbidity and mortality, be investigated. One costly way to perform such investigations could be multicenter studies where simulation-based team training could be implemented in one half of the involved centers and patient important outcomes could be measured over time. Second, we find that the interventions in the studies in this review are heterogenetic. Research should aim to identify the effects of different interventions rather than comparing an intervention to itself (via pre-post tests). For the benefit of comparing future studies, it will be necessary to provide evidence that investigates the length of the simulation training versus the level of training obtained. As an example, it could be interesting to investigate if 2-day courses are superior compared to 1-day courses.

Limitations

This review has some limitations. First, the literature review was limited to 10 years from the date of the search. Important studies outside this timeframe may have been missed. Second, the results may have been influenced by publication bias, in that studies with negative results toward simulation have not been published. In this study, this bias may have pushed the results toward a favorable analysis of simulation training.

Conclusions

The included studies in this review suggest positive outcomes in terms of the benefit of simulation-based team training. However, these studies are associated with a serious risk of bias and report on low levels using the Kirkpatrick Model. Hence, more rigorous research is needed to investigate the evidence of the benefit of team-based simulation in EM and critical care contexts.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

PRISMA:

Preferred Reporting Items for Systematic Reviews and Meta-Analysis

PROSPERO:

Prospective Register of Systematic Reviews

EM:

Emergency medicine

PICO:

Population, intervention, comparison, and outcomes

CASP:

Critical Appraisal Skills Programme

HFAS:

Human Factors Attitude Survey

Ottawa GRS:

Ottawa Crisis Resource Management Global Rating Scale

LST:

Latent safety threats

SAQ:

Safety Attitudes Questionnaire

ALS:

Advanced Life Support

ROBINS-I:

A tool for assessing Risk of Bias in Non-randomized Studies of Interventions

References

  1. 1.

    Issenberg SB, McGaghie WC, Petrusa ER, Lee Gordon D, Scalese RJ. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Med Teach. 2005;27:10–28.

    Article  Google Scholar 

  2. 2.

    Okuda Y, Bond W, Bonfante G, McLaughlin S, Spillane L, Wang E, et al. National growth in simulation training within emergency medicine residency programs, 2003-2008. Acad Emerg Med. 2008;15:1113–6.

    Article  Google Scholar 

  3. 3.

    Cook DA, Hatala R, Brydges R, Zendejas B, Szostek JH, Wang AT, et al. Technology-enhanced simulation for health professions education: a systematic review and meta-analysis. JAMA. 2011;306:978–88.

    CAS  PubMed  Google Scholar 

  4. 4.

    Ilgen JS, Sherbino J, Cook DA. Technology-enhanced simulation in emergency medicine: a systematic review and meta-analysis. Acad Emerg Med. 2013;20:117–27.

    Article  Google Scholar 

  5. 5.

    Ten Eyck RP. Simulation in emergency medicine training. Pediatr Emerg Care. 2011; 27: 333-341; quiz 42-44.

  6. 6.

    Boet S, Bould MD, Fung L, Qosa H, Perrier L, Tavares W, et al. Transfer of learning and patient outcome in simulated crisis resource management: a systematic review. Can J Anaesth. 2014;61:571–82.

    Article  Google Scholar 

  7. 7.

    Kirkpatrick. Evaluating training programs: the four levels. 3rd ed. San Francisco: Berrett-Koehler Publishers Inc; 2006.

    Google Scholar 

  8. 8.

    Arora S, Miskovic D, Hull L, Moorthy K, Aggarwal R, Johannsson H, et al. Self vs. expert assessment of technical and non-technical skills in high fidelity simulation. Am J Surg. 2011;202:500–6.

    Article  Google Scholar 

  9. 9.

    Riem N, Boet S, Bould MD, Tavares W, Naik VN. Do technical skills correlate with non-technical skills in crisis resource management?: a simulation study. Br J Anaesth. 2012;109:723–8.

    CAS  Article  Google Scholar 

  10. 10.

    Gordon M, Baker P, Catchpole K, Darbyshire D, Schocken D. Devising a consensus definition and framework for non-technical skills in healthcare to support educational design: a modified Delphi study. Med Teach. 2015;37:572–7.

    Article  Google Scholar 

  11. 11.

    Flin R, Maran N. Identifying and training non-technical skills for teams in acute medicine. Qual Saf Health Care. 2004;13(Suppl 1):i80–4.

    Article  Google Scholar 

  12. 12.

    Patterson MD, Blike GT, Nadkarni VM. In situ simulation: challenges and results. In: Henriksen K, Battles JB, Keyes MA, Grady ML, editors. Advances in patient safety: New directions and alternative approaches. Vol. 3. Performance and Tools. AHRQ Publication No. 08-0034-3. Rockville, MD: Agency for Healthcare Research and Quality; 2008.

  13. 13.

    Massoth C, Roder H, Ohlenburg H, Hessler M, Zarbock A, Popping DM, et al. High-fidelity is not superior to low-fidelity simulation but leads to overconfidence in medical students. BMC Med Educ. 2019;19:29.

    Article  Google Scholar 

  14. 14.

    Whitmore SP, Gunnerson KJ, Haft JW, Lynch WR, VanDyck T, Hebert C, et al. Simulation training enables emergency medicine providers to rapidly and safely initiate extracorporeal cardiopulmonary resuscitation (ECPR) in a simulated cardiac arrest scenario. Resuscitation. 2019;138:68–73.

    Article  Google Scholar 

  15. 15.

    Tiel Groenestege-Kreb D, van Maarseveen O, Leenen L. Trauma team. Br J Anaesth. 2014;113:258–65.

    CAS  Article  Google Scholar 

  16. 16.

    Gjeraa K, Moller TP, Ostergaard D. Efficacy of simulation-based trauma team training of non-technical skills: a systematic review. Acta Anaesthesiol Scand. 2014;58:775–87.

    CAS  Article  Google Scholar 

  17. 17.

    Young AK, Maniaci MJ, Simon LV, Lowman PE, McKenna RT, Thomas CS, Cochuyt JJ, Vadeboncoeur TF. Use of a simulation-based advanced resuscitation training curriculum: impact on cardiopulmonary resuscitation quality and patient outcomes. J Intensive Care Soc. 2020;21(1):57–63.

    Article  Google Scholar 

  18. 18.

    Davis D, Warrington SJ. Simulation training and skill assessment in emergency medicine. [Updated 2020 May 24]. In: StatPearls [Internet]. Treasure Island (FL): StatPearls Publishing.

  19. 19.

    Sterne JAC, Hernán MA, Reeves BC, Savović J, Berkman ND, Viswanathan M, Henry D, Altman DG, Ansari MT, Boutron I, Carpenter JR, Chan AW, Churchill R, Deeks JJ, Hróbjartsson A, Kirkham J, Jüni P, Loke YK, Pigott TD, Ramsay CR, Regidor D, Rothstein HR, Sandhu L, Santaguida PL, Schünemann HJ, Shea B, Shrier I, Tugwell P, Turner L, Valentine JC, Waddington H, Waters E, Wells GA, Whiting PF, JPT H. ROBINS-I: a tool for assessing risk of bias in non-randomized studies of interventions. BMJ. 2016;355:i4919. https://0-doi-org.brum.beds.ac.uk/10.1136/bmj.i4919.

    Article  PubMed  PubMed Central  Google Scholar 

  20. 20.

    Critical Appraisal Skills Programme. CASP qualitative checklist. Retrieved from https://casp-uk.net/casp-tools-checklists/. Accessed July 2020.

  21. 21.

    Popay J, Roberts H , Sowden A , Petticrew M , Arai L, Rodgers M , Britten N, Roen K, Duffy S. Guidance on the conduct of narrative synthesis in systematic reviews: a product of the ESRC methods programme. Lancaster, 2006.

  22. 22.

    Bond WF, Barker LT, Cooley KL, Svendsen JD, Tillis WP, Vincent AL, et al. A simple, low-cost method to integrate telehealth interprofessional team members during in situ simulation. Simul Healthc. 2019;14:129–36.

    Article  Google Scholar 

  23. 23.

    Brewster DJ, Barrett JA, Gherardin E, O’Neill JA, Sage D, Hanlon G. Evaluating team-based inter-professional advanced life support training in intensive care: a prospective observational study. Anaesth Intensive Care. 2017;45:79–87.

    CAS  Article  Google Scholar 

  24. 24.

    Chan CH, Chan TN, Yuen MC, Tung WK. Evaluation of a simulation-based workshop on clinical performance for emergency physicians and nurses. World J Emerg Med. 2015;6:16–22.

    Article  Google Scholar 

  25. 25.

    Couto TB, Barreto JKS, Marcon FC, Mafra A, Accorsi TAD. Detecting latent safety threats in an interprofessional training that combines in situ simulation with task training in an emergency department. Adv Simul (Lond). 2018;3:23.

    Article  Google Scholar 

  26. 26.

    Dagnone JD, Takhar A, Lacroix L. The simulation Olympics: a resuscitation-based simulation competition as an educational intervention. CJEM. 2012;14(6):363–8.

    Article  Google Scholar 

  27. 27.

    Hicks CM, Kiss A, Bandiera GW, Denny CJ. Crisis resources for emergency workers (CREW II): results of a pilot study and simulation-based crisis resource management course for emergency medicine residents. CJEM. 2012;14:354–62.

    Article  Google Scholar 

  28. 28.

    Marker S, Mohr M, Ostergaard D. Simulation-based training of junior doctors in handling critically ill patients facilitates the transition to clinical practice: an interview study. BMC Med Educ. 2019;19:11.

    Article  Google Scholar 

  29. 29.

    Meurling L, Hedman L, Sandahl C, Fellander-Tsai L, Wallin CJ. Systematic simulation-based team training in a Swedish intensive care unit: a diverse response among critical care professions. BMJ Qual Saf. 2013;22:485–94.

    Article  Google Scholar 

  30. 30.

    Paltved C, Bjerregaard AT, Krogh K, Pedersen JJ, Musaeus P. Designing in situ simulation in the emergency department: evaluating safety attitudes amongst physicians and nurses. Adv Simul (Lond). 2017;2:4.

    Article  Google Scholar 

  31. 31.

    Parsons JR, Crichlow A, Ponnuru S, Shewokis PA, Goswami V, Griswold S. Filling the gap: simulation-based crisis resource management training for emergency medicine residents. West J Emerg Med. 2018;19:205–10.

    Article  Google Scholar 

  32. 32.

    Rasmussen MB, Dieckmann P, Barry Issenberg S, Ostergaard D, Soreide E, Ringsted CV. Long-term intended and unintended experiences after advanced life support training. Resuscitation. 2013;84:373–7.

    Article  Google Scholar 

  33. 33.

    Truta TS, Boeriu CM, Copotoiu SM, Petrisor M, Turucz E, Vatau D, et al. Improving nontechnical skills of an interprofessional emergency medical team through a one day crisis resource management training. Medicine (Baltimore). 2018;97:e11828.

    Article  Google Scholar 

  34. 34.

    Wong AH, Auerbach MA, Ruppel H, Crispino LJ, Rosenberg A, Iennaco JD, et al. Addressing dual patient and staff safety through a team-based standardized patient simulation for agitation management in the emergency department. Simul Healthc. 2018;13:154–62.

    Article  Google Scholar 

  35. 35.

    Flowerdew L, Brown R, Vincent C, et al. Development and validation of a tool to assess emergency physicians’ nontechnical skills. Ann Emerg Med 2012;59:376–85.

  36. 36.

    Jirativanont T, Raksamani K, Aroonpruksakul N, Apidechakul P, Suraseranivongse S. Validity evidence of non-technical skills assessment instruments in simulated anaesthesia crisis management. Anaesth Intensive Care. 2017;45:469–75.

    CAS  Article  Google Scholar 

  37. 37.

    Grogan EL, Stiles RA, France DJ, Speroff T, Morris JA Jr, Nixon B, et al. The impact of aviation-based teamwork training on the attitudes of health-care professionals. J Am Coll Surg. 2004;199:843–8.

    Article  Google Scholar 

  38. 38.

    Kim J, Neilipovitz D, Cardinal P, Chiu M, Clinch J. A pilot study using high-fidelity simulation to formally evaluate performance in the resuscitation of critically ill patients: the University of Ottawa critical care medicine, high-fidelity simulation, and crisis resource management I study. Crit Care Med. 2006;34:2167–74.

    Article  Google Scholar 

  39. 39.

    Fletcher JD, Wind AP. Cost considerations in using simulations for medical training. Mil Med. 2013;178:37–46.

    CAS  Article  Google Scholar 

Download references

Acknowledgements

Not applicable

Funding

Corporate HR, MidtSim, Central Denmark Region has supported this work by funding 40 working days to be divided between MN and JW. CP is a co-author but also Medical Director of MidtSim. CP had the same rights and duties, according to the Vancouver criteria, as the other co-authors. Thus, she had no special veto right in any scientific matter relating to the results and conclusions of this review. The details of the funding by MidtSim are defined by contract. Both parties had the right to discontinue the contract if insurmountable disagreements arose.

Author information

Affiliations

Authors

Contributions

MN, SH, and JW drafted the manuscript. All authors critically revised the manuscript and approved the final version. The authors read and approved the final manuscript.

Corresponding author

Correspondence to Jesper Weile.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

None of the other authors have any disclosures.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Weile, J., Nebsbjerg, M.A., Ovesen, S.H. et al. Simulation-based team training in time-critical clinical presentations in emergency medicine and critical care: a review of the literature. Adv Simul 6, 3 (2021). https://0-doi-org.brum.beds.ac.uk/10.1186/s41077-021-00154-4

Download citation

Keywords

  • Emergency medicine
  • Critical care
  • Simulation-based team training