Back to Journals » Advances in Medical Education and Practice » Volume 6

Simulation training for foundation doctors on the management of the acutely ill patient

Authors Cachia M, Pace-Bardon M, Balzan G, Tilney R, Micallef J, Balzan M

Received 17 September 2015

Accepted for publication 28 October 2015

Published 16 December 2015 Volume 2015:6 Pages 657—663

DOI https://doi.org/10.2147/AMEP.S96566

Checked for plagiarism Yes

Review by Single anonymous peer review

Peer reviewer comments 2

Editor who approved publication: Dr Md Anwarul Azim Majumder



Monique Cachia,1 Michael Pace-Bardon,2 Gabriella Balzan,2 Russel Tilney,2 Josef Micallef,2 Martin Balzan2

1Department of Medicine, 2Department of Respiratory Medicine, Mater Dei Hospital, Msida, Malta

Background: A study evaluating subjective trainee responses to simulation training organized by the Malta Foundation Program in particular whether this changed their clinical practice.
Method: Feedback using a standardized questionnaire was obtained from 120 (M=55%) participants. A 0–10 Likert scale was used to evaluate responses.
Results: Participants scored the simulation sessions as “useful” at 7.7 (95% confidence interval [CI] 7.4–8.0), rated “the overall experience” at 7.5 (95% CI 7.2–7.8), and thought it made a change in “daily practice” at 5.83 (95% CI 5.4–6.3). The score for the tutor “creating a satisfactory learning environment” and “quality of simulator equipment” was 7.8 (95% CI 7.6–8.1) and 7.7 (95% CI 7.4–8), respectively. Trainees rated “how close was the simulation to a real-life scenario” as 6.24 (95% CI 5.9–6.6). When asked whether the presence of colleagues hindered or helped, the majority were neutral 50 (41.7%), 36 (30%) said it hindered, while only 21 (28.3%) felt it helped. In contrast, 94 (78.33%) stated it was useful to observe colleagues while only 5 (4.2%) stated it was not. Likelihood for future participation was 7.4 (95% CI 7–7.8). Trainees recommended a median of 3 (interquartile range 2–5) simulations per year.
Conclusion: Trainees rated the sessions as useful and asked for more sessions possibly at an undergraduate level. Rating for equipment and tutors was positive; however, some felt that the effect on daily practice was limited. Most were comfortable observing others and uncomfortable being observed. The value of increasing sessions to 3–4 per year, timing them before clinical attachments and audiovisual prebriefing for candidates naïve to simulation needs to be evaluated in future studies.

Keywords: simulation, foundation, training, acute, medical, mannequin

Background

Simulation-based medical training has become a recommended and useful training tool for both undergraduate and postgraduate medical trainees. Advances in simulator technology together with rapidly rising public expectations in the quality of medical care have led to increasing efforts at replacing or supplementing traditional learning methods with simulator-based medical education replicating realistic and life-like clinical encounters.1

The Malta Foundation Program for newly qualified doctors is an affiliated school of the UK Foundation Program that includes a 2-year residency training program after graduation from medical school leading to a medical license to practice in Malta or the UK. It includes structured teaching, hands-on training, and continuous assessments. It has been making use of simulation-based training since 2009.2

Constructivism in education occurs when the learner engages in an experience with a tutor and then “constructs” his own interpretation of that experience. Simulation training provides an opportunity to observe the potential benefit of this method of teaching. This audit was aimed at obtaining subjective responses from learners to see if simulation training helped transform their medical knowledge into changed and hopefully improved medical practice.3

Methodology

The Malta foundation Program has organized regular 3-hour simulation training sessions for every trainee in each foundation year focusing on care of the acutely ill patient in which trainees were faced with various acute hospital-based case scenarios. Table 1 shows some commonly used clinical scenarios and Table 2 shows the main educational objectives during a typical scenario, for example, “acute myocardial infarction”.

Table 1 Commonly used clinical scenarios
Abbreviation: COPD, chronic obstructive pulmonary disease.

Table 2 Educational objectives during a typical scenario “acute myocardial infarction (AMI)”
Abbreviations: AMI, acute myocardial infarction; BP, blood pressure; ECG, electrocardiogram; ST, segment elevation.

The tutors present prepared clinical scenarios at each session. The main educational objectives of the sessions included proper assessment of the situation, taking a short focused history, examination, appropriate investigations, and management. Candidates were also assessed on their ability to correctly prioritize, refer when appropriate, and communicate well with the patient and their colleagues.

Setting

Simulation sessions were held in the simulation training room at the postgraduate training center at Mater Dei Hospital.

The sessions were hands on and utilized a mannequin – the Gaumard HAL® S3201 (Gaumard Scientific, Miami, FL, USA) Advanced tetherless patient simulator with many of the physiological and anatomical responses of a human, including variable vital signs pupil reaction, cyanosis, fits (manifested as a vibratory movement of the mannequin) and heart, breath and bowel sounds which could be altered by the tutors according to the scenario presented. A trained nurse was available at each scenario to assist the participants by delivering the items needed, such as oxygen masks, drip sets, and administering any fluids or drugs being prescribed by the trainee. Participants were also able to practice other skills such as intravenous cannula insertion and bloodletting on a separate model.

A maximum of eight participants were allowed per session and then assessed in pairs. The remaining six participants were asked to view the session from a separate observation room fully equipped with audiovisual aids giving them access to the scenario room and all vital signs. The tutors present during each session consisted of a registered specialist consultant in internal medicine, two senior postgraduate trainees in internal medicine, and a registered nurse.

The three tutors were seated in the control room behind a glass screen within the simulation room and were in full control of the vital and clinical signs as well as dubbing the voice of the simulator mannequin. The nurse assisted the participants realistically during the actual scenario. Trainees were allowed telephone access to their seniors or colleagues whose role was played by one of the three tutors.

Each case scenario lasted 15–20 minutes and participants were asked to manage the patient’s presenting complaint as they would during any acute real-life situation. After each scenario, the tutors provided both the participants and observers with a 15-minute debriefing session that would include giving feedback to the two trainees being assessed. Trainees were graded from A to D based on their performance, with D being the lowest pass mark. Trainees obtaining an F would fail the session and thus require a repeat session.

Design and data collection

Data protection permission was obtained from the hospital data protection officer and ethics committee in line with EU legislation. Questionnaires were distributed and collected over a 14-day period in 2015. Data collection was done in a separate venue and at a later date from the simulation sessions. Written informed consent was obtained from the Foundation doctors who answered 14 questions related to their experience during the simulation training sessions organized by the Malta Foundation Program. Data for calculation of cost was provided by Malta Foundation Program school administration. Validation of the questionnaire was done by peer review of two specialists in internal medicine. Internal consistency of the questions was tested with Cronbach alpha. For Questions 3, 5, 6, 11, and 12 it was 0.8, and for Questions 4 and 8 it was 0.6. Figure 1 shows the questionnaire that was used and the 0–10 point Likert scale where 5 was neutral, 0 scored the lowest response, and 10 the highest response.

Figure 1 Simulation training questionnaire.

The questionnaires were analyzed using Microsoft Excel and Minitab 17 (Minitab Inc., State College, PA, USA). The average scores out of 10, standard deviation, confidence intervals (CIs), median, and interquartile range (IQR) were calculated. The 95% CI of the response was calculated from the total number, mean, and standard deviation.

Results

A total of 120 participants (n=120) completed the questionnaire out of 210 Foundation Program doctors in 2015; 70 first year, 45 second year, and 5 unknown. The main results are shown in Table 3 and Figure 2.

Figure 2 Median, interquartile range, and range for questions 3–12.
Abbreviation: Q, question.

Table 3 Trainee response of simulation session using a 10-point Likert scale

From a Likert scale of 0–10, participants scored the simulation session as “useful” at 7.7 (95% CI 7.4–8.0), though providing a moderate change in “daily practice” at 5.83 (95% CI 5.4–6.3). When asked “how close was the simulation to a real-life scenario” the score was 6.24 (95% CI 5.9–6.6).

The score for the tutor “creating a satisfactory learning environment” was 7.8 (95% CI 7.6–8.1) and that for “quality of simulator equipment” was 7.7 (95% CI 7.4–8). The effect of the presence of colleagues on performance was evaluated among trainees. The majority n=50 (41.7%) were neutral, n=36 (30%) felt it hindered their performance while only 21 (28.3%) felt it helped. In contrast, the majority n=94 (78.33%) stated it was helpful to observe colleagues.

The likelihood for future participation was 7.4 (95% CI 7–7.8). Trainees rated “the overall experience” at 7.5 (95% CI 7.2–7.8) and recommended a median of 3 (IQR 2–5) simulations per year.

In 2014, a total of three trainees obtained a failure during their session and thus required a repeat session during this same year.

The questionnaire also allowed participants to voice their suggestions for possible improvement with many requesting more sessions at an earlier stage and with some commenting on technical difficulties that were encountered during the sessions.

Costs were calculated as shown in Table 4 bearing in mind the depreciation of equipment by 20% per year. A total of 25 simulation sessions were organized during 2014 for 200 foundation program trainees; with sessions consisting of eight trainees and five staff members (one nurse, one technician, and three medical tutors).

Table 4 Total cost/trainee/session

Discussion

It has been reported that simulation-based medical training is a useful educational tool. It promotes skill acquisition through active participant practice, experiential and repetitive learning, reflection, and feedback in a controlled safe environment that simulates a real-life scenario.4,5 Simulation training is being widely used in anesthesia, endoscopy training, trauma, as well as advanced cardiac life support.6

Analysis of results shows that trainee responses were mostly positive. Participants stated that the sessions were useful (7.68) and that tutors created a satisfactory learning environment (7.83). The quality of the simulator equipment was also praised (7.7). Respondents felt that the simulations were of high educational value and requested a median of 3 simulation sessions per year rather than the current yearly session. On the other hand, despite the positive feedback in almost all areas, when asked about whether simulation led to a change in clinical day-to-day practice, the responses were more neutral.

This could be explained by the following considerations. First of all, many trainees attended simulation sessions later in the year with the result that they would have already gained significant first hand experience in their daily clinical practice. Second, many trainees felt that the simulation session did not “closely simulate a real-life scenario”, which could account for the moderate response to “a change in daily practice”. This was also the first experience to this kind of training for 1st year trainees who might have had higher expectations of this special training method being unaware of its inevitable limitations. A more positive response on change in daily practice might have been achieved by providing a prebriefing session, perhaps with audiovisual description of simulation so as to help naïve trainees in familiarizing with the simulation environment prior to the actual simulation sessions.7 However, 2nd year trainees would have an advantage in this respect as they would have had their first session the previous year.

A number of participants requested that sessions be organized earlier on during their foundation training, if not at undergraduate level. Undergraduate simulation training as well as simulation in the 1st month of the program would be ideal especially if followed by two sessions during the year. This is, however, difficult to achieve in view of time constraints and trainer availability together with increasing costs to the employer.

In fact a research study performed by Miles et al8 evaluated UK Foundation doctors’ induction experiences and results showed that trainees felt they are having inadequate inductions prior to rotations. Perhaps introducing simulation sessions earlier on during induction may ameliorate the induction experience.

In a similar study performed by the University of Hertfordshire in 2008, results were mostly positive.9 On the other hand this same study differed from our study, in that the majority of participants did not feel that the presence of their peers hindered their performance. In our case there were more individuals who felt being observed by their peers hindered their performance (30%) as opposed to those who did not (28.3%), while the majority remained neutral on the matter (41.7%). This discrepancy could be explained by the fact that in the Hertfordshire study, participants were from different professions and possibly different centers, while all trainees in our study were doctors from the same year working in the same foundation school after 5 years of medical school together. The rationale for this would be embarrassment or fear of being watched, supervised, judged, or criticized after their simulated scenario.9

Furthermore, evaluation of performance might have played a role in this since debriefing and feedback was given to all trainees together and trainees failing the scenario would require a repeat session.

An end of session debriefing ensures adequate individualized feedback is given to each trainee allowing participants to reflect on their own performance and create memorable experiences that will aid in lifelong learning.10 A systematic review in 2005 highlighted that feedback and debriefing occurring during simulation training sessions are among the most important parts of simulation-based medical education.11

Limitations of this study included the small sample size and feedback based on a single 3-hour session with limited scenarios. The study did not analyze or quantify improved patient outcomes following the simulation session or compare this to a control group consisting of trainees who were traditionally trained but looked specifically at subjective trainee responses. Negative experience by some trainees at the session might have led to negative feedback and may not truly reflect the outcomes of that session. Another limitation is that responses were not taken immediately after the session. On the other hand, the small number of tutors and the identical use of mannequin and simulation center equipment were positive aspects to this study.

Simulation sessions allow the trainee to practice in ways that would not be possible in a real-life scenario as this would be unethical and unsafe to the sick patient. It provides a controlled environment permitting multiple practice attempts where mistakes or failures will not put patients’ welfare at stake.12

This form of training assesses individual learning as well as the participant’s abilities of working within a team – a crucial concept of the multidisciplinary nature of modern medical care.5 It targets specific trainee needs by allowing repetition and providing consistency that would be difficult to achieve when dealing with real patients.13 Simulation training also allows exposure to rare complex scenarios and gives multiple trainees the same learning opportunities.12 Simulation training has even been shown to be more cost-effective than conventional learning techniques by leading to improved standards in patient care.14 In a meta-analysis done by McGaghie et al,1 simulation-based medical education with deliberate practice was found to yield better results than traditional clinical education. Improved patient practice and outcome was seen in laparoscopic surgeries, bronchoscopies, and managing difficult obstetric deliveries.

In a study performed by Karakus et al,15 computer-based simulation training in emergency medicine was found to have led to a statistically significant improvement in skill acquisition and success rate in managing complicated cases when compared to a control group of students who were traditionally trained.

This form of training has proved to be of utmost benefit as it is able to target multiple trainees within a given set timeframe despite time restrictions for medical education.16

Limitations associated with simulation training include the cost of the simulator equipment, instructors, and venue.17 In our case, the relatively high costs were mainly due to the cost of the mannequin, and the small number of sessions per year.

The lack of human emotion provided by the mannequin as well as the lack of environmental confounding factors, which may lead to possible deviations from reality within the controlled setting.18 Lack of adequate physical signs including the quality of sound generated by the mannequin may also account for the lack of realism and may be deceiving for the trainee.19 Trainees may fail to manage the mannequin as a real person and thus omit important practices such as proper history taking, full examination, consent, and use of protective equipment. Consequently, this may lead to unsafe practice, which may be reflected in day-to-day management of patients.19

Conclusion

Foundation doctors expressed overall satisfaction with the quality of the simulation training session and thought it was a positive learning experience. The quality of the equipment was good, the tutors were thought to have created a favorable educational environment and most trainees were willing to participate in future sessions. On the other hand, respondents expressed the opinion that simulation sessions led to a moderate change in their daily clinical practice. The value of undergraduate sessions, increasing sessions to 3–4 per year, and audiovisual prebriefing for candidates naïve to simulation are important suggestions that need evaluation to optimize simulation training.

Acknowledgments

We would like to acknowledge the Malta Foundation Program directors Dr Tonio Piscopo and Mr Kevin Cassar and the head of the postgraduate training program Dr Ray Galea for permission in performing this study. No organization or programs have provided any funding sources for this study.

Author contributions

All authors contributed toward data analysis, drafting, and revising the paper and agree to be accountable for all aspects of the work.

Disclosure

The authors report no conflicts of interest in this work.


References

1.

McGaghie WC, Issenberg SB, Cohen ER, Barsuk JH, Wayne DB. Does simulation-based medical education with deliberate practice yield better results than traditional clinical education? A meta-analytic comparative review of the evidence. Acad Med. 2011;86(6):706–711.

2.

The Foundation Programme Malta; 2015. Available from: http://www.fpmalta.com.

3.

Bradley C. The role of high-fidelity clinical simulation in teaching and learning in the health professions; 2011. Available from: http://www.kcl.ac.uk/study/learningteaching/kli/research/hern/hern-j4/Claire-Bradley-hernjvol4.pdf.

4.

Weaver SJ, Salas E, Lyons R, et al. Simulation-based team training at the sharp end: a qualitative study of simulation-based team training design, implementation, and evaluation in healthcare. J Emerg Trauma Shock. 2010;3(4):369–377.

5.

Perkins GD. Simulation in resuscitation training. Resuscitation. 2007;73(2):202–211.

6.

Wayne DB, Didwania A, Feinglass J, Fudala MJ, Barsuk JH, McGaghie WC. Simulation-based education improves quality of care during cardiac arrest team responses at an academic teaching hospital: a case-control study. Chest. 2008;133(1):56–61.

7.

Crichton F, Joseph S. 0068 Managing the expectations of foundation doctors new to simulation. BMJ STEL. 2014;1(Suppl 1):A46.

8.

Miles S, Kellett J, Leinster SJ. Foundation doctors’ induction experiences. BMC Med Educ. 2015;15:118.

9.

Alinier G, Harwood C, Harwood P, Montague S, Huish E, Ruparelia K. Development of a programme to facilitate interprofessional simulation-based training for final year undergraduate healthcare students. Higher Education Academy/University of Hertfordshire; 2008.

10.

Newby JP, Keast J, Adam WR. Simulation of medical emergencies in dental practice: development and evaluation of an undergraduate training programme. Aust Dent J. 2010;55(4):399–404.

11.

Issenberg SB, McGaghie WC, Petrusa ER, Lee Gordon D, Scalese RJ. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Med Teach. 2005; 27(1):10–28.

12.

Al-Elq AH. Simulation-based medical teaching and learning. J Family Community Med. 2010;17(1):35–40.

13.

McKenna KD, Carhart E, Bercher D, Spain A, Todaro J, Freel J. Simulation Use in Paramedic Education Research (SUPER): a descriptive study. Prehosp Emerg Care. 2015;19(3):432–440.

14.

Cohen ER, Feinglass J, Barsuk JH, et al. Cost savings from reduced catheter-related bloodstream infection after simulation-based education for residents in a medical intensive care unit. Simul Healthc. 2010; 5(2):98–102.

15.

Karakus A, Duran L, Yavuz Y, Altintop L, Caliskan F. Computer-based simulation training in emergency medicine designed in the light of malpractice cases. BMC Med Educ. 2014;14:155.

16.

Grant DJ, Marriage SC. Training using medical simulation. Arch Dis Child. 2012;97(3):255–259.

17.

Okuda Y, Bryson EO, DeMaria S Jr, et al. The utility of simulation in medical education: what is the evidence? Mt Sinai J Med. 2009; 76(4):330–343.

18.

Good ML. Patient simulation for training basic and advanced clinical skills. Medical Educ. 2003;37 (Suppl 1):14–21.

19.

Weller JM, Nestel D, Marshall SD, Brooks PM, Conn JJ. Simulation in clinical teaching and learning. Med J Aust. 2012;196(9):594.

Creative Commons License © 2015 The Author(s). This work is published and licensed by Dove Medical Press Limited. The full terms of this license are available at https://www.dovepress.com/terms.php and incorporate the Creative Commons Attribution - Non Commercial (unported, v3.0) License. By accessing the work you hereby accept the Terms. Non-commercial uses of the work are permitted without any further permission from Dove Medical Press Limited, provided the work is properly attributed. For permission for commercial use of this work, please see paragraphs 4.2 and 5 of our Terms.