Virtual Alternative to the Oral Examination for Emergency Medicine Residents

Author Affiliation
Jillian McGrath, MD The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio
Nicholas Kman, MD The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio
Douglas Danforth, PhD The Ohio State University Wexner Medical Center, Department of Obstetrics & Gynecology, Columbus, Ohio
David P. Bahner, MD The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio
Sorabh Khandelwal, MD The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio
Daniel R. Martin, MD The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio
Rollin Nagel, PhD The Ohio State University College of Medicine, Office of Evaluation, Curriculum Research and Development, Columbus, Ohio
Nicole Verbeck, MPH The Ohio State University College of Medicine, Office of Evaluation, Curriculum Research and Development, Columbus, Ohio
David P. Way, MEd The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio
Richard Nelson, MD The Ohio State University Wexner Medical Center, Department of Emergency Medicine, Columbus, Ohio

Introduction
Methods
Results
Discussion
Limitations
Conclusion

ABSTRACT

Introduction

The oral examination is a traditional method for assessing the developing physician’s medical knowledge, clinical reasoning and interpersonal skills. The typical oral examination is a face-to-face encounter in which examiners quiz examinees on how they would confront a patient case. The advantage of the oral exam is that the examiner can adapt questions to the examinee’s response. The disadvantage is the potential for examiner bias and intimidation. Computer-based virtual simulation technology has been widely used in the gaming industry. We wondered whether virtual simulation could serve as a practical format for delivery of an oral examination. For this project, we compared the attitudes and performance of emergency medicine (EM) residents who took our traditional oral exam to those who took the exam using virtual simulation.

Methods

EM residents (n=35) were randomized to a traditional oral examination format (n=17) or a simulated virtual examination format (n=18) conducted within an immersive learning environment, Second Life (SL). Proctors scored residents using the American Board of Emergency Medicine oral examination assessment instruments, which included execution of critical actions and ratings on eight competency categories (1–8 scale). Study participants were also surveyed about their oral examination experience.

Results

We observed no differences between virtual and traditional groups on critical action scores or scores on eight competency categories. However, we noted moderate effect sizes favoring the Second Life group on the clinical competence score. Examinees from both groups thought that their assessment was realistic, fair, objective, and efficient. Examinees from the virtual group reported a preference for the virtual format and felt that the format was less intimidating.

Conclusion

The virtual simulated oral examination was shown to be a feasible alternative to the traditional oral examination format for assessing EM residents. Virtual environments for oral examinations should continue to be explored, particularly since they offer an inexpensive, more comfortable, yet equally rigorous alternative.

INTRODUCTION

Simulation-based education and assessment strategies have become increasingly popular in medical education. Healthcare simulations are being used in individual and group settings for both formative and summative assessments.1,2 The use of immersive learning environments (ILE) for education provides learners with a sense of being immersed in the simulated environment while experiencing it as real. Partial immersive environments involve a virtual world that consists of three dimensions (3-D) displayed on a two-dimensional (2-D) computer screen.3,4 Educational research using 3-D virtual worlds and their effect on learning outcomes is limited.5 With a predicted paradigm shift in medical education where immersive environments continue to expand in the personal and professional lives of learners, it is imperative to explore and understand the implications and limitations of these ILEs.6

A number of ILEs have been developed in the past ten years with varying rates of adoption by the education community. Second Life (SL) is a virtual 3-D platform that allows individuals from any geographic location to interact in a virtual environment. Accordingly, SL provides opportunities for remote virtual simulation experiences.5,7 In SL, users are represented in a virtual world by their avatars. An avatar is an online, self-created, animated characterization of the user that can act in any role (doctor, patient, nurse, or teacher) and perform programmed tasks (Figure 1). SL has been successfully used in medical and public health education.7,8 Specifically, the platform has been used to model doctor-patient relationships, teach clinical diagnosis, train for disasters, virtually tour the human anatomy, and to conduct physical examinations.4,8-11

wjem-16-336-g001

Figure 1. Avatar patient in an emergency department examination bay.

The American Board of Emergency Medicine (ABEM) administers an oral board examination semiannually to residency-trained emergency medicine (EM) physicians. Passage of the oral board examination is required for EM board certification.12 The purpose of the oral board examination is to assess the candidates’ medical knowledge, clinical reasoning, and interpersonal skills. In the current format, candidates travel to a central assessment venue to take the oral board examination. The examination requires the candidate to verbally explain how they would handle various patient cases to an examiner. Many residency programs offer “mock” or practice oral examinations to prepare residents for the ABEM oral boards. Residents at this academic EM residency program participate in an annual mock oral examination, which is conducted in the traditional format, a face-to-face interaction with an examiner. The purpose of this study was to assess the feasibility of using immersive virtual simulation technology to administer oral examinations to EM residents and to evaluate the potential of this platform as an alternative to the traditional face-to-face oral examination.

METHODS

We used a prospective, stratified-random control group study design to evaluate two methods of administering an oral examination: the traditional face-to-face method and the immersive virtual simulation method. To create a virtual environment we used SL (Second Life 2.0 Viewer, Linden Research, Inc. (Linden Lab), San Francisco, CA). Second Life Viewer is free, open-access computer software; however, fees are required to purchase virtual real estate or to construct virtual environments. We constructed a virtual emergency department (ED) for this study on virtual real estate purchased by one of the authors for another project. Both real estate and building costs were covered by internal institutional grants. The study was conducted at an American, university-based, three-year EM residency training program. Residents at all three levels, program years (PGY) 1–3 were included.

EM residents (n=35) were randomly assigned to one of two groups using a stratified approach to ensure that each group had an equal number of residents from each of three levels of training (PGY1–3). The first group was administered the oral examination using the immersive virtual interface with the examiner at a different physical location than the examinee (Figure 1). The second group served as a control and was administered the oral examination using the traditional format: a face-to-face patient case scenario that was managed with the examiner present in the same room as the examinee. Both groups were administered the same case scenario in which the examinee was expected to diagnose and manage a patient with ST-elevation acute myocardial infarction and resuscitate the patient after cardiac arrest. The study was reviewed and approved by our institution’s behavioral sciences institutional review board.

In the immersive virtual condition, the examinee managed the patient case using the physician avatar to play the role of the physician (Figure 2). The faculty proctor played the role of the patient using the patient avatar. The examinee and proctor were in remote physical locations and communicated via headset and computer. Resident-examinees verbally interviewed the patient-proctor for historical details and physical exam findings. A collection of pertinent diagnostic data was created in PowerPoint and subsequently loaded into an image viewer in the immersive virtual environment. The faculty proctor controlled the image viewer, allowing diagnostic data (initial and repeat vital signs, laboratory reports, and diagnostic imaging) to be displayed in the virtual examination room in real time when requested by the examinee or at appropriate times during the case (Figure 2). Identical images were printed on paper and offered in similar sequence for the traditional oral exam format. Two faculty proctors administered all virtual examinations and two other faculty proctors administered all traditional oral exams. Access to a video demonstration of the virtual examination can be found at http://vimeo.com/user29472626/videos (Password = OSUEMSL2, case sensitive).

wjem-16-336-g002

Figure 2. Physician (examinee) avatar examining a patient in an emergency department examination bay after requesting chest radiograph.

We used the ABEM instruments for scoring resident performance and documenting execution of “critical actions” on both the virtual and traditional oral examination conditions. Using this instrumentation, proctors scored examinees on eight performance items using an 8-point rating scale. The items on the instrument represent the eight ABEM competency categories. Proctors also used the ABEM checklist to document whether the examinee executed 10 “critical actions” during their work on the case. Traditional and virtual groups were compared on: the number of critical actions they executed, their original scores on the eight performance items, and on passage of the performance items as defined by ABEM. The ABEM standard for passing is a composite score greater than or equal to 5.75 on the 8-point scale. For our purposes, all performance items were dichotomized into pass-fail variables to compare groups on pass-rates across each competency category.

Participants were surveyed regarding their opinions about the oral examination experience. The immersive virtual group received a survey comprised of 13 items that used a 5-point Likert response set (from 1=Strongly Disagree to 5= Strongly Agree). Questions inquired about past experiences with the format, any difficulties encountered during the exams, and format preference for the virtual group. Questions further elicited the level of perceived realism, objectivity, efficiency and intimidation during the examinations. The traditional oral examination group received a shorter 6-item survey comprised of questions designed to compare their experience with that of the virtual group.

We used Fisher’s exact tests for analyzing 2 × 2 tables to compare the groups on number of critical actions executed and on pass rates for each of the eight examination items (ABEM competencies and overall clinical competence). Independent t-tests were used to compare the groups on mean item scores for each of the eight items. We used Bonferroni corrections to control for family-wise Type 1 error rates for each set of multiple comparisons (critical action and pass-fail comparisons, and comparisons between competency category scores).13

RESULTS

Fisher’s exact test showed no significant differences between traditional and virtual examinees in the number of critical actions executed (Table 1) and showed no significant difference in pass rates between traditional and virtual examinees (Table 2). We compared groups on examination scores for the ABEM’s individual competency categories and found no significant differences (Table 3). However, moderate effect sizes were observed for many of the ABEM competency categories, with all but two mean differences favoring the virtual examination group (data acquisition and interpersonal relations competencies). The assessment results observed during this study were consistent with results of mock oral examinations administered in prior years.

Table 1. Frequencies, (percentages), and Fisher’s exact test value for 17 Traditional Oral Exam Group residents and 18 Immersive Virtual Exam Group residents on execution of 10 critical actions during an oral examination case.

table 1

CPR, cardiopulmonary resuscitation; Vtach, ventricular tachycardia; Vfib, ventricular fibrillation
*A family-wise Bonferroni correction was used to control for Type I error rates (finding significant differences by chance). The corrected p-value considered for statistical significance is equal to 0.005.
**Critical actions were documented by the proctors using a checklist during the examination.
Table 2. Frequencies, (percentages), and Fisher’s exact test value for 17 Traditional Oral Exam Group residents and 18 Immersive Virtual Exam Group residents on passing or failing 8 American Board of Emergency Medicine competency categories.
 table 2*A family-wise Bonferroni correction was used to control for Type I error rates (finding significant differences by chance). The corrected p-value considered for statistical significance is equal to 0.006.
**A score of 5.75 or greater was required for passing each competency category.
Table 3. Means, (standard deviations), independent t-test results, and effect sizes for 17 Traditional Oral Exam Group residents and 18 Immersive Virtual Exam Group residents on 8 American Board of Emergency Medicine competency category scores.
table 3
*A family-wise Bonferroni correction was used to control for Type I error rates (finding significant differences by chance). The corrected p-value considered for statistical significance is equal to 0.006.
**Scores were assigned by proctors using a standard ABEM 1-8 scale.

The mock oral examination is a program requirement for our residency program. Accordingly, all examinees, regardless of training level, had experience with the traditional face-to-face oral examination format prior to this study. Most of the examinees (57%) had never used SL prior to the examination. Examinees reported that both formats were realistic (Traditional 80% vs. Virtual 86%). All examinees perceived the examinations to be fair, objective, and efficient in either format. None of the examinees in the virtual group found the examination to be intimidating (Traditional 40% vs. Virtual 0%), and many reported that the virtual format was less intimidating than traditional oral exams that they had experienced in the past (77%). Most of the virtual examinees reported a preference for the virtual examination (79% agree and 21% neutral) over the traditional format (Table 4).

Table 4. Means, (standard deviations), independent t-test results, and effect sizes for 10 Traditional Oral Exam Group residents and 14 Immersive Virtual Exam Group residents on six post-examination evaluation items over the interface they experienced: Immersive Environment or Traditional Face-to-Face interface with a proctor.

table 4

 

*A family-wise Bonferroni correction was used to control for Type I error rates (finding significant differences by chance). The corrected p-value considered for statistical significance is equal to 0.006.
**Option key: 1= Strongly Disagree, 2= Disagree, 3= Neutral, 4= Agree, 5= Strongly Agree.

DISCUSSION

Criticisms of the traditional oral examination process have been raised by EM residents, including recent residency graduates preparing for the ABEM oral examination.14 These criticisms can be classified into three domains or issues; fidelity, validity, and logistics.

The first issue has to do with the oral examination fidelity. While the goal of the ABEM oral examination has been to replicate a realistic ED encounter; the traditional oral examination format, with its face-to-face questioning process, remains somewhat artificial. Also related to fidelity is the manner in which the patient information is communicated to the examinee. Until recently, ABEM examinees have been unable to visualize their patient during the test. Additionally, physical examination findings and diagnostic imaging were presented on paper, rather than in the medium in which it would be encountered in a real ED. To address this issue, ABEM has recently committed to incorporating some computer-based images into their oral examination (eOrals); however, the specific details of this change have yet to be revealed.12

A second issue involves the validity of the oral examination format. Threats to validity involve both the effects of performance anxiety and examiner bias. Performance anxiety can occur when an examinee is confronted with an unfamiliar examiner during a face-to-face encounter in an unfamiliar environment. Furthermore, despite formal scoring systems and examiner training, examiner bias also remains a threat to the validity of the traditional oral exam format.15 The literature suggests that bias is common towards candidates with good interpersonal skills, good communication skills, and those who are physically attractive.16 Finally, a third issue with the traditional ABEM oral examination format involves logistics, such as the expense of preparation and travel to the testing site.15

The virtual examination offers many potential benefits to address the issues involved with the traditional oral examination. The virtual examination offers higher fidelity realism than the traditional oral exam by immersing the examinee in a setting resembling one in which actual patient care is delivered. The virtual examination involves dynamic interaction between the examinee as physician and the examiner as patient. Patient information, such as vital signs and diagnostic imaging results, are presented in a more authentic manner in the virtual environment.

Threats to examination validity are also minimized in the virtual examination format. The potential for anxiety produced by a one-on-one encounter with a stranger is alleviated by the virtual world encounter. Examiner biases resulting from the personal encounter with the examinee are also minimized.3 Finally, because the virtual examination can be administered through the electronic medium of the world-wide web, logistic concerns can be addressed. Delivering the oral exam through a virtual reality platform eliminates the need for examinees and examiners to travel, providing economic and time savings.

The results of this study offer confirmation that virtual examination results are comparable to those of the traditional oral examinations for assessing EM residents. In fact, we observed moderate effect sizes favoring the SL group, even with relatively small samples, on five of the eight ABEM competencies; suggesting that with bigger samples we might have demonstrated significantly better performance on these competencies through the virtual examination platform.

One can envision the application of advanced virtual simulation technology as a way to alleviate some of the barriers encountered in the current process. In addition to reported ease of use and perception by many that this was a more realistic experience, the virtual examination format is adaptable. The virtual oral examination could be administered from any remote location with computer access, at any time of day. Thus, oral examinations could be completed while on away rotations, while travelling, at home rather than in an office setting, or at a remote testing site.

The implementation of virtual technology in resident assessment required a time commitment for brief training of faculty to use the system to administer the examinations. Direct costs included purchasing space in the virtual world and costs for building the desired assessment environment.17 Examiners of the immersive virtual oral examination required 2–3 hours of training in order to develop and monitor the data display. Communication via microphone and computer did not require specific training for the faculty or residents participating in the examinations. Some examinees reported feedback or echoing in the headset. Such impediments can be eliminated through use of higher quality microphones, headsets and computer systems.

Resident feedback regarding the use of the immersive environment for the oral examination in EM was overwhelmingly positive. Many examinees expressed an interest in even more advanced capability to interact with the virtual patient and within the examination room. With currently available animation and programming capabilities, items in the room can be made more interactive. The examinee might click on the patient-avatar’s body to perform physical examination skills or a virtual IV pole to order IV fluids. They might also instruct a nurse avatar to perform programmed tasks. The virtual environment could be expanded to a multi-case format, requiring concurrent care of multiple patients requiring the examinee-physician avatar to transition between patient rooms. More complex immersive assessment environments may offer higher fidelity assessment potential without additional cost or barriers to ease of use. Transition to an automated scenario, without a real-life examiner, could be achieved through application of artificial intelligence. The Unity 3-D platform is an example of another immersive environment that is ideally suited to support the virtual assessment format as it offers higher levels of fidelity and is reported to be easier to use by both examinees and proctors.18 In addition, it can be highly customized and configured to provide secure examinations without requiring the type of third party support which is necessary with platforms such as Second Life.

LIMITATIONS

This study was conducted at a single academic training site and therefore the number of assessment subjects was not large. Our intent was to evaluate the feasibility and value of the virtual examination; however, because each resident experienced only one examination format, they were unable to preference one format over the other. Prior experience with the traditional face-to-face oral examination made it possible for the immersive exam participants to contrast the virtual with the traditional oral exam experience. As voices were not modified in this virtual examination, there is the potential for bias to be introduced based on voice recognition of the examinee or examiner. This is a bias that could be avoided in a larger-scale examination format using anonymous proctors, technology to modify or standardize the proctor’s voice, or creation of an automated case using artificial intelligence. Because only one examiner evaluated each examinee, inter-rater reliability was not evaluated; therefore, inter-rater reliability remains an issue to be studied in future research. Other aspects of testing via virtual simulation require additional exploration before such technology is more broadly adopted for general use in a high-stakes oral board examination. Further research is needed to study faculty perceptions about the virtual examination experience or evaluate the reproducibility of results using this platform.

CONCLUSION

The virtual simulated oral examination is a feasible alternative to the traditional oral examination format for EM residents. This study used Second Life as a platform for the virtual examination; however, we believe that other immersive learning environments should be evaluated. Future studies should focus on identifying and developing the most user-friendly platforms for virtual oral examination and continue to assess applications of virtual examination in other areas of medical education.

Footnotes

Section Editor: Michael Epter, DO

Full text available through open access at http://escholarship.org/uc/uciem_westjem

Address for Correspondence: Jillian McGrath, MD, The Ohio State University Wexner Medical Center, 760 Prior Hall, 376 W. 10th Avenue, Columbus, OH 43210-1238. Email: jillian.mcgrath@osumc.edu. 3 / 2015; 16:336 – 343

Submission history: Revision received October 22, 2014; Submitted January 5, 2015; Accepted January 6, 2015

Conflicts of Interest: By the WestJEM article submission agreement, all authors are required to disclose all affiliations, funding sources and financial or management relationships that could be perceived as potential sources of bias. The authors disclosed none. The authors have no affiliation or financial agreement with Second Life or Linden Research, Inc.

REFERENCES

1. Vozenilek J, Huff JS, Reznek M, et al. See one, do one, teach one: Advanced technology in medical education. Acad Emer Med. 2004;11:1149-1154.

2. McGaghie WC. Use of simulation to assess competence and improve healthcare. Med Sci Educ. 2011;21(3S):261-363.

3. Immersive Learning Training and Education White Paper. Daden Limited. Avaiable at: www.daden.co.uk/resources/white-papers/. Accessed Jan 6, 2014

4. Andreatta PB, Maslowski E, Petty S, et al. Virtual reality triage training provides a viable solution for disaster-preparedness. Acad Emer Med. 2010;17(8):870-876.

5. Hansen M. Versatile, immersive, creative and dynamic virtual 3-D healthcare learning environments: A review of the literature. J Med Internet Res. 2008;10:e26.

6. Taekman JM, Shelley K. Virtual environments in healthcare: Immersion, disruption, and flow. Int Anesthesiol Clin. 2010;48(3):101-121.

7. Boulos M, Kamel N, Hetherington L, et al. Second life: an overview of the potential of 3-D virtual worlds in medical and health education. Health Info Libr J. 2007;24:233-245.

8. Schwaab (McGrath) J, Kman N, Nagel R, et al. Using second life virtual simulation environment for mock oral emergency medicine examination. Acad Emer Med. 2011;18(5):559-562.

9. Danforth D, Procter M, Heller R, et al. Development of virtual patient simulations for medical education. J Virtual World Res. 2009;2(2). Avaiable at: https://journals.tdl.org/jvwr/article/view/707/503. Accessed Jun 24, 2010

10. Simon S, Avatar II. The hospital. 2010. Avaiable at: www.wsj.com.

11. Yellowlees P, Cook J, Marks S, et al. Can virtual reality be used to conduct mass prophylaxis clinic training? A pilot program. Biosecur and Bioterror. 2008;6:36-44.

12. Examination information for candidates: Oral certification Exam. Avaiable at: www.abem.org. Accessed Apr 14, 2014

13. Bland JM, Altman DG. Multiple significance tests: The Bonferroni method. BMJ. 1995;310(21):170.

14. Bianchi L, Gallagher EJ, Horte R, et al. Interexaminer agreement on the American Board of Emergency Medicine Oral Certification Examination. Ann Emerg Med. 2003;41:859-64.

15. Platts-Mills TF, Lewin MR, Madsen T. The oral certification examination. Ann Emerg Med. 2006;47(3):278-282.

16. Lunz ME, Bashook PG. Relationship between candidate communication ability and oral certification examination scores. Med Educ. 2008;42:1227-1233.

17. Second Life Viewer (Version 2.0.1) [Software]. 2010. Avaiable at: http://secondlife.com/support/downloads/?lang=en-US.

18. Unity (Version 4.5) [Software]. 2014. Avaiable at: http://unity3D.com/unity/download.