Abstract
As part of an ongoing evaluation of our residency program, a needs assessment was performed to assess resident and attending perspectives on current methods of surgical skills training in the operating room.
Participants included obstetrics-gynecology residents and faculty at a university program. Two surveys were developed and validated. Results were analyzed with 2-sample t tests, comparing Likert scores. Findings were significant if the difference between means was >1.
Thirty of 31 residents and 40 of 60 attending physicians responded to the survey. Residents and attending physicians agreed that the surgical skills training program needs improvement (difference in mean, −0.39; confidence interval [CI]: −0.98 to 0.20). The areas of most disagreement were regarding feedback on surgical skills and instrument handling (difference in mean, 2.53; CI: 1.81–3.26, and difference in mean, 2.24; CI: 1.44–3.05).
A significant proportion of surgical skills training during residency occurs as on-the-job training, and operating room time provides a key learning opportunity. This report demonstrates that there is a noteworthy difference in the perception of attending physicians and residents about the quality of teaching and feedback that is currently occurring in the operating room. The difference in perspectives among residents and attending physicians reported in this survey suggests a need for improved communication and systematic feedback in order to capitalize on operating room time as a critical surgical skills training arena.
Background
The earliest training models for surgical skills training were built on an apprenticeship model. In 1928, delegates of the American Medical Association formally endorsed residency training in accordance with this model.1 The concept of residency training came from the German training system and supported teaching surgical skills through observation of an expert in the field, with progression to performance of a specific procedure based on the decision of that expert.1 In the 1990s there was a shift in thinking about optimal approaches to surgical skills training. The introduction of laparoscopy presented a new set of surgical skills that required mastery by both trainees and expert surgeons. As a result, the term expert was called into question, and the best means of training was reevaluated. Whereas expertise had previously been based on professional experience and the opinion of one's peers, there became a new concern for superior performance and reproducibility.2 The Calman Report reiterated these concerns on a national level in 1993, stressing the need for standardization of teaching surgical skills.1
In the evolution of surgical skills training, many different models for skill acquisition have been used to determine the best means of teaching surgical skills. Dreyfus and Dreyfus described stages of skill acquisition, including novice, advanced beginner, competent, proficient, expert, and master.3 This model focuses on the importance of pattern recognition, intuition, and reflection as key elements of skill acquisition.3 Peyton described a 4-stage approach to teaching manual skills. It includes instructor demonstration; instructor deconstruction with breakdown into simple steps; formulation, with the instructor demonstrating the skill as the student recites the steps; and ultimately, performance by the student.1 Perhaps the most widely accepted model for surgical skills training is drawn from studies of teaching pilots, musicians, and athletes through the use of a 3-step model. The first step is the conscious phase, in which each step is thought out. The second step is the shared control phase, in which simple tasks are performed without much thought, while complex tasks still require a step-by-step approach. Finally, in the automatic phase, complex procedures are performed easily.4 These models place importance on repetition of a specific skill critical for both retention and improved performance.1 Ericsson2 labeled “deliberate practice” as one critical element to achieving expertise, along with other important factors such as having a well-defined goal, motivation to improve, feedback, and gradual refinement.
In contemporary medical education, the effort to build on these models has led toward an emphasis on teaching surgical skills outside the operating room. Simulation techniques and surgical skill trainers have proven to be a worthwhile and appropriate practice tool, and use of such training models in resident education has grown.5 The importance of teaching outside the operating room is reinforced by the continuing development of new surgical technologies including robotics, focus on reduction of medical errors, and resident work hour restrictions.6
New teaching tools developed for outside the operating room may complement and improve surgical skills training. Nevertheless, the operating room remains the true forum of performance for a surgeon and is still of primary value as a teaching environment for the resident surgeon.7 It is in the operating room where a surgeon combines decision making with the technical aspects of surgery5; therefore, the goal for superior surgical skills training can only be achieved through teaching improvements both inside and outside the operating room, creating a balance of practice and performance.
We are concerned that the emphasis on training outside the operating room may inadvertently lead to de-emphasis on the importance of intraoperative teaching. In the effort to improve surgical skills training overall, opportunities and techniques to improve teaching within the operating room must also be emphasized, including the development of feedback techniques and learning tools specifically for the operating room. We speculated that the perception of the quantity and quality of intraoperative teaching varied between residents and faculty. Specifically, we wondered whether a disparity exists between faculty and residents regarding perception of preparation for surgery, intraoperative teaching, and feedback. In this context, a needs assessment survey was performed among obstetrics-gynecology residents and attending physicians, with the ultimate goal of using the results to improve surgical skills training in the operating room.
Methods
This study was performed as a quality improvement effort as a part of the Accreditation Council for Graduate Medical Education program requirements. Thirty-two obstetrics-gynecology residents (postgraduate years 1−4) and 60 attending surgical obstetrics-gynecology faculty members were identified as participants in a residency training program of 1 academic institution. Two different surveys were created using the 9-step process suggested by Stone8 to create a useful and valid questionnaire. One survey was designed to elicit the opinions of attending physicians and the other to determine the opinions of residents regarding surgical skills training. Specific questions focused on resident preparation prior to a surgical case, attending physician teaching in the operating room, and feedback from attending physicians. The questions of each survey were structured to ask similar questions from either attending physician's or resident's perspective so they could be compared. Responses were obtained using a Likert scale from 1 to 9 (figure1).
Likert Scale With Definitions, Used by Physicians to Express Their Degree of Agreement
Likert Scale With Definitions, Used by Physicians to Express Their Degree of Agreement
Graphic Representation of Agreement Between Resident and Attending Physicians on Points Relating to Feedback
Graphic Representation of Agreement Between Resident and Attending Physicians on Points Relating to Feedback
The surveys were distributed to residents and faculty using an online survey website service (SurveyMonkey, Menlo Park, CA). The surveys were sent out via e-mail with a short explanation that the survey would be anonymous, and that it would be used in an effort to improve surgical skills training at the institution (figure 2).
Results were analyzed with a 2-sample t test, comparing mean score for attending physicians versus residents. The hypothesis for the t test was that the difference in means was >1. This hypothesis was chosen because of the definition of the categories for the 9-point Likert scale. A difference in response between attending physician and resident of >1, with a 95% confidence interval that did not cross 1, was therefore considered significant.
Results
Thirty-one of 32 residents responded to the survey (97% response rate) and 40 of 60 attending physicians responded (67% response rate).
There were several areas in which residents and attending physician agreed. Both groups proposed that the surgical skills program needs improvement, with an average attending physician score of 6.98 (defined as “agree”) and an average resident score of 7.37. Attending physicians and residents agreed that, in general, they discussed the planned procedure prior to the case (mean score of 6.58 vs 5.83; mean difference, 0.74; confidence interval [CI]: 0.06−1.43) and that attending physicians ask pertinent questions regarding the case (mean difference, 0.91; CI: 0.19–1.63). Finally, attending physicians and residents agreed that attending physicians give feedback specific to each step of the procedure (mean difference, 0.85; CI: 0.10−1.61).
There were 2 questions for which a trend toward disagreement was noted, with the difference in Likert score not reaching statistical significance. These questions were specific to anatomy, and the first question addressed residents' review of anatomy prior to a case (4.88 vs 6.43; mean difference, 1.56; CI: −2.15 to −0.96). Similarly, the trend showed that attending physicians are more likely to agree that they demonstrate anatomy (6.87 vs 5.17; mean difference 1.71; CI: 1.00−2.41) (table1).
There was a disparity in perception of feedback in the operating room. Although attending physicians agreed (7.00) that they that they give frequent feedback regarding surgical skills, residents disagreed (4.47). Similarly, attending physicians agreed that they give specific and constructive feedback during a case (7.08), feedback regarding instrument handling (7.11), formative feedback on what residents did well (6.53), and opportunities for improvement (6.42). Residents, on the other hand, were more likely to disagree with these statements, with mean scores ranging from 4.60 to 4.86. The difference in means between residents and attending physicians for these questions all reach statistical significance, ranging from 1.82 to 2.53, with confidence intervals that are >1 and do not cross 1 (table 2).
Both residents and attending physicians agreed that there is feedback on proper surgical technique (mean t scores of 7.34 and 5.70, respectively), but the difference in means for this question, 1.63, did reach statistical significance (table 3). Finally, although attending physicians were more likely to disagree that most residents review how procedures are done prior to a case (mean score, 4.80), residents were more likely to agree (mean score, 6.90). This difference was statistically significant, with a difference in means of −2.10 (CI: −2.72 to −1.48, table 4).
Discussion
As new teaching tools are developed, it is critical to remember the fundamental value of the operating room as a learning environment for surgeons. In order to create a balanced surgical skills training program, residency programs must continue to strive to improve the communication and feedback between residents and attending physicians in the operating room. This type of one-on-one focused learning in a setting in which both technical skill and judgment are evaluated is crucial to the development of expert surgeons.
Teaching in the operating room must be examined with the same critical eye applied to surgical skills training outside the operating room, and new techniques and models of learning must be incorporated in this setting as well. The survey presented here was therefore undertaken to improve surgical training in the operating room. The goal was to examine areas in which residents' and attending physicians' perceptions of teaching differed. This information can then be used to determine new tools and teaching techniques to improve the learning environment of the operating room.
The findings in this investigation suggest that even though residents and attending physicians agree that productive learning currently occurs in the operating room, there is room for improvement. The areas in which attending physicians and residents disagreed the most were specific to feedback in the operating room. As this is a critical element in adult learning,9 this survey exemplifies the need for tools to improve communication and to better align the perceptions among residents and attending physicians with regards to feedback in the operating room.
This survey is limited by the small number of participants; however, the survey is specific to one particular hospital, and the response rate (particularly from residents) was very high, giving a comprehensive view of practices and perceptions in this department. It is not clear whether the responses obtained may be extrapolated to other programs or are unique to our own program. Furthermore, this survey asks for a general opinion from both residents and attending physicians and does not account for the fact that some attending physicians may be very good at giving feedback and some attending physicians may not be as skilled in this area. Therefore, more specific data may be important when structuring how to improve communication and provide improved feedback on surgical skills. Importantly, procedure-based specialty residency programs may benefit from similar self-evaluations as demonstrated in this report.
Conclusions
Our survey revealed specific areas with room for improvement in teaching in the operating room. The goal of this study was to produce information that could be used in a productive way to improve surgical skills training in the operating room. By recognizing the differences in perceptions between residents and attending physicians, specific education tools and guidelines may be developed in order to bridge the communication gap and increase the feedback to residents in the operating room. By achieving this type of intervention, the program can help to expedite the transition of resident physicians to becoming expert surgeons and to improve the teaching skills of attending physicians.
References
Author notes
Kimberly L. Levinson, MD, MPH, is a Resident Postgraduate Year-3, Johns Hopkins Hospital; Joyce N. Barlin, MD, is a Resident Postgraduate Year-4, Johns Hopkins Hospital; Kristiina Altman, MD, is Assistant Professor of Obstetrics and Gynecology in the Department of Obstetrics and Gynecology, Johns Hopkins Bayview; and Andrew J. Satin, MD, is Chair, Department of Obstetrics and Gynecology, Johns Hopkins Bayview.
Editor's Note: The online version of this article contains survey questions.