Background Although Clinical Competency Committees (CCCs) were implemented to facilitate the goals of competency-based medical education, implementation has been variable, and we do not know if and how these committees affected programs and assessment in graduate medical education (GME).

Objective To explore the roles CCCs fulfill in GME and their effect on trainees, faculty, and programs.

Methods We conducted a narrative review of CCC primary research with the following inclusion criteria: all articles must be research in nature, focused on GME and specifically studying CCCs, and published in English language journals from January 2013 to November 2022.

Results The main results are as follows: (1) The primary role of the CCC (decision-making on trainee progress) is mostly described in “snapshots” (ie, focusing on a single aspect of this role at a single point in time); (2) CCCs are taking on secondary roles, some of which were anticipated (eg, remediation, feedback) whereas others were “unanticipated” (eg, use of CCC data to validate trainee self-assessment, predict trainee performance in other settings such as certifying examinations, investigate gender bias in assessment); and (3) Articles briefly mentioned short-term outcomes of CCCs at the level of the trainees, faculty, and programs. However, most studies described interventions to aid CCC work and did not specifically aim at investigating short-term (eg, curriculum changes) or long-term outcomes (eg, improved patient outcomes).

Conclusions CCCs fulfill a range of roles in assessment beyond their intended purpose. A more systematic approach is needed to investigate the outcomes of CCC implementation on GME.

The primary role of Clinical Competency Committees (CCCs) in graduate medical education (GME) is to render judgements about trainees’ performance against a set of criteria and standards (eg, competency Milestones in the United States) and to make recommendations to program directors about advancement.1  In addition to this primary role, there has been a recognition of evolving secondary roles that these committees can play in residency training.2  Although neither a mandate nor rarely made explicit, these secondary roles result from CCCs being privy to “seeing” and synthesizing all their trainees’ assessment data and may include remediating trainees, critiquing the quantity and quality of assessment data, providing faculty development, identifying curricular gaps and redundancies, and serving a quality improvement function for the assessment system.2  To date, however, there has been no synthesis of the existing literature to describe roles actually fulfilled by CCCs and their effect on GME and assessment systems in particular. Without such exploration, we risk not providing CCC members with the time allotment, faculty development, and administrative support needed, thus limiting their ability to ensure that trainees are ready for graduation and capable of providing safe and effective care. Therefore, the purpose of this narrative review is to provide an overview of the published research on the roles CCCs fulfill in GME and their effect on residency training and assessment.

We chose to conduct a narrative review because they are particularly useful for exploring topics that are complex and under-studied, and they allow the inclusion of a wide variety of studies.3  Our approach was informed by the recommendations published by Sukhera, which offer guidance on key decisions in the review process.3  According to Sukhera, narrative reviews are (1) “flexible” in which the initial scope may change through the review process,” (2) well suited for topics that are “under-researched,” (3) “require a meaningful synthesis of research evidence that may be complex or broad,” and (4) “require detailed, nuanced description and interpretation.”3  Our study met each of these criteria. First, although we started the review broadly (eg, including perspectives and policy documents), we realized that we needed to focus on research articles to gather data about how CCCs are actually being implemented and studied. Second, our focus on the roles of CCCs and their associated outcomes is indeed an “under-researched” area in GME. Third, the CCC literature to date is complex (ie, represented by varying implementation approaches).4  Fourth, we provided a “nuanced”3  analysis of the descriptions offered in the literature of CCCs as detailed below.

Search Strategy

To our knowledge, the first mandate for CCCs was put forth in January 2013, when the Next Accreditation System went into effect. Thus, we searched PubMed, Scopus, ERIC, and Google Scholar databases from January 1, 2013 to November 1, 2022. During the study, the search was updated twice because there was a delay in our process due to the COVID-19 pandemic and the emergence of pertinent CCC literature. We used “Clinical Competency Committee” as a search term to identify publications where the CCC was mentioned in the title and/or abstract. The search strategy was conducted under the guidance of a librarian at the leading author’s institution.

Publication Selection Process

The Figure summarizes the search process. One author (A.E.) performed a preliminary review of the search results. One hundred seventeen articles were initially identified and, after removal of duplicates, 111 were selected for “title and abstract review.” Two researchers (A.E. and either M.G. or S.H.) independently screened 60 abstracts for inclusion criteria and subsequently resolved discrepancies through discussion. This approach proved to be critical to the richness of our discussions centered on developing definitions for some of the key concepts involved in conceptualizing our work, such as “secondary roles” and “impacts.” A.E. subsequently reviewed all remaining abstracts. Following full-text review of selected abstracts, ultimately, 84 articles were included for data extraction.5-88 

Figure

Flowchart of the Literature Search and Study Selection Process

Abbreviation: CCC, Clinical Competency Committee.

Figure

Flowchart of the Literature Search and Study Selection Process

Abbreviation: CCC, Clinical Competency Committee.

Close modal

Inclusion/Exclusion Criteria

Inclusion criteria were the following: all articles must be research in nature, focused on GME and specifically studying CCCs (ie, the term CCC was used in the abstract and/or title), and published in English language journals from January 2013 to November 2022. Articles were excluded if they focused on other decision-making groups such as education committees, were written as perspectives, or situated in undergraduate medical education. The authors discussed these criteria at length, opting to focus the review on only GME, as the process for using CCCs in undergraduate medical education is still evolving.

Data Extraction

A data extraction form was created in Microsoft Excel. The categories (codes) for data extraction included demographic program information, description of the primary and secondary role(s) of CCCs, and the effect of these roles at various levels of the assessment system and context (eg, on the trainee, training program, specialty). Using the preliminary coding framework, 3 authors (A.E., M.G., S.H.) independently reviewed 14 randomly selected articles. Preliminary findings were discussed in the entire research team (A.E., M.G., S.H., E.H.) to resolve any discrepancies and to further refine the coding framework. Using the final coding framework, A.E. subsequently reviewed the remaining articles, while the other authors (M.G., S.H., E.H.) “spot checked” the extracted data for accuracy and completeness. Any articles co-authored by one or more members of the research team were reviewed independently by a colleague (Lisa Conforti or Raghdah Al-Bualy) with content expertise in CCCs. Table 1 presents a taxonomy of themes capturing both primary and secondary CCC roles.

Table 1

Summary of the Articles Included in This Review

Summary of the Articles Included in This Review
Summary of the Articles Included in This Review

We were aware of the posited secondary roles mentioned in the CCC guidebook,2  but we did not know if this would capture the breadth of what was published. Thus, in our taxonomy, we differentiated “anticipated” secondary roles (ie, those that had been mentioned as theoretically possible in the CCC guidebook2 ) versus those that we had not already seen described (“unanticipated”). We used the competency-based medical education (CBME) logic model developed by Van Melle et al to define effects of CCCs on GME training.89  A logic model is a visual representation of how components of a process interact to effect short- and long-term outcomes.89  Thus, changes in the assessment system (eg, new rotations, assessment tools) or actions and attitudes of stakeholders resulting from CCCs fulfilling primary and secondary roles were identified as effects or outcomes. We furthermore characterized outcomes as “short-term” (eg, changes that can be measured in 1 or 2 CCC cycles, for example, residents using CCC feedback to create individualized learning plans) versus “long-term” (eg, those changes that take a much longer time to identify and measure such as linking trainee learning to improved patient outcomes or determining whether CCC ratings influence revisions to the Milestones).

Reflexivity

We had ongoing discussions to determine how our a priori definitions and assumptions of CCC roles and experiential knowledge might influence our interpretation of the data. All authors have published under the broad heading of assessment. A.E. works for the Accreditation Council for Graduate Medical Education (ACGME) and has contributed to the CCC literature. E.H. worked for the ACGME at the time of the study and has contributed to the Milestones and CCC literature. M.G. has been involved in design of assessments in which decision-making relies on CCCs. She also has experiential knowledge of how CCCs work and how they can affect assessment quality. S.H. has been involved in the design and execution of programs of assessment and curricula of undergraduate medical education in which decision-making is done by a committee, and she has experiential knowledge on quality assurance of assessment programs.

Table 2 presents an overview of the demographic characteristics of the studies included in our review. Evaluation of the quality of the studies included in our review identified most studies to be original research published in peer-reviewed journals, focusing on specific elements in the process of CCC implementation and decision-making. The majority were single-institution, single-specialty studies conducted in the United States involving CCC review of trainees at the residency level. See Table 1 for a brief description of each of the studies included in the review.

Table 2

Demographic Data of the CCC Literature

Demographic Data of the CCC Literature
Demographic Data of the CCC Literature

CCC Primary Role and Potential Associated Effect

The primary role of the CCC is to assess trainees and make recommendations to program directors regarding developmental progress.1  We found that most articles did not describe the primary role in its entirety (ie, from review of the assessment data to making a judgement and communicating it to programs and trainees). Rather, the focus was on various aspects such as trying to make sense of difficult data,5  the use of undocumented assessment data,6  role conceptualization,7  or implementation of new assessment tools.8  See data for themes and examples. Some articles focused on the process of Milestones ratings and patterns in performance scoring.9-16  An ethnography of CCCs furthermore revealed marked variability in the approaches CCCs use to conduct their work, such as the strategies used in decision-making and types of assessment data used.4  We also found little inquiry as to how decisions about trainee performance are actually made. The CCC process (before, during, and after meetings) as described by Al-Bualy and colleagues outlines multiple steps during CCC work.17  Our review reveals that still very little is known about how these steps are actually performed.

Although we did not find any studies that explicitly aimed to investigate the effect of CCC implementation on programs and assessment systems, there were multiple examples of “CCC-inspired changes” aimed to facilitate CCC work such as instituting new (1) curricula,18  (2) assessments,19-23  (3) ways of organizing the assessment data,24-29  and (4) assessors.30,31  Beyond these CCC-inspired changes, there were also some mentions of “short-term” outcomes involving the trainees, program, or CCC itself. For instance, at the level of the trainees, one program implemented a web-based direct observation assessment tool with 38% of residents subsequently reporting more immediate feedback on their operative skills.19  At the level of the program, Mamtani et al describe development and implementation of a curriculum and instructional strategies to support assessment of Patient Safety and Quality Improvement Milestones.18  Similarly, Pack and colleagues describe CCCs providing training programs with input for development of new learning experiences, based on what CCCs count as evidence for their decision-making and to fill gaps within the assessment system.32 

Lastly, there were a number of instances of short-term outcomes on the CCC itself. In some cases, authors were able to report a resultant change to their process. For instance, the score cards used in one general surgery program to consolidate all relevant assessment data to a single page card reduced the length of the CCC meeting from 126 minutes to 106 minutes, yet the time available to actually discuss each resident increased from 1 minute prior to using the score cards to 5 minutes after implementation.28  Nabors and colleagues described reduced deliberation time during CCC meetings following a change in the assessment process (ie, CCC members assigning their ratings prior to the CCC meeting).33  Other studies reported the effect of including allied health professionals as assessors on the robustness of CCC decision-making. For example, Bedy and colleagues surveyed emergency medicine CCC members and found that they perceived pharmacists’ assessments as useful to their judgement of resident performance.31  In another study, nurses’ end-of-rotation assessments of emergency medicine residents did not correlate with final CCC proficiency levels; however, their descriptive comments were thought to be “invaluable” for identifying areas for improvement.30  Pack and colleagues found that through the CCCs struggle with problematic evidence, members unpacked their assumptions about the source of the data used and reflected on how they perceive certain types of data as useful evidence.5  All of these examples suggest the potential effect of CCCs at multiple levels in residency training and assessment, not only on trainees, curricula, and assessment programs but also on the CCC members themselves, for example through development of shared mental models around decision-making processes, including usefulness of assessment data.

CCC Secondary Roles and Potential Associated Effect

We found only one study that was specifically designed to explore CCC secondary roles.32  In this study by Pack and colleagues, CCCs deliberately contributed to ongoing evaluation of the curriculum and assessment in residency training and provided meaningful feedback to program directors about program limitations, thereby documenting an impact on CCC members, other faculty, and the program.32  As the vast majority of articles included in our review were not designed to explore CCC secondary roles, we extrapolated the secondary role of the CCC based on the narrative provided by the authors. For instance, some CCCs went beyond identifying struggling residents to assume other responsibilities in the remediation process. Warburton and colleagues describe their CCC not only identifying struggling learners but also referring them to the Early Intervention Remediation Committee.34  In another setting, CCC members helped create individualized learning plans and monitored residents’ progress.35  In these descriptions of secondary roles, the associated outcomes were not reported.

In addition to these “anticipated” secondary roles, we also found a myriad of secondary roles which had not been ascribed to CCCs previously and denoted these as “unanticipated” secondary roles (see Table 1 for themes and examples). The very presence of a CCC, and the data generated by these committees, allowed for “CCC data facilitated research” (ie, use of CCC data to investigate a variety of assessment-related topics). Several studies, for example, investigated the correlation between resident self-assessment and CCC scores.36-42  In various specialties and settings, CCC data was used to check for gender-based differences in Milestones ratings43-47  (ie, for evidence of gender bias in assessment). In those studies that did not find any significant gender-based differences,43-46  some authors hypothesized that CCCs may serve to neutralize biases in the assessments they receive.46  CCC data were also used to investigate correlations between Milestone ratings and performance in other settings, such as certifying or in-training examinations.48,49  Lastly, CCC data was used to support the evaluation of curricular activities or changes. For instance, Frey-Vogel and colleagues compared pediatric interns’ performance scores in a simulated setting with CCC performance ratings.50  In all of the examples mentioned above, CCC data was seen as an instrumental resource for investigating a variety of assessment-related questions that arise in GME.

To our knowledge this is the first narrative review of the CCC literature focusing on CCC roles and the effect of these committees on residency training and assessment. Our work revealed 3 key findings: (1) most CCC studies address components of the primary role but do not explore the entire process; (2) CCCs fulfil secondary roles that we did not anticipate; and (3) despite the myriad of short-term “CCC-inspired changes,” there is a lack of documented long-term outcomes.

By revealing the lack of a comprehensive approach to studying their primary and secondary roles, our review illustrates the complex nature of the work of CCCs. Not only are we still trying to understand how CCCs accomplish their intended role, but our findings also suggest their actual role includes a broad range of secondary roles, some of which are unanticipated. Thus, our review illuminates the breadth of roles CCC play in GME assessment systems. With a clearer understanding of the roles these committees actually play, programs are in a better position to advocate for resources to support CCC efforts. A more systematic approach to investigating CCC processes may not only assist CCCs in developing their expertise, but also provide meaningful input into how assessment systems can be improved to ensure fair and transparent assessment of learning as well as enhance assessment for learning (feedback and learning opportunities for trainees), contributing to achievement of overarching goals in CBME.

CCC performance ratings have been used as a source of validity evidence in the context of self-assessments, exploration of bias in assessment decisions, and prediction of performance in other settings—an unanticipated role. This seems to suggest that CCC ratings are considered to be a “gold standard.” Although findings from these studies may help critically analyze and improve assessment practices in specific settings, more work is needed to justify this approach. Further research into both anticipated and unanticipated secondary CCC roles can help training programs leverage the data they are already collecting to evaluate their practices and make improvements to their education and assessment systems. Therefore, we want to emphasize that even though secondary CCC roles (both anticipated and unanticipated) were not the original intent, they should not be minimized. More work is needed to determine the full scope of these roles and their effects on assessment systems, and we call for a collaborative national research agenda centered on studying specific CCC roles and their associated short- and long-term outcomes.

We also need to determine how implementation of CCCs contributes to achievement of intended outcomes in CBME (ie, graduating trainees who can provide safe and effective care). Although we found multiple examples of “CCC-inspired changes” or “outputs” such as new assessment tools, learning experiences, and data organization platforms, we did not find any studies that focused on investigation of long-term outcomes (eg, a change in feedback culture). To better understand if and how implementation of CCCs contributes to CBME goals, we therefore recommend that future studies consider using the logic model to study CCC impact, both short- and long-term, and intended as well as unintended outcomes.

There are a number of limitations to this work. Our review is limited to English-language studies only. Since studying the roles and associated outcomes of CCCs was not the explicit aim of most of the articles we reviewed, it is possible that we made inferences without fully understanding the contexts described. Although our taxonomy can serve as a first step for conceptualizing CCC processes and effects, it will need to be revisited and revised as future work is performed. Most of the studies were conducted in the United States and it is possible that our findings and conclusions may not be generalizable to other contexts. Thus, future studies should investigate CCC roles and associated outcomes in other contexts.

Although this narrative review identified broader CCC roles than anticipated, there were significant gaps in the literature regarding descriptions of these roles and their associated effects. The lack of articles specifically focused on investigating the outcomes of implementing CCCs in GME is a key finding and a launching point for future work.

The authors would like to thank Lisa Conforti, Raghdah Al-Bualy, Laura Edgar, and Lorelei Lingard.

1. 
Accreditation Council for Graduate Medical Education
.
ACGME Common Program Requirements (Residency)
.
2. 
Andolsek
K,
Padmore
J,
Hauer
K,
Ekpenyong
A,
Edgar
L,
Holmboe
E.
Clinical Competency Committees: A Guidebook for Programs
. 3rd ed.
Accreditation Council for Graduate Medical Education
;
2020
.
3. 
Sukhera
J.
Narrative reviews: flexible, rigorous, and practical
.
J Grad Med Educ
.
2022
;
14
(
4
):
414
-
417
.
4. 
Ekpenyong
A,
Edgar
L,
Wilkerson
L,
Holmboe
ES.
A multispecialty ethnographic study of clinical competency committees (CCCs)
.
Med Teach
.
2022
;
44
(
11
):
1228
-
1236
.
5. 
Pack
R,
Lingard
L,
Watling
CJ,
Chahine
S,
Cristancho
SM.
Some assembly required: tracing the interpretative work of clinical competency committees
.
Med Educ
.
2019
;
53
(
7
):
723
-
734
.
6. 
Tam
J,
Wadhwa
A,
Martimianakis
MA,
Fernando
O,
Regehr
G.
The role of previously undocumented data in the assessment of medical trainees in clinical competency committees
.
Perspect Med Educ
.
2020
;
9
(
5
):
286
-
293
.
7. 
Hauer
KE,
Chesluk
B,
Iobst
W,
et al.
Reviewing residents’ competence: a qualitative study of the role of clinical competency committees in performance assessment
.
Acad Med
.
2015
;
90
(
8
):
1084
-
1092
.
8. 
Mount
CA,
Short
PA,
Mount
GR,
Schofield
CM.
An end-of-year oral examination for internal medicine residents: an assessment tool for the clinical competency committee
.
J Grad Med Educ
.
2014
;
6
(
3
):
551
-
554
.
9. 
Beeson
MS,
Hamstra
SJ,
Barton
MA,
et al.
Straight line scoring by clinical competency committees using emergency medicine milestones
.
J Grad Med Educ
.
2017
;
9
(
6
):
716
-
720
.
10. 
Schwartz
A,
Balmer
DF,
Borman-Shoap
E,
et al.
Shared mental models among clinical competency committees in the context of time-variable, competency-based advancement to residency
.
Acad Med
.
2020
;
95
(suppl 11 Association of American Medical Colleges Learn Serve Lead: Proceedings of the 59th Annual Research in Medical Education Presentations):
95
-
102
.
11. 
Schumacher
DJ,
Bartlett
KW,
Elliott
SP,
et al.
Milestone ratings and supervisory role categorizations swim together, but is the water muddy?
Acad Pediatr
.
2019
;
19
(
2
):
144
-
151
.
12. 
Schumacher
DJ,
West
DC,
Schwartz
A,
et al.
Longitudinal assessment of resident performance using entrustable professional activities
.
JAMA Netw Open
.
2020
;
3
(
1
):
e1919316
.
13. 
Dehon
E,
Jones
J,
Puskarich
M,
Sandifer
JP,
Sikes
K.
Use of emergency medicine milestones as items on end-of-shift evaluations results in overestimates of residents’ proficiency level
.
J Grad Med Educ
.
2015
;
7
(
2
):
192
-
196
.
14. 
Schumacher
DJ,
King
B,
Barnes
MM,
et al.
Influence of clinical competency committee review process on summative resident assessment decisions
.
J Grad Med Educ
.
2018
;
10
(
4
):
429
-
437
.
15. 
Hamstra
SJ,
Yamazaki
K,
Barton
MA,
Santen
SA,
Beeson
MS,
Holmboe
ES.
A national study of longitudinal consistency in ACGME milestone ratings by clinical competency committees: exploring an aspect of validity in the assessment of residents’ competence
.
Acad Med
.
2019
;
94
(
10
):
1522
-
1531
.
16. 
Kelleher
M,
Kinnear
B,
Sall
DR,
et al.
Warnings in early narrative assessment that might predict performance in residency: signal from an internal medicine residency program
.
Perspect Med Educ
.
2021
;
10
(
6
):
334
-
340
.
17. 
Al-Bualy
R,
Ekpenyong
A,
Holmboe
E.
Effective clinical competency committee meeting practices: before, during, and after
.
Acad Med
.
2023
;
98
(
11
):
1340
.
18. 
Mamtani
M,
Scott
KR,
DeRoos
FJ,
Conlon
LW.
Assessing EM patient safety and quality improvement milestones using a novel debate format
.
West J Emerg Med
.
2015
;
16
(
6
):
943
-
946
.
19. 
Cooney
CM,
Cooney
DS,
Bello
RJ,
Bojovic
B,
Redett
RJ,
Lifchez
SD.
Comprehensive observations of resident evolution: a novel method for assessing procedure-based residency training
.
Plast Reconstr Surg
.
2016
;
137
(
2
):
673
-
678
.
20. 
Cullen
MJ,
Zhang
C,
Sackett
PR,
Thakker
K,
Young
JQ.
Can a situational judgment test identify trainees at risk of professionalism issues? A multi-institutional, prospective cohort study
.
Acad Med
.
2022
;
97
(
10
):
1494
-
1503
.
21. 
Rebel
A,
DiLorenzo
A,
Nguyen
D,
et al.
Should objective structured clinical examinations assist the clinical competency committee in assigning anesthesiology milestones competency?
Anesth Analg
.
2019
;
129
(
1
):
226
-
234
.
22. 
Bryant
BH.
Feasibility of an entrustable professional activity for pathology resident frozen section training
.
Acad Pathol
.
2021
;
8
:
23742895211041757
.
23. 
Rassbach
CE,
Blankenburg
R.
A novel pediatric residency coaching program: outcomes after one year
.
Acad Med
.
2018
;
93
(
3
):
430
-
434
.
24. 
Yen
SPF,
Lewis
PJ.
Implementation of a departmental resident web-based database
.
Acad Radiol
.
2022
;
29
(
1
):
144
-
149
.
25. 
Stahl
CC,
Collins
E,
Jung
SA,
et al.
Implementation of entrustable professional activities into a general surgery residency
.
J Surg Educ
.
2020
;
77
(
4
):
739
-
748
.
26. 
Friedman
KA,
Raimo
J,
Spielmann
K,
Chaudhry
S.
Resident dashboards: helping your clinical competency committee visualize trainees’ key performance indicators
.
Med Educ Online
.
2016
;
21
:
29838
.
27. 
Almufarrej
F,
O’Brien
M,
Shahait
A,
Nava
G.
Feasibility of smartphone application in plastic surgery operative assessments
.
Plast Reconstr Surg Glob Open
.
2022
;
10
(
2
):
e4085
.
28. 
Dougherty
D,
Weiler
EJ,
Sbraccia
JC,
Crowe
CB,
Lindsey
HJ.
A novel dashboard for improving the efficiency of a general surgery residency clinical competency committee meeting
.
Am J Surg
.
2022
;
224
(
1 Pt B
):
363
-
365
.
29. 
Johna
S,
Woodward
B.
Navigating the next accreditation system: a dashboard for the milestones
.
Perm J
.
2015
;
19
(
4
):
61
-
63
.
30. 
Regan
L,
Cope
L,
Omron
R,
Bright
L,
Bayram
JD.
Do end-of-rotation and end-of-shift assessments inform clinical competency committees’ (CCC) decisions?
West J Emerg Med
.
2018
;
19
(
1
):
121
-
127
.
31. 
Bedy
SC,
Goddard
KB,
Stilley
JAW,
Sampson
CS.
Use of emergency department pharmacists in emergency medicine resident milestone assessment
.
West J Emerg Med
.
2019
;
20
(
2
):
357
-
362
.
32. 
Pack
R,
Lingard
L,
Watling
C,
Cristancho
S.
Beyond summative decision making: illuminating the broader roles of competence committees
.
Med Educ
.
2020
;
54
(
6
):
517
-
527
.
33. 
Nabors
C,
Forman
L,
Peterson
SJ,
et al.
Milestones: a rapid assessment method for the clinical competency committee
.
Arch Med Sci
.
2017
;
13
(
1
):
201
-
209
.
34. 
Warburton
KM,
Goren
E,
Dine
CJ.
Comprehensive assessment of struggling learners referred to a graduate medical education remediation program
.
J Grad Med Educ
.
2017
;
9
(
6
):
763
-
767
.
35. 
Blau
JA,
Atia
AN,
Powers
DB.
Clinical competency committees in plastic surgery residency
.
Plast Reconstr Surg Glob Open
.
2021
;
9
(
9
):
e3833
.
36. 
Lyle
B,
Borgert
AJ,
Kallies
KJ,
Jarman
BT.
Do attending surgeons and residents see eye to eye? An evaluation of the Accreditation Council for Graduate Medical Education milestones in general surgery residency
.
J Surg Educ
.
2016
;
73
(
6
):
e54
-
e58
.
37. 
Watson
RS,
Borgert
AJ,
O Heron
CT,
et al.
A multicenter prospective comparison of the Accreditation Council for Graduate Medical Education milestones: clinical competency committee vs. resident self-assessment
.
J Surg Educ
.
2017
;
74
(
6
):
e8
-
e14
.
38. 
Tichter
AM,
Mulcare
MR,
Carter
WA.
Interrater agreement of emergency medicine milestone levels: resident self-evaluation vs clinical competency committee consensus
.
Am J Emerg Med
.
2016
;
34
(
8
):
1677
-
1679
.
39. 
Ross
FJ,
Metro
DG,
Beaman
ST,
et al.
A first look at the Accreditation Council for Graduate Medical Education anesthesiology milestones: implementation of self-evaluation in a large residency program
.
J Clin Anesth
.
2016
;
32
:
17
-
24
.
40. 
Srikumaran
D,
Tian
J,
Ramulu
P,
et al.
Ability of ophthalmology residents to self-assess their performance through established milestones
.
J Surg Educ
.
2019
;
76
(
4
):
1076
-
1087
.
41. 
Meier
AH,
Gruessner
A,
Cooney
RN.
Using the ACGME milestones for resident self-evaluation and faculty engagement
.
J Surg Educ
.
2016
;
73
(
6
):
e150
-
e157
.
42. 
Chow
I,
Nguyen
VT,
Losee
JE,
et al.
Milestones in plastic surgery: attending assessment versus resident assessment
.
Plast Reconstr Surg
.
2019
;
143
(
2
):
e425
-
e432
.
43. 
Kwasny
L,
Shebrain
S,
Munene
G,
Sawyer
R.
Is there a gender bias in milestones evaluations in general surgery residency training?
Am J Surg
.
2021
;
221
(
3
):
505
-
508
.
44. 
Santen
SA,
Yamazaki
K,
Holmboe
ES,
Yarris
LM,
Hamstra
SJ.
Comparison of male and female resident milestone assessments during emergency medicine residency training: a national study
.
Acad Med
.
2020
;
95
(
2
):
263
-
268
.
45. 
Zuckerbraun
NS,
Levasseur
K,
Kou
M,
et al.
Gender differences among milestone assessments in a national sample of pediatric emergency medicine fellowship programs
.
AEM Educ Train
.
2020
;
5
(
3
):
e10543
.
46. 
Hauer
KE,
Jurich
D,
Vandergrift
J,
et al.
Gender differences in milestone ratings and medical knowledge examination scores among internal medicine residents
.
Acad Med
.
2021
;
96
(
6
):
876
-
884
.
47. 
Athy
S,
Talmon
G,
Samson
K,
Martin
K,
Nelson
K.
Faculty versus resident self-assessment using pathology milestones: how aligned are we?
Acad Pathol
.
2021
;
8
:
23742895211060526
.
48. 
Turner
JA,
Fitzsimons
MG,
Pardo
MC
, et al.
Effect of performance deficiencies on graduation and board certification rates: a 10-yr multicenter study of anesthesiology residents
.
Anesthesiology
.
2016
;
125
(
1
):
221
-
229
.
49. 
Kimbrough
MK,
Thrush
CR,
Barrett
E,
Bentley
FR,
Sexton
KW.
Are surgical milestone assessments predictive of in-training examination scores?
J Surg Educ
.
2018
;
75
(
1
):
29
-
32
.
50. 
Frey-Vogel
AS,
Scott-Vernaglia
SE,
Carter
LP,
Huang
GC.
Simulation for milestone assessment: use of a longitudinal curriculum for pediatric residents
.
Simul Healthc
.
2016
;
11
(
4
):
286
-
292
.
51. 
Schauer
DP,
Kinnear
B,
Kelleher
M,
Sall
D,
Schumacher
DJ,
Warm
EJ.
Developing the expected entrustment score: accounting for variation in resident assessment
.
J Gen Intern Med
.
2022
;
37
(
14
):
3670
-
3675
.
52. 
Roshan
A,
Wagner
N,
Acai
A,
et al.
Comparing the quality of narrative comments by rotation setting
.
J Surg Educ
.
2021
;
78
(
6
):
2070
-
2077
.
53. 
Donato
AA,
Alweis
R,
Wenderoth
S.
Design of a clinical competency committee to maximize formative feedback
.
J Community Hosp Intern Med Perspect
.
2016
;
6
(
6
):
33533
.
54. 
Abbott
KL,
George
BC,
Sandhu
G,
et al.
Natural language processing to estimate clinical competency committee ratings
.
J Surg Educ
.
2021
;
78
(
6
):
2046
-
2051
.
55. 
Rozenshtein
A,
Mullins
ME,
Deitte
LA,
et al.
“What program directors think” II: results of the 2013 and 2014 annual surveys of the Association of Program Directors in
Radiology. Acad Radiol
.
2015
;
22
(
6
):
787
-
793
.
56. 
Lloyd
RB,
Park
YS,
Tekian
A,
Marvin
R.
Understanding assessment systems for clinical competency committee decisions: evidence from a multisite study of psychiatry residency training programs
.
Acad Psychiatry
.
2020
;
44
(
6
):
734
-
740
.
57. 
Sebesta
EM,
Cooper
KL,
Badalato
GM.
Program director perceptions of usefulness of the Accreditation Council for Graduate Medical Education milestones system for urology resident evaluation
.
Urology
.
2019
;
124
:
28
-
32
.
58. 
Smit
MP,
de Hoog
M,
Brackel
HJL,
ten Cate
O,
Gemke
RJBJ.
A national process to enhance the validity of entrustment decisions for Dutch pediatric residents
.
J Grad Med Educ
.
2019
;
11
(
suppl 4
):
158
-
164
.
59. 
Duitsman
ME,
Fluit
CRMG,
van Alfen-van der Velden
JAEM,
et al.
Design and evaluation of a clinical competency committee
.
Perspect Med Educ
.
2019
;
8
(
1
):
1
-
8
.
60. 
Ekpenyong
A,
Becker
K.
What resources do clinical competency committees (CCCs) require to do their work? A pilot study comparing CCCs across specialties
.
Med Teach
.
2021
;
43
(
1
):
86
-
92
.
61. 
Ekpenyong
A,
Baker
E,
Harris
I,
et al.
How do clinical competency committees use different sources of data to assess residents’ performance on the internal medicine milestones? A mixed methods pilot study
.
Med Teach
.
2017
;
39
(
10
):
1074
-
1083
.
62. 
Shumway
NM,
Dacus
JJ,
Lathrop
KI,
Hernandez
EP,
Miller
M,
Karnad
AB.
Use of milestones and development of entrustable professional activities in 2 hematology/oncology training programs
.
J Grad Med Educ
.
2015
;
7
(
1
):
101
-
104
.
63. 
Chen
F,
Arora
H,
Martinelli
SM.
Use of key performance indicators to improve milestone assessment in semi-annual clinical competency committee meetings
.
J Educ Perioper Med
.
2017
;
19
(
4
):
e611
.
64. 
Van Osdol
AD,
O’Heron
CT,
Jarman
BT.
“Same as it ever was”—only different: a successful ACGME innovative project to equip PGY I residents to take at-home call
.
J Surg Educ
.
2014
;
71
(
6
):
e104
-
e110
.
65. 
Kearney
AM,
Rokni
AM,
Gosain
AK.
The Accreditation Council for Graduate Medical Education milestones in integrated plastic surgery programs: how competency-based assessment has been implemented
.
Plast Reconstr Surg
.
2022
;
149
(
4
):
1001
-
1007
.
66. 
Schumacher
DJ,
Martini
A,
Sobolewski
B,
et al.
Use of resident-sensitive quality measure data in entrustment decision making: a qualitative study of clinical competency committee members at one pediatric residency
.
Acad Med
.
2020
;
95
(
11
):
1726
-
1735
.
67. 
Goyal
N,
Folt
J,
Jaskulka
B,
et al.
Assessment methods and resource requirements for milestone reporting by an emergency medicine clinical competency committee
.
Med Educ Online
.
2018
;
23
(
1
):
1538925
.
68. 
Kou
M,
Baghdassarian
A,
Khanna
K,
et al.
Guiding fellows to independent practice: current trends in pediatric emergency medicine fellow supervision
.
Pediatr Emerg Care
.
2022
;
38
(
10
):
517
-
520
.
69. 
Schumacher
DJ,
Poynter
S,
Burman
N,
et al.
Justifications for discrepancies between competency committee and program director recommended resident supervisory roles
.
Acad Pediatr
.
2019
;
19
(
5
):
561
-
565
.
70. 
Schumacher
DJ,
Martini
A,
Bartlett
KW,
et al.
Key factors in clinical competency committee members’ decisions regarding residents’ readiness to serve as supervisors: a national study
.
Acad Med
.
2019
;
94
(
2
):
251
-
258
.
71. 
Schumacher
DJ,
Schwartz
A,
Zenel
JA
, et al.
Narrative performance level assignments at initial entrustment and graduation: integrating EPAs and milestones to improve learner assessment
.
Acad Med
.
2020
;
95
(
11
):
1736
-
1744
.
72. 
Stehman
CR,
Hochman
S,
Fernández-Frackelton
M,
et al.
Professionalism milestones assessments used by emergency medicine residency programs: a cross-sectional survey
.
West J Emerg Med
.
2019
;
21
(
1
):
152
-
159
.
73. 
Schumacher
DJ,
Martini
A,
Kinnear
B,
et al.
Facilitators and inhibitors to assessing entrustable professional activities in pediatric residency
.
Acad Pediatr
.
2021
;
21
(
4
):
735
-
741
.
74. 
Yaghmour
NA,
Poulin
LJ,
Bernabeo
EC,
et al.
Stages of milestones implementation: a template analysis of 16 programs across 4 specialties [published correction appears in J Grad Med Educ. 2022;14(6):732. doi:10.4300/JGME-D-22-00841.1]
.
J Grad Med Educ
.
2021
;
13
(
suppl 2
):
14
-
44
.
75. 
Balach
T,
Conforti
LN,
Sangha
S,
Edgar
L,
Ames
E.
Qualitative study of orthopedic surgery milestones 1.0: burdens and benefits
.
J Surg Educ
.
2022
;
79
(
5
):
1259
-
1269
.
76. 
Nabors
C,
Peterson
SJ,
Forman
L,
et al.
Operationalizing the internal medicine milestones—an early status report
.
J Grad Med Educ
.
2013
;
5
(
1
):
130
-
137
.
77. 
Maranich
AM,
Hemmer
PA,
Uijtdehaage
S,
Battista
A.
ACGME milestones in the real world: a qualitative study exploring response process evidence
.
J Grad Med Educ
.
2022
;
14
(
2
):
201
-
209
.
78. 
Hanson
MN,
Pryor
AD,
Jeyarajah
DR,
et al.
Implementation of entrustable professional activities into fellowship council accredited programs: a pilot project
.
Surg Endosc
.
2023
;
37
(
4
):
3191
-
3200
.
79. 
Lewis
KO,
Hathaway
SB,
Bratcher
D,
Blowey
D,
Knapp
JF.
Current milestones assessment practices, needs, and challenges of program directors: a collective case study in a pediatric hospital setting
.
Cureus
.
2021
;
13
(
4
):
e14585
.
80. 
Yamazaki
K,
Holmboe
ES,
Hamstra
SJ.
An empirical investigation into milestones factor structure using national data derived from clinical competency committees
.
Acad Med
.
2022
;
97
(
4
):
569
-
576
.
81. 
Schumacher
DJ,
Michelson
C,
Poynter
S,
et al.
Thresholds and interpretations: how clinical competency committees identify pediatric residents with performance concerns
.
Med Teach
.
2018
;
40
(
1
):
70
-
79
.
82. 
Conforti
LN,
Yaghmour
NA,
Hamstra
SJ,
et al.
The effect and use of milestones in the assessment of neurological surgery residents and residency programs
.
J Surg Educ
.
2018
;
75
(
1
):
147
-
155
.
83. 
Ketteler
ER,
Auyang
ED,
Beard
KE,
et al.
Competency champions in the clinical competency committee: a successful strategy to implement milestone evaluations and competency coaching
.
J Surg Educ
.
2014
;
71
(
1
):
36
-
38
.
84. 
Klutts
JS,
Guerin
LA,
Bruch
LA,
et al.
Pathology milestones: assessing clinical competency by committee
.
Acad Pathol
.
2015
;
2
(
4
):
2374289515614003
.
85. 
Bienstock
JL,
Shivraj
P,
Yamazaki
K,
et al.
Correlations between Accreditation Council for Graduate Medical Education obstetrics and gynecology milestones and American Board of Obstetrics and Gynecology qualifying examination scores: an initial validity study
.
Am J Obstet Gynecol
.
2021
;
224
(
3
):
308.e1
-
308.e25
.
86. 
Mink
R,
Herman
BE,
Carraccio
C,
et al.
Agreement of program directors with clinical competency committees for fellow entrustment
.
J Med Educ Curric Dev
.
2020
;
7
:
2382120520936613
.
87. 
Mink
RB,
Schwartz
A,
Herman
BE,
et al.
Validity of level of supervision scales for assessing pediatric fellows on the common pediatric subspecialty entrustable professional activities
.
Acad Med
.
2018
;
93
(
2
):
283
-
291
.
88. 
Mikhaeil-Demo
Y,
Holmboe
E,
Gerard
EE,
et al.
Simulation-based assessments and graduating neurology residents’ milestones: status epilepticus milestones
.
J Grad Med Educ
.
2021
;
13
(
2
):
223
-
230
.
89. 
Van Melle
E,
Hall
A,
Schumacher
D,
et al.
Capturing outcomes of competency-based medical education: the call and the challenge
.
Med Teach
.
2021
;
43
(
7
):
794
-
800
.

The online supplementary data contains a visual abstract.

Funding: The authors report no external funding source for this study.

Conflict of interest: The authors declare they have no competing interests.

Disclaimer: Andem Ekpenyong, MD, MHPE, works part-time for the Accreditation Council for Graduate Medical Education (ACGME), and at the time of writing, Eric S. Holmboe, MD, worked full-time for the ACGME.

Supplementary data