I am not a robot, and I can prove it…just by clicking on all the picture tiles with bicycles or buses. These CAPTCHA tests, or Completely Automated Public Turing tests to tell Computers and Humans Apart, protect websites from bots and spam. Will residency program directors soon need their own form of CAPTCHA to protect the integrity of the residency application process from artificial intelligence (AI)? The 2023-2024 residency application season is the first in which generative AI has been a popularly recognized and easily accessed tool. While its use has advantages, it also threatens the sincerity and reliability of important data points in application evaluation.
In November 2022, the San Francisco-based tech company OpenAI released ChatGPT—a free online chat bot capable of writing prose that is nearly indistinguishable from human-generated text. It is now one of several such services available online. Generative Pre-trained Transformer (GPT) describes large language model artificial neural networks that are “trained” on large data sets and can perform natural language processing tasks such as generating new text and analyzing text for specified content.1 After the March 2023 update, OpenAI announced on its website:
“We’ve trained a model called ChatGPT which interacts in a conversational way. The dialogue format makes it possible for ChatGPT to answer follow-up questions, admit its mistakes, challenge incorrect premises, and reject inappropriate requests.”2
The use of AI was quickly adopted by university students. An informal survey of nearly 5000 Stanford students found that 17% used ChatGPT in fall 2022 for assignments and examinations (the term in which ChatGPT became publicly available).3 Over 25% felt that AI use, even if answers are unedited, was not a violation of honor code.3 While there are no data on AI use by health professional students, one must assume parallels with the general university population.
Because AI use is widespread at universities, many applicants will consider using it when writing a personal statement (PS). The PS has long been the applicant’s opportunity to narrate their journey through medical school and to their specialty of choice, and to highlight their unique qualities and motivations.4 Using AI to assist in writing them can improve efficiency, help with brainstorming, and improve communication for people for whom English is a second language.5 It also improves equity in access to application assistance services, which can cost in the thousands of dollars.6
On the other hand, multiple online forums advise against using AI for the PS because it seems “inauthentic and unoriginal—and therefore not as good as” an applicant’s own work.7 An AI-written PS lacks the narrative voice, context, and specific details that make it personal.8 Another notable drawback of AI-generated work is its propensity for lengthy and potentially excessive text, a factor that poses a barrier to reader efficiency. No data yet exist on the applicant’s perspective on the use of AI in PS generation.
Through letters of recommendation (LORs), faculty advocate for students and highlight their longitudinal relationships.9 Similarly to the PS, AI can assist in generating an LOR, which can improve writing efficiency and reduce the use of biased language,10 with which LORs have long been fraught.11 There are no studies of how frequently AI is used to generate a professional school LOR. An article in The Atlantic, “The End of Recommendation Letters: Professors, Like Their Students, Use ChatGPT to Get Out of Doing Their Assignments,” reveals one popular sentiment against the use of AI in letter writing.12 Because AI-generated letters usually lack the personal connection, emotional intelligence, nuances, and subjective judgement of a human-written letter,13 heavy editing is advisable if AI is used for the first draft. If AI is used in text generation, authorship is not attributed to the AI; at least this is the stance for scientific journals.14 Ultimately, the signer of an LOR takes responsibility for its content, which has always been the case.
Program directors may also look for assistance from AI. Application inflation necessitates improved efficiency in application review.15,16 AI could streamline application screening by looking for specific combinations of content. AI could mitigate human bias, for example, by identifying biased language in an LOR,17 but could also propagate it. AI must be “trained” on the attractive features of an application, and if bias exists in the training data, that bias can be codified. When Amazon piloted the use of AI in application screening, the training data set included a majority of male employees, which unintentionally led to weighting “non-maleness” as a deficit in the application. Amazon discontinued use of this faulty paradigm once the bias was discovered.18
While AI offers advantages and threats to the application process, identifying its use is challenging. The majority of program directors reviewing the PS found AI-generated ones acceptable and did not suspect that they were not human-written.6 Faculty with an MD or PhD could also not consistently determine which medical research abstracts were written by humans vs AI.19 Software for AI detection is publicly available and often free, but would require additional time in the already compressed application review period, as each piece of prose would need to be uploaded or pasted into the detection site. Additionally, AI-detecting software may be inaccurate, particularly when evaluating text from a writer for whom English is a second language.20 The Mozilla Foundation, whose mission is to maintain open and publicly accessible internet services, warns: “Detector tools will always be imperfect, which makes them nearly useless…One cannot accuse a student of using AI…based on the output of a detector tool that you know has a 10% chance of giving a false positive.”21 At least for now, it seems impossible to know with certainty if the content of prose is the product of the author and contains accurate information. However, this has always been the case. There have always been threats to authenticity and concern over the ethics of using such potentially flawed datapoints in admissions.22 AI only exacerbates them.
Currently the only guideline from the Association of American Medical Colleges on the use of AI in residency applications states: “[the PS] must be your own work and not the work of another author or the product of artificial intelligence.”23 Applicants submitting applications to medical school through American Medical College Application Service are required to attest that “all written passages…are my own and have not been written, in part or in whole, by another author and are not the product of artificial intelligence.”24 However, such an attestation is not required for residency application. These policies do not reflect the complexities of AI use. For example, is it allowable to brainstorm using AI and then edit the text heavily? Can AI be used to edit one’s own work to improve grammar and readability?
Until there are more nuanced and generally agreed-upon guidelines at a national level on the use of AI in application preparation, graduate medical education (GME) programs should consider posting explicit policies on their websites regarding the use of AI in preparing application materials.25 AI could even aid in the generation of such a policy (Box 1). However, author-generated guidelines are clearer and more concise, which highlights the difference in word efficiency between AI and human text (Box 2). Any limitations should be stated with a purpose behind them, although such policies are presently unenforceable. There have always been concerns over the use of the PS and the LOR in applicant evaluation; AI only exacerbates these and may ultimately lead to a decline in the importance of these data points in the application package. The national GME community needs to revisit the methods of holistic application review in light of the challenges posed by AI.
Respect Ethical Considerations: Ensure that the use of AI aligns with ethical guidelines, including privacy, consent, and fairness. Avoid using AI in a way that may infringe upon others’ rights or values.
Clearly Disclose AI Use: If you choose to incorporate AI tools or algorithms into your application materials, clearly indicate which aspects were generated or reviewed using AI. Transparency is essential to maintain trust and credibility.
AI as a Supplement, Not a Replacement: Use AI as a supplement to your application, not as a substitute for your own work and effort. Your application should primarily reflect your skills, experiences, and personal input.
Maintain Authenticity: Your application materials should genuinely represent your qualifications, personality, and experiences. Avoid using AI to artificially enhance your qualifications or present yourself in a misleading way.
Provide Context: Explain why you chose to use AI in your application and how it adds value to your materials. Clarify how the AI tools contributed to your presentation and understanding.
Beware of Plagiarism: Ensure that any content generated by AI tools, such as personal statements, essays, or letters of recommendation, is unique and not plagiarized. Plagiarism is a serious offense in academic and professional settings.
Check for Errors: AI-generated content may have errors or inaccuracies. Always review and proofread AI-generated materials to correct any mistakes and ensure the highest quality.
Be Prepared to Discuss AI Use: If invited for an interview or during the selection process, be ready to discuss your AI use in detail. Explain how it complements your qualifications and demonstrates your technological proficiency.
Use AI Appropriately: Only use AI where it genuinely adds value to your application. Avoid using AI for tasks that are traditionally expected to be completed by the applicant, such as personal statements or letters of intent.
Respect Program Guidelines: Always follow the specific guidelines provided by the residency program. If the program explicitly prohibits the use of AI in application materials, respect and adhere to these rules.
These guidelines aim to strike a balance between encouraging applicants to utilize technology responsibly and ensuring that their applications reflect their genuine qualifications and abilities. The ultimate goal is to maintain fairness and integrity in the application process while allowing applicants to harness the benefits of AI when it’s appropriate and ethical.
Generated in response to the instruction (generated from ChatGPT on 10/25/23): Can you produce a set of guidelines for residency programs to provide to applicants to discourage them from using AI as part of their application?
Follow and Respect Program Specific Guidelines: Prior to submission of application materials, ensure you are aware of each program’s guideline surrounding AI usage.
Be Aware of Plagiarism and Maintain Authenticity: If utilizing AI to create application materials, be aware that created elements may qualify as plagiarism. A strong application will maintain the applicant’s authentic experiences and perspective.
Disclose AI Use Clearly: For all aspects of the application where AI is used, it should be clearly stated and disclosed to application reviewers.
Proofread and Ensure Errors Are Corrected: Careful review of all AI-generated materials is essential prior to submission for edits and correction of errors.
Selection of AI Usage: The applicant must carefully consider what aspects of the application could be amenable for AI usage. Avoid using AI for tasks that should reflect the individual applicant’s original insight and work (personal statement, letter of intent) in order to maximize the authenticity and impact of these components of the application.