<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Dermatol</journal-id><journal-id journal-id-type="publisher-id">derma</journal-id><journal-id journal-id-type="index">29</journal-id><journal-title>JMIR Dermatology</journal-title><abbrev-journal-title>JMIR Dermatol</abbrev-journal-title><issn pub-type="epub">2562-0959</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v8i1e72371</article-id><article-id pub-id-type="doi">10.2196/72371</article-id><article-categories><subj-group subj-group-type="heading"><subject>Short Paper</subject></subj-group></article-categories><title-group><article-title>Image Generation of Common Dermatological Diagnoses by Artificial Intelligence: Evaluation Study of the Potential for Education and Training Purposes</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Kooper-Johnson</surname><given-names>Sarah</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lim</surname><given-names>Subin</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Aldhalaan</surname><given-names>Jumana</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Cobos</surname><given-names>Gabriela</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Kodumudi</surname><given-names>Vijay</given-names></name><degrees>MBA, MPH, MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Loo</surname><given-names>Daniel S</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Mervis</surname><given-names>Joshua</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Petrillo</surname><given-names>Madalyn</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Robinson</surname><given-names>Sarah N</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Nguyen</surname><given-names>Bichchau Michelle</given-names></name><degrees>MBA, MPH, MD</degrees><xref ref-type="aff" rid="aff1"/></contrib></contrib-group><aff id="aff1"><institution>Tufts Medical Center</institution><addr-line>260 Tremont Street</addr-line><addr-line>Boston</addr-line><addr-line>MA</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Parker</surname><given-names>Eva</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Lian</surname><given-names>Cheryl</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Tiryaki</surname><given-names>Volkan Mujdat</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Sarah Kooper-Johnson, MD, Tufts Medical Center, 260 Tremont Street, Boston, MA, 02116, United States, 1 (617) 636-0156; <email>sarahkj111@gmail.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>16</day><month>12</month><year>2025</year></pub-date><volume>8</volume><elocation-id>e72371</elocation-id><history><date date-type="received"><day>08</day><month>02</month><year>2025</year></date><date date-type="rev-recd"><day>12</day><month>08</month><year>2025</year></date><date date-type="accepted"><day>26</day><month>08</month><year>2025</year></date></history><copyright-statement>&#x00A9; Sarah Kooper-Johnson, Subin Lim, Jumana Aldhalaan, Gabriela Cobos, Vijay Kodumudi, Daniel S Loo, Joshua Mervis, Madalyn Petrillo, Sarah N Robinson, Bichchau Michelle Nguyen. Originally published in JMIR Dermatology (<ext-link ext-link-type="uri" xlink:href="http://derma.jmir.org">http://derma.jmir.org</ext-link>), 16.12.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Dermatology, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="http://derma.jmir.org">http://derma.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://derma.jmir.org/2025/1/e72371"/><abstract><sec><title>Background</title><p>The integration of artificial intelligence (AI) into dermatology holds promise for education and diagnostic purposes, particularly through image generation, which has not been well studied.</p></sec><sec><title>Objective</title><p>This study aimed to assess whether AI image generation software can generate accurate images of classic dermatological conditions and whether they are recognizable as computer-generated.</p></sec><sec sec-type="methods"><title>Methods</title><p>Images of 10 dermatological conditions were generated using DALLE-2 and DALLE-3 programs. These images were randomized among clinical photographs and distributed to dermatology residents and attending physicians. Participants were instructed to (1) identify AI-generated images and (2) provide their diagnosis.</p></sec><sec sec-type="results"><title>Results</title><p>AI-generated images were detected as computer-generated in 70.8% (85/120) of cases. Correct diagnoses were made based on all AI images 40.83% (49/120) of the time. This was significantly lower than the 72.0% (46/60) recognition rate for clinical photographs (<italic>P</italic>&#x003C;.001). DALLE-2 images were diagnosed correctly less frequently (25.0%, 15/60) than DALLE-3 images (56.6%, 34/60; <italic>P</italic>&#x003C;.001).</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>AI-generated images of common dermatological conditions are becoming more accurate. This holds great implications for education but should be used with caution as further research is needed with more advanced, specific, and inclusive training data. Limitations include the use of AI image generators created by a single parent company as well as the use of a limited set of diagnoses.</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>technology</kwd><kwd>morphology</kwd><kwd>education</kwd><kwd>diagnostic</kwd><kwd>image generation</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>The study and practice of dermatology is rooted in visual inspection and pattern recognition of classic disease morphology. Patient images in textbooks and archives aid in learning these patterns and developing differential diagnoses along with morphological descriptions. However, these images are not always immediately available and may have implications regarding patient privacy. Furthermore, historically, there has been significant limitations of availability of images depicting dermatological conditions in skin of color [<xref ref-type="bibr" rid="ref1">1</xref>].</p><p>The use of artificial intelligence (AI) applications in dermatology has been described in many articles in which algorithms have neared or surpassed the diagnostic accuracy of clinicians when it comes to analyzing images [<xref ref-type="bibr" rid="ref2">2</xref>]. New developments in AI have also reached mainstream headlines, as applications such as ChatGPT are able to perform well on dermatology boards style questions [<xref ref-type="bibr" rid="ref3">3</xref>] and provide appropriate answers to patient questions [<xref ref-type="bibr" rid="ref4">4</xref>] using complex neural networks. Their parent-company, OpenAI, also released a text-to-image generator in 2021 called DALL-E trained on text-images pairs sourced from across the internet. The autoregressive model produces images from scratch based on the text that the user inputs [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>] .</p><p>This could have implications for the image-based field of dermatology, particularly in education. If AI can generate images of various conditions, these could be used alongside patient images as a quick reference tool and aid in the development of differential diagnoses. The aim of this study is to assess whether successive iterations of DALL-E can produce accurate images of classic dermatological conditions and whether these images can convincingly resemble clinical images or if they will be recognized as AI-generated.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Ethical Considerations</title><p>Given that this study did not include patient intervention or patient data, patient consent, de-identification, and privacy protections were not indicated. Additionally, this work was exempt from approval by the institutional review board..</p></sec><sec id="s2-2"><title>Study Overview</title><p>A list of 10 common dermatologic conditions (vitiligo, urticaria, rosacea, acne, milia, cellulitis, molluscum contagiosum, psoriasis, seborrheic dermatitis, and melasma) was generated by OpenAI&#x2019;s DALLE-2 and DALLE-3. Clinical photographs, previously used as teaching Kodachromes, and deemed diagnostic of these 10 conditions by an attending physicians with &#x003E;5 years of dermatology experience (BN) were also collected from an academic dermatologist with permission from patients for their use. The phrase &#x201C;Please generate a photorealistic image of X on the Y&#x201D; was used with each query where X was the dermatologic condition and Y was the location of the lesion or condition. For example, for acne, the phrase &#x201C;please generate a photorealistic image of acne on the face&#x201D; was inputted. During beta testing, it was found that using the diagnosis name as opposed to a morphological description yielded more accurate results.</p><p>Ten images from each generator were compiled and randomized among the 10 clinical photographs, a 50% ratio of images. A quiz containing all 30 of these images was administered to 2 second-year and 1 third-year dermatology resident, along with 3 attending physicians with &#x003E;5 years of dermatology experience. Participants were instructed to (1) identify AI-generated images and (2) provide their diagnosis.</p></sec><sec id="s2-3"><title>Statistical Analysis</title><p>Two-sample z tests were used to determine if one image generator produced images that were more likely to be recognized as AI and to assess if there was a significant difference in the capability of these algorithms to produce recognizable and diagnosable images. These tests were also performed to see if residents or attending physicians were able to recognize AI images or come up with correct diagnoses at different rates. The results were also broken down by diagnosis category, for example inflammatory disorders and pigmentary disorder, and two-sample z tests were conducted to determine if the image generators were more recognizable as AI or more accurate with any particular category of diagnosis. Statistical significance level was set to &#x03B1;=.05. All statistical analyses were performed in Microsoft Excel.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><p>Upon subjective evaluation of AI-generated images, those generated by DALLE-2 exhibited a higher degree of photorealism compared to DALLE-3. However, these images tended to be non-specific and generally failed to accurately represent specific diagnoses. DALLE-3 images were more cartoonish in nature but showcased specific dermatological findings such as scale indicative of psoriasis. Of the 20 AI-generated images, none depicted individuals of older age, and only 2 images appeared to represent skin of color. These images can be found in <xref ref-type="fig" rid="figure1">Figure 1</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Conditions studied and images generated from DALLE-2 and DALLE-3 Software.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="derma_v8i1e72371_fig01.png"/></fig><p>In the survey of dermatology residents and attending physicians, AI-generated images were detected as computer-generated in 70.8% (85/120) of cases. DALLE-2 images were recognized as AI-generated less often (41.66%, 25/60) than DALLE-3 images (100%, 60/60; <italic>P</italic>&#x003C;.001). Clinical photographs were mistakenly perceived as AI-generated in only 1.66% (1/60) of instances. Residents recognized all AI images more often than attending physicians (75% [45/60] vs 66.67% [40/60]; <italic>P</italic>=.0224) and recognized DALLE-2 images as AI generated more often than attending physicians (50% [15/30] vs 33.33% [10/30]; <italic>P</italic>=.001). Both groups recognized DALLE-3 images as AI generated 100% of the time.</p><p>In terms of producing recognizable images, correct diagnoses were made based on all AI images 40.83% (49/120) of the time. This was significantly lower than the 71.67% (43/60) recognition rate for clinical photographs (<italic>P</italic>&#x003C;.001). When comparing image generators, DALLE-2 images were diagnosed correctly less frequently, at 25.0% (15/60), compared with DALLE-3 images, which were diagnosed correctly 56.67% (34/60) of the time (<italic>P</italic>&#x003C;.001). Attending physicians were able to correctly identify DALLE-2 images more often than residents (36.67% [11/30] vs 20% [6/30]; <italic>P</italic>=.001). The rate of correct diagnosis was the same for residents and attending physicians for DALLE-3 images (56.67%, 34/60). DALLE-2 images representing urticaria, psoriasis, and seborrheic dermatitis were misclassified the most often in this set. The DALLE-3 image representing cellulitis was misclassified the most often in this set. The full analysis results are shown in <xref ref-type="table" rid="table1">Table 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Quiz results broken down by respondent type describing how often respondents recognized images as artificial intelligence (AI) generated and how often the correct diagnosis was given.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Respondents</td><td align="left" valign="bottom">Clinical images</td><td align="left" valign="bottom">All AI images</td><td align="left" valign="bottom">DALLE-2 images</td><td align="left" valign="bottom">DALLE-3 images</td><td align="left" valign="bottom"><italic>P</italic> value between DALLE-2 and DALLE-3</td></tr></thead><tbody><tr><td align="left" valign="top">Percentage recognized as AI</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>All respondents</td><td align="left" valign="top">1.66 (1/60)</td><td align="left" valign="top">70.83 (85/120)</td><td align="left" valign="top">41.66 (25/60)</td><td align="left" valign="top">100 (60/60)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Resident respondents</td><td align="left" valign="top">0 (0/30)</td><td align="left" valign="top">75 (45/60)</td><td align="left" valign="top">50 (15/30)</td><td align="left" valign="top">100 (30/30)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Attending physician respondents</td><td align="left" valign="top">3.33 (1/30)</td><td align="left" valign="top">66.67 (40/60)</td><td align="left" valign="top">33.33 (10/30)</td><td align="left" valign="top">100 (30/30)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top"><italic>P</italic> value between resident and attending physician responses</td><td align="left" valign="top">.747</td><td align="left" valign="top">.022</td><td align="left" valign="top">.001</td><td align="left" valign="top">1</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Percentage of given correct diagnosis</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>All respondents</td><td align="left" valign="top">71.67 (43/60)</td><td align="left" valign="top">40.80 (49/120)</td><td align="left" valign="top">25 (15/60)</td><td align="left" valign="top">56.67 (34/60)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Resident respondents</td><td align="left" valign="top">80 (24/30)</td><td align="left" valign="top">38.33 (23/60)</td><td align="left" valign="top">20 (6/30)</td><td align="left" valign="top">56.67 (34/60)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Attending physician respondents</td><td align="left" valign="top">63.33 (19/30)</td><td align="left" valign="top">46.67 (28/60)</td><td align="left" valign="top">36.67 (11/30)</td><td align="left" valign="top">56.67 (34/60)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top"><italic>P</italic> value between resident and attending physician responses</td><td align="left" valign="top">.001</td><td align="left" valign="top">.002</td><td align="left" valign="top">.001</td><td align="left" valign="top">1</td><td align="left" valign="top"/></tr></tbody></table></table-wrap><p>When broken down by disease category, AI-generated images of infectious diseases from all programs were recognized as AI the most often at 83.33% (20/24). Across all categories, DALLE-3 images were recognized significantly more often than DALLE-2 images (<italic>P</italic>&#x003C;.001 for each). Infectious etiologies also had low rates of correct diagnosis, being correctly labeled only 16.67% (2/12) of the time for both DALLE-2 and DALLE-3 images. Otherwise, DALLE-3 images were given the correct diagnosis more often than DALLE-2 images across categories, with pigmentary disorders being the most accurate at 91.67% (11/12) recognition. The full analysis results are shown in <xref ref-type="table" rid="table2">Table 2</xref>.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Quiz results broken down by disease category describing often respondents recognized images as artificial intelligence (AI) generated and how often the correct diagnosis was given.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Disease category</td><td align="left" valign="bottom">Inflammatory</td><td align="left" valign="bottom">Infectious</td><td align="left" valign="bottom">Pigmentary</td><td align="left" valign="bottom">Other</td></tr></thead><tbody><tr><td align="left" valign="top">Percentage recognized as AI (All AI programs)</td><td align="left" valign="top">65 (39/60)</td><td align="left" valign="top">83.33 (20/24)</td><td align="left" valign="top">70.83 (17/24)</td><td align="left" valign="top">66.67 (8/12)</td></tr><tr><td align="left" valign="top">Percentage given correct diagnosis (clinical images)</td><td align="left" valign="top">80 (24/30)</td><td align="left" valign="top">66.67 (8/12)</td><td align="left" valign="top">83.33 (10/12)</td><td align="left" valign="top">16.67 (1/6)</td></tr><tr><td align="left" valign="top">Percentage given correct diagnosis (all AI programs)</td><td align="left" valign="top">36.67 (22/60)</td><td align="left" valign="top">16.67 (4/24)</td><td align="left" valign="top">70.83 (17/24)</td><td align="left" valign="top">41.67 (5/12)</td></tr><tr><td align="left" valign="top"><italic>P</italic> value between all AI and clinical images for correct diagnosis</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">.019</td><td align="left" valign="top">.012</td></tr><tr><td align="left" valign="top">Percentage recognized as AI (DALLE-2)</td><td align="left" valign="top">30 (9/30)</td><td align="left" valign="top">66.67 (8/12)</td><td align="left" valign="top">41.67 (5/12)</td><td align="left" valign="top">50 (3/6)</td></tr><tr><td align="left" valign="top">Percentage recognized as AI (DALLE-3)</td><td align="left" valign="top">100 (30/30)</td><td align="left" valign="top">100 (12/12)</td><td align="left" valign="top">100 (12/12)</td><td align="left" valign="top">100 (6/6)</td></tr><tr><td align="left" valign="top"><italic>P</italic> value between DALLE-2 and DALLE-3 for AI recognition</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Percentage given correct diagnosis (DALLE-2)</td><td align="left" valign="top">26.67 (8/30)</td><td align="left" valign="top">16.67 (2/12)</td><td align="left" valign="top">41.67 (5/12)</td><td align="left" valign="top">0 (0/6)</td></tr><tr><td align="left" valign="top">Percentage given correct diagnosis (DALLE-3)</td><td align="left" valign="top">46.67 (14/30)</td><td align="left" valign="top">16.67 (2/12)</td><td align="left" valign="top">91.67 (11/12)</td><td align="left" valign="top">83.33 (5/6)</td></tr><tr><td align="left" valign="top"><italic>P</italic> value between DALLE-2 and DALLE-3 for correct diagnosis</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x003E;.99</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><p>The objective of this work was to determine if the AI image generator DALL-E is able to produce realistic and accurate depictions of common dermatologic conditions. As the field of AI continues to advance, its integration into dermatology holds promise for educational and diagnostic purposes. However, this study reveals nuances in the accuracy and recognition of AI-generated images compared to clinical photographs.</p><p>One notable aspect of the results is the varying recognition rates of AI-generated images, with DALLE-3 images being recognized as computer-generated more often than DALLE-2 images. This was seen across disease categories and across levels of training. This could be due to controversy surrounding these tools and implications regarding images that were too realistic such that later iterations of the software were made to generate images that were more cartoonish in nature. It is pertinent, however, that residents were better able to recognize computer-generated images, particularly the more photorealistic ones. This may indicate that the younger generation is more in tune with this technology and may have implications for its future use.</p><p>Nevertheless, it was seen that the later software, DALLE-3, produced more accurate images, with the correct diagnosis being assigned more often than with DALLE-3 images. DALLE-3 performed particularly well when generating images of pigmentary disorders. This indicates that the programs could be receiving more training data on this topic over the course of time and indicates promise for future programs. However, attending physicians were more likely to give the correct diagnosis for DALLE-2 generated images, suggesting that there may be more subtle signs present in these images that indicate the correct diagnosis. Neither program generated images that were correctly diagnosed as often as clinical images, indicating room for improvement on this front as well.</p><p>Finally, skin of color was represented only twice in the dataset. Although skin of color images could be generated upon instruction, it is evident that this is not the default setting for these tools. Considering that patients of color already face barriers to dermatological diagnoses and treatment [<xref ref-type="bibr" rid="ref7">7</xref>], it is imperative that training sets of images for any type of algorithm are inclusive and representative of a diverse population.</p><p>Limitations of this study include the use of AI image generators created by a single parent company as well as the use of a limited set of diagnoses. Further research is needed in this area with more advanced and specific training data. As with any new technology, careful research should be conducted as it improves to ensure that if and when it is utilized for patient care, it upholds standards of safety, accuracy, and equity.</p></sec></body><back><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Perlman</surname><given-names>KL</given-names> </name><name name-style="western"><surname>Williams</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Egbeto</surname><given-names>IA</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>DX</given-names> </name><name name-style="western"><surname>Siddiquee</surname><given-names>N</given-names> </name><name name-style="western"><surname>Park</surname><given-names>JH</given-names> </name></person-group><article-title>Skin of color lacks representation in medical student resources: a cross-sectional study</article-title><source>Int J Womens Dermatol</source><year>2021</year><month>03</month><volume>7</volume><issue>2</issue><fpage>195</fpage><lpage>196</lpage><pub-id pub-id-type="doi">10.1016/j.ijwd.2020.12.018</pub-id><pub-id pub-id-type="medline">33937492</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brinker</surname><given-names>TJ</given-names> </name><name name-style="western"><surname>Hekler</surname><given-names>A</given-names> </name><name name-style="western"><surname>Enk</surname><given-names>AH</given-names> </name><etal/></person-group><article-title>Deep learning outperformed 136 of 157 dermatologists in a head-to-head dermoscopic melanoma image classification task</article-title><source>Eur J Cancer</source><year>2019</year><month>05</month><volume>113</volume><fpage>47</fpage><lpage>54</lpage><pub-id pub-id-type="doi">10.1016/j.ejca.2019.04.001</pub-id><pub-id pub-id-type="medline">30981091</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="web"><article-title>Brown university</article-title><source>OpenAI&#x2019;s GPT-4 performs to a high degree on board-style dermatology questions - PubMed</source><access-date>2024-06-04</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://pubmed-ncbi-nlm-nih-gov.revproxy.brown.edu/38131454/">https://pubmed-ncbi-nlm-nih-gov.revproxy.brown.edu/38131454/</ext-link></comment></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ferreira</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Chu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Grant-Kels</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Ogunleye</surname><given-names>T</given-names> </name><name name-style="western"><surname>Lipoff</surname><given-names>JB</given-names> </name></person-group><article-title>Evaluation of ChatGPT dermatology responses to common patient queries</article-title><source>JMIR Dermatol</source><year>2023</year><month>11</month><day>17</day><volume>6</volume><fpage>e49280</fpage><pub-id pub-id-type="doi">10.2196/49280</pub-id><pub-id pub-id-type="medline">37976093</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Ramesh</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pavlov</surname><given-names>M</given-names> </name><name name-style="western"><surname>Goh</surname><given-names>G</given-names> </name><etal/></person-group><article-title>Zero-shot text-to-image generation</article-title><comment>Preprint posted online on  Feb 26, 2021</comment><pub-id pub-id-type="doi">10.48550/arXiv.2102.12092</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="web"><article-title>OpenAI</article-title><source>DALL&#x00B7;E 2</source><access-date>2023-02-23</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/dall-e-2/">https://openai.com/dall-e-2/</ext-link></comment></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hooper</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shao</surname><given-names>K</given-names> </name><name name-style="western"><surname>Feng</surname><given-names>H</given-names> </name></person-group><article-title>Racial/ethnic health disparities in dermatology in the United States, part 1: overview of contributing factors and management strategies</article-title><source>J Am Acad Dermatol</source><year>2022</year><month>10</month><volume>87</volume><issue>4</issue><fpage>723</fpage><lpage>730</lpage><pub-id pub-id-type="doi">10.1016/j.jaad.2021.12.061</pub-id><pub-id pub-id-type="medline">35143914</pub-id></nlm-citation></ref></ref-list></back></article>