Logo Logo

Wiegand, Tim Luca Till; Jung, Leonard Ben; Gudera, Jonas Anton; Schuhmacher, Luisa Sophie; Moehrle, Paulina; Rischewski, Jon Felix; Mehrzad, Pardiss; Jeong, Subin; Nguyen, Lisa Ha; Poeschla, Michael; Velezmoro, Laura Isabella; Kruk, Linus; Dimitriadis, Konstantinos; Koerte, Inga Katharina (2025): Demographic inaccuracies and biases in the depiction of patients by artificial intelligence text-to-image generators. npj Digital Medicine, 8: 459. ISSN 2398-6352

[thumbnail of s41746-025-01817-6.pdf] Creative Commons Namensnennung (CC BY)
Veröffentlichte Publikation
s41746-025-01817-6.pdf

Abstract

The wide usage of artificial intelligence (AI) text-to-image generators raises concerns about the role of AI in amplifying misconceptions in healthcare. This study therefore evaluated the demographic accuracy and potential biases in the depiction of patients by four commonly used text-to-image generators. A total of 9060 images of patients with 29 different diseases was generated using Adobe Firefly, Bing Image Generator, Meta Imagine, and Midjourney. Twelve independent raters determined the sex, age, weight, and race and ethnicity of the patients depicted. Comparison to the real-world epidemiology showed that the generated images failed to depict demographical characteristics such as sex, age, and race and ethnicity accurately. In addition, we observed an over-representation of White and normal weight individuals. Inaccuracies and biases may stem from non-representative and non-specific training data as well as insufficient or misdirected bias mitigation strategies. In consequence, new strategies to counteract such inaccuracies and biases are needed.

Publikation bearbeiten
Publikation bearbeiten