Currie Geoffrey, Hewis Johnathan, Ebbs Phillip
Charles Sturt University, Wagga Wagga, Australia; Baylor College of Medicine, Houston, USA.
Charles Sturt University, Port Macquarie, Australia.
Australas Emerg Care. 2025 Jun;28(2):103-109. doi: 10.1016/j.auec.2024.11.003. Epub 2024 Dec 2.
In Australia, almost 50 % of paramedics are female yet they remain under-represented in stereotypical depictions of the profession. The potentially transformative value of generative artificial intelligence (AI) may be limited by stereotypical errors, misrepresentations and bias. Increasing use of text-to-image generative AI, like DALL-E 3, could reinforce gender and ethnicity biases and, therefore, is important to objectively evaluate.
In March 2024, DALL-E 3 was utilised via GPT-4 to generate a series of individual and group images of Australian paramedics, ambulance officers, police officers and firefighters. In total, 82 images were produced including 60 individual-character images, and 22 multiple-character group images. All 326 depicted characters were independently analysed by three reviewers for apparent gender, age, skin tone and ethnicity.
Among first responders, 90.8 % (N = 296) were depicted as male, 90.5 % (N = 295) as Caucasian, 95.7 % (N = 312) as a light skin tone, and 94.8 % (N = 309) as under 55 years of age. For paramedics and police the gender distribution was a statistically significant variation from that of actual Australian workforce data (all p < 0.001). Among the images of individual paramedics and ambulance officers (N = 32), DALL-E 3 depicted 100 % as male, 100 % as Caucasian and 100 % with light skin tone.
Gender and ethnicity bias is a significant limitation for text-to-image generative AI using DALL-E 3 among Australian first responders. Generated images have a disproportionately high misrepresentation of males, Caucasians and light skin tones that are not representative of the diversity of paramedics in Australia today.
在澳大利亚,近50%的护理人员为女性,但在该职业的刻板印象描绘中,她们的代表性仍然不足。生成式人工智能(AI)的潜在变革价值可能会受到刻板印象错误、错误表述和偏见的限制。像DALL-E 3这样的文本到图像生成式AI的使用日益增加,可能会强化性别和种族偏见,因此,客观评估很重要。
2024年3月,通过GPT-4使用DALL-E 3生成了一系列澳大利亚护理人员、救护人员、警察和消防员的个人及群体图像。总共生成了82张图像,包括60张个人角色图像和22张多角色群体图像。三位评审员对所有326个描绘的角色进行了独立分析,以确定其明显的性别、年龄、肤色和种族。
在急救人员中,90.8%(N = 296)被描绘为男性,90.5%(N = 295)为白种人,95.7%(N = 312)为浅肤色,94.8%(N = 309)年龄在55岁以下。护理人员和警察的性别分布与澳大利亚实际劳动力数据存在统计学上的显著差异(所有p < 0.001)。在个人护理人员和救护人员的图像中(N = 32),DALL-E 3将其100%描绘为男性,100%为白种人,100%为浅肤色。
性别和种族偏见是使用DALL-E 3的文本到图像生成式AI在澳大利亚急救人员中的一个重大限制。生成的图像对男性、白种人和浅肤色的错误表述比例过高,不能代表当今澳大利亚护理人员的多样性。