Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

AI-Generated Images of Substance Use and Recovery: Mixed Methods Case Study

View through CrossRef
Abstract Background Images created with generative artificial intelligence (AI) tools are increasingly used for health communication due to their ease of use, speed, accessibility, and low cost. However, AI-generated images may bring practical and ethical risks to health practitioners and the public, including through the perpetuation of stigma against vulnerable and historically marginalized groups. Objective To understand the potential value of AI-generated images for health care and public health communication, we sought to analyze images of substance use disorder and recovery generated with ChatGPT. Specifically, we sought to investigate: (1) the default visual outputs produced in response to a range of prompts about substance use disorder and recovery, and (2) the extent to which prompt modification and guideline-informed prompting could mitigate potentially stigmatizing imagery. Methods We performed a mixed-methods case study examining depictions of substance use and recovery in images generated by ChatGPT 4.o. We generated images (n=84) using (1) prompts with colloquial and stigmatizing language, (2) prompts that follow best practices for person-first language, (3) image prompts written by ChatGPT, and (4) a custom GPT informed by guidelines for images of SUD. We then used a mixed-methods approach to analyze images for demographics and stigmatizing elements. Results Images produced in the default ChatGPT model featured primarily White men (81%, n=34). Further, images tended to be stigmatizing, featuring injection drug use, dark colors, and symbolic elements such as chains. These trends persisted even when person-first language prompts were used. Images produced by the guideline-informed custom GPT were markedly less stigmatizing; however, they featured almost only Black women (74%, n=31). Conclusions Our findings confirm prior research about stigma and biases in AI-generated images and extend this literature to substance use. However, our findings also suggest that (1) images can be improved when clear guidelines are provided and (2) even with guidelines, iteration is needed to create an image that fully concords with best practices.
Title: AI-Generated Images of Substance Use and Recovery: Mixed Methods Case Study
Description:
Abstract Background Images created with generative artificial intelligence (AI) tools are increasingly used for health communication due to their ease of use, speed, accessibility, and low cost.
However, AI-generated images may bring practical and ethical risks to health practitioners and the public, including through the perpetuation of stigma against vulnerable and historically marginalized groups.
Objective To understand the potential value of AI-generated images for health care and public health communication, we sought to analyze images of substance use disorder and recovery generated with ChatGPT.
Specifically, we sought to investigate: (1) the default visual outputs produced in response to a range of prompts about substance use disorder and recovery, and (2) the extent to which prompt modification and guideline-informed prompting could mitigate potentially stigmatizing imagery.
Methods We performed a mixed-methods case study examining depictions of substance use and recovery in images generated by ChatGPT 4.
o.
We generated images (n=84) using (1) prompts with colloquial and stigmatizing language, (2) prompts that follow best practices for person-first language, (3) image prompts written by ChatGPT, and (4) a custom GPT informed by guidelines for images of SUD.
We then used a mixed-methods approach to analyze images for demographics and stigmatizing elements.
Results Images produced in the default ChatGPT model featured primarily White men (81%, n=34).
Further, images tended to be stigmatizing, featuring injection drug use, dark colors, and symbolic elements such as chains.
These trends persisted even when person-first language prompts were used.
Images produced by the guideline-informed custom GPT were markedly less stigmatizing; however, they featured almost only Black women (74%, n=31).
Conclusions Our findings confirm prior research about stigma and biases in AI-generated images and extend this literature to substance use.
However, our findings also suggest that (1) images can be improved when clear guidelines are provided and (2) even with guidelines, iteration is needed to create an image that fully concords with best practices.

Related Results

Hydatid Disease of The Brain Parenchyma: A Systematic Review
Hydatid Disease of The Brain Parenchyma: A Systematic Review
Abstarct Introduction Isolated brain hydatid disease (BHD) is an extremely rare form of echinococcosis. A prompt and timely diagnosis is a crucial step in disease management. This ...
Current therapeutic strategies for erectile function recovery after radical prostatectomy – literature review and meta-analysis
Current therapeutic strategies for erectile function recovery after radical prostatectomy – literature review and meta-analysis
Radical prostatectomy is the most commonly performed treatment option for localised prostate cancer. In the last decades the surgical technique has been improved and modified in or...
Breast Carcinoma within Fibroadenoma: A Systematic Review
Breast Carcinoma within Fibroadenoma: A Systematic Review
Abstract Introduction Fibroadenoma is the most common benign breast lesion; however, it carries a potential risk of malignant transformation. This systematic review provides an ove...
Hydatid Cyst of The Orbit: A Systematic Review with Meta-Data
Hydatid Cyst of The Orbit: A Systematic Review with Meta-Data
Abstarct Introduction Orbital hydatid cysts (HCs) constitute less than 1% of all cases of hydatidosis, yet their occurrence is often linked to severe visual complications. This stu...

Back to Top