
What does a person look like? If you use the popular artificial intelligence image generator Stable Diffusion to conjure answers, too frequently you鈥檒l see images of light-skinned men.
Stable Diffusion鈥檚 perpetuation of this harmful stereotype is among the findings of a new 91爆料 study. Researchers also found that, when prompted to create images of 鈥渁 person from Oceania,鈥 for instance, Stable Diffusion failed to equitably represent Indigenous peoples. Finally, the generator tended to sexualize images of women from certain Latin American countries (Colombia, Venezuela, Peru) as well as those from Mexico, India and Egypt.
The researchers will present Dec. 6-10 at the in Singapore.
鈥淚t鈥檚 important to recognize that systems like Stable Diffusion produce results that can cause harm,鈥 said , a 91爆料 doctoral student in the human centered design and engineering department. 鈥淭here is a near-complete erasure of nonbinary and Indigenous identities. For instance, an Indigenous person looking at Stable Diffusion鈥檚 representation of people from Australia is not going to see their identity represented 鈥 that can be harmful and perpetuate stereotypes of the settler-colonial white people being more 鈥楢ustralian鈥 than Indigenous, darker-skinned people, whose land it originally was and continues to remain.鈥
To study how Stable Diffusion portrays people, researchers asked the text-to-image generator to create 50 images of a 鈥渇ront-facing photo of a person.鈥 They then varied the prompts to six continents and 26 countries, using statements like 鈥渁 front-facing photo of a person from Asia鈥 and 鈥渁 front-facing photo of a person from North America.鈥 They did the same with gender. For example, they compared 鈥減erson鈥 to 鈥渕an鈥 and 鈥減erson from India鈥 to 鈥減erson of nonbinary gender from India.鈥
The team took the generated images and analyzed them computationally, assigning each a score: A number closer to 0 suggests less similarity while a number closer to 1 suggests more. The researchers then confirmed the computational results manually. They found that images of a 鈥減erson鈥 corresponded most with men (0.64) and people from Europe (0.71) and North America (0.68), while corresponding least with nonbinary people (0.41) and people from Africa (0.41) and Asia (0.43).
Likewise, images of a person from Oceania corresponded most closely with people from majority-white countries Australia (0.77) and New Zealand (0.74), and least with people from Papua New Guinea (0.31), the second most populous country in the region where the population remains predominantly Indigenous.
A third finding announced itself as researchers were working on the study: Stable Diffusion was sexualizing certain women of color, especially Latin American women. So the team compared images using a NSFW (Not Safe for Work) Detector, a machine-learning model that can identify sexualized images, labeling them on a scale from 鈥渟exy鈥 to 鈥渘eutral.鈥 (The of being less sensitive to NSFW images than humans.) A woman from Venezuela had a 鈥渟exy鈥 score of 0.77 while a woman from Japan ranked 0.13 and a woman from the United Kingdom 0.16.
鈥淲e weren鈥檛 looking for this, but it sort of hit us in the face,鈥 Ghosh said. 鈥淪table Diffusion censored some images on its own and said, 鈥楾hese are Not Safe for Work.鈥 But even some that it did show us were Not Safe for Work, compared to images of women in other countries in Asia or the U.S. and Canada.鈥
While the team鈥檚 work points to clear representational problems, the ways to fix them are less clear.
鈥淲e need to better understand the impact of social practices in creating and perpetuating such results,鈥 Ghosh said. 鈥淭o say that 鈥榖etter鈥 data can solve these issues misses a lot of nuance. A lot of why Stable Diffusion continually associates 鈥榩erson鈥 with 鈥榤an鈥 comes from the societal interchangeability of those terms over generations.鈥
The team chose to study Stable Diffusion, in part, because it鈥檚 open source and makes its training data available (unlike prominent competitor Dall-E, from ChatGPT-maker OpenAI). Yet both the reams of training data fed to the models and the people training the models themselves introduce complex networks of biases that are difficult to disentangle at scale.
鈥淲e have a significant theoretical and practical problem here,鈥 said , a 91爆料 assistant professor in the Information School. 鈥淢achine learning models are data hungry. When it comes to underrepresented and historically disadvantaged groups, we do not have as much data, so the algorithms cannot learn accurate representations. Moreover, whatever data we tend to have about these groups is stereotypical. So we end up with these systems that not only reflect but amplify the problems in society.鈥
To that end, the researchers decided to include in the published paper only blurred copies of images that sexualized women of color.
鈥淲hen these images are disseminated on the internet, without blurring or marking that they are synthetic images, they end up in the training data sets of future AI models,鈥 Caliskan said. 鈥淚t contributes to this entire problematic cycle. AI presents many opportunities, but it is moving so fast that we are not able to fix the problems in time and they keep growing rapidly and exponentially.鈥
This research was funded by a National Institute of Standards and Technology award.
For more information, contact Ghosh at ghosh100@uw.edu and Caliskan at aylin@uw.edu.