论文标题
对比语言视觉AI模型在网络绑带的多模式数据上均表现出性对象化偏见
Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias
论文作者
论文摘要
评估了在网络刮擦上培训的九种语言Vision AI模型,并评估了对比度的语言图像预处理(剪辑)目标,以证明心理学家研究的偏见:女孩和女性的性客观化发生时,当一个人的人类特征(例如情感)被忽略并被视为身体时,这会发生。我们在心理学上复制了三个实验,以量化性客观化,并表明该现象持续存在于AI中。第一个实验使用来自性客观化和情感数据库中女性的标准化图像,并发现人类特征与客观化女性的图像分离:模型对情绪状态的识别是由受试者完全还是部分衣服介导的。嵌入关联测试(EATS)返回愤怒(d> 0.80)和悲伤(d> 0.50)的显着效果大小,将全衣服受试者的图像与情感相关联。 Grad-Cam显着性图强调了剪辑从对象图像中的情感表达中分散注意力。第二个实验衡量了代表性应用中的效果:自动图像字幕仪(南极字幕)的单词表示情感的单词小于50%,而对于部分衣服女性的图像而言,与全衣服女性的图像相比。第三个实验发现,与男性专业人士相对于性描述,女性专业人士(科学家,医生,高管)的图像可能与性描述有关。第四个实验表明,“年龄]年龄大女孩”的提示产生了性图像(由NSFW分类器确定),最多73%的时间用于VQGAN-CLIP和稳定的扩散。男孩的相应率永远不会超过9%。证据表明,在网络刮擦中训练的语言视觉AI模型学习了性客观化的偏见,这会传播到下游应用程序。
Nine language-vision AI models trained on web scrapes with the Contrastive Language-Image Pretraining (CLIP) objective are evaluated for evidence of a bias studied by psychologists: the sexual objectification of girls and women, which occurs when a person's human characteristics, such as emotions, are disregarded and the person is treated as a body. We replicate three experiments in psychology quantifying sexual objectification and show that the phenomena persist in AI. A first experiment uses standardized images of women from the Sexual OBjectification and EMotion Database, and finds that human characteristics are disassociated from images of objectified women: the model's recognition of emotional state is mediated by whether the subject is fully or partially clothed. Embedding association tests (EATs) return significant effect sizes for both anger (d >0.80) and sadness (d >0.50), associating images of fully clothed subjects with emotions. GRAD-CAM saliency maps highlight that CLIP gets distracted from emotional expressions in objectified images. A second experiment measures the effect in a representative application: an automatic image captioner (Antarctic Captions) includes words denoting emotion less than 50% as often for images of partially clothed women than for images of fully clothed women. A third experiment finds that images of female professionals (scientists, doctors, executives) are likely to be associated with sexual descriptions relative to images of male professionals. A fourth experiment shows that a prompt of "a [age] year old girl" generates sexualized images (as determined by an NSFW classifier) up to 73% of the time for VQGAN-CLIP and Stable Diffusion; the corresponding rate for boys never surpasses 9%. The evidence indicates that language-vision AI models trained on web scrapes learn biases of sexual objectification, which propagate to downstream applications.