Sarah Wyer sarah.wyer@durham.ac.uk
PGR Student Doctor of Philosophy
Sarah Wyer sarah.wyer@durham.ac.uk
PGR Student Doctor of Philosophy
Professor Sue Black sue.black@durham.ac.uk
Professor
This study explores the occurrence and implications of sexualized violence against women in text completion tasks performed by GPT-3 models. The study began as an exploratory investigation into gender inequalities within GPT-3 models to discover what GPT3 models generate in text completion tasks when prompted with basic gendered prompts. For example, "Men can…", "Women can…". Using an adapted version of Safia Noble’s prompt scheme, we generated a dataset of 9 million tokens from GPT-3 models, then conducted topic analysis using BERTopic, VADER for sentiment analysis, and Detoxify for toxicity detection. The results were unexpected. Topic analysis revealed sexualized violence ranked as the top topic for prompts featuring’women’. Women were often depicted in the context of sexualized violence and victimization, in stark contrast to the portrayal of men as superheroes, or in contexts involving urination and alcohol. For example, a randomly selected output generated by the prompt. "Why are women so", generated "Why are women so arrogant about being raped?". Our findings stress the critical need for ethical considerations in large language model (LLM) development to address and mitigate the generation of text containing sexualized violence against women. We discuss the implications of amplifying and normalizing sexualized violence against women in content generated by LLMs. Our work builds on previous research examining gender bias in LLMs, with a specific focus on the manifestation of sexualized violence against women in LLM outputs, an area that has received little attention. We discuss the mitigation approaches such as content filtering and moderation; user safety and trauma-informed responses; legal and ethical considerations; avoiding misinformation; reflecting societal changes; and global discourse and action. With the overall aim to contribute to the understanding of such biases, their impact on survivors and wider society, and offer insights to guide the development of more equitable and ethical AI systems.
Wyer, S., & Black, S. (online). Algorithmic bias: sexualized violence against women in GPT-3 models. AI and Ethics, https://doi.org/10.1007/s43681-024-00641-0
Journal Article Type | Article |
---|---|
Acceptance Date | Nov 24, 2024 |
Online Publication Date | Jan 15, 2025 |
Deposit Date | Jan 20, 2025 |
Publicly Available Date | Jan 20, 2025 |
Journal | AI and Ethics |
Electronic ISSN | 2730-5961 |
Publisher | Springer |
Peer Reviewed | Peer Reviewed |
DOI | https://doi.org/10.1007/s43681-024-00641-0 |
Public URL | https://durham-repository.worktribe.com/output/3342589 |
Published Journal Article (Advance Online Version)
(3.1 Mb)
PDF
Publisher Licence URL
http://creativecommons.org/licenses/by/4.0/
Self-Regulated Sample Diversity in Large Language Models
(2024)
Presentation / Conference Contribution
Empirical comparison of text-based mobile apps similarity measurement techniques
(2019)
Journal Article
Digital Inclusion in Nothern England: Training Women from Underrepresented Communities in Tech: A Data Analytics Case Study
(2020)
Presentation / Conference Contribution
Clustering Mobile Apps Based on Mined Textual Features
(2016)
Presentation / Conference Contribution
Optimising IT Security Research via a Low Cost, Instantly Available, Cloud Based Cyber Range
(2024)
Presentation / Conference Contribution
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
Apache License Version 2.0 (http://www.apache.org/licenses/)
Apache License Version 2.0 (http://www.apache.org/licenses/)
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search