TY - GEN
T1 - "I wouldn't say offensive but⋯"
T2 - 6th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2023
AU - Gadiraju, Vinitha
AU - Kane, Shaun
AU - Dev, Sunipa
AU - Taylor, Alex
AU - Wang, Ding
AU - Denton, Emily
AU - Brewer, Robin
N1 - Funding Information:
We thank members of the Technology, AI, Society, and Culture (TASC), People and AI Research (PAIR), and many other teams at Google, for their invaluable feedback on this work. We also thank our participants for collaborating with us on this project: Minh Ha, Imran Ahmed, Liza McCollum, Aziza Rodriguez, Lucy Gogolushko, Liv Milner, Nicole harris, Eric Dixon, Matthew Janusauskas, Timm Sinnen, Carlos Mitchell, Ruthie Clark, Pauline Perera, Brenda Gutierrez Baeza, Kalani Helekunihi, Sarah Diaz, Loop Salazar, Julie Bothe, Shari Eberts, Jon Taylor, Sonny Wasilowski, Brett Schuilwerve, Carol E Conway, Kristen McDevitt, Sara Montgomery, Steve Lu, Earl Dillon, Sadeepa Munasinghe, Christopher Reardon, Jana Schroeder, Denise Lance, Lisan Hasnain, Blake Sinnett, Hai Nguyên Ly, Hank Chiuppi, Mike Reiser, Doug Langley, Brandon Misch, and those who prefer to remain anonymous.
Publisher Copyright:
© 2023 Owner/Author.
PY - 2023/6/12
Y1 - 2023/6/12
N2 - Large language models (LLMs) trained on real-world data can inadvertently reflect harmful societal biases, particularly toward historically marginalized communities. While previous work has primarily focused on harms related to age and race, emerging research has shown that biases toward disabled communities exist. This study extends prior work exploring the existence of harms by identifying categories of LLM-perpetuated harms toward the disability community. We conducted 19 focus groups, during which 56 participants with disabilities probed a dialog model about disability and discussed and annotated its responses. Participants rarely characterized model outputs as blatantly offensive or toxic. Instead, participants used nuanced language to detail how the dialog model mirrored subtle yet harmful stereotypes they encountered in their lives and dominant media, e.g., inspiration porn and able-bodied saviors. Participants often implicated training data as a cause for these stereotypes and recommended training the model on diverse identities from disability-positive resources. Our discussion further explores representative data strategies to mitigate harm related to different communities through annotation co-design with ML researchers and developers.
AB - Large language models (LLMs) trained on real-world data can inadvertently reflect harmful societal biases, particularly toward historically marginalized communities. While previous work has primarily focused on harms related to age and race, emerging research has shown that biases toward disabled communities exist. This study extends prior work exploring the existence of harms by identifying categories of LLM-perpetuated harms toward the disability community. We conducted 19 focus groups, during which 56 participants with disabilities probed a dialog model about disability and discussed and annotated its responses. Participants rarely characterized model outputs as blatantly offensive or toxic. Instead, participants used nuanced language to detail how the dialog model mirrored subtle yet harmful stereotypes they encountered in their lives and dominant media, e.g., inspiration porn and able-bodied saviors. Participants often implicated training data as a cause for these stereotypes and recommended training the model on diverse identities from disability-positive resources. Our discussion further explores representative data strategies to mitigate harm related to different communities through annotation co-design with ML researchers and developers.
KW - algorithmic harms
KW - artificial intelligence
KW - chatbot
KW - data annotation
KW - dialog model
KW - disability representation
KW - large language models
KW - qualitative
UR - http://www.scopus.com/inward/record.url?scp=85163677708&partnerID=8YFLogxK
U2 - 10.1145/3593013.3593989
DO - 10.1145/3593013.3593989
M3 - Conference contribution
AN - SCOPUS:85163677708
T3 - ACM International Conference Proceeding Series
SP - 205
EP - 216
BT - Proceedings of the 6th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2023
PB - Association for Computing Machinery
Y2 - 12 June 2023 through 15 June 2023
ER -