Abstract / Description of output
This presentation reports on a case study investigating how Natural
Language Processing, a field that applies computational methods
such as Machine Learning to human-written texts, can support
the measurement and evaluation of gender biased language
in archival catalogs. Working with English descriptions from the
catalog metadata of the University of Edinburgh’s Archives, we
created an annotated dataset and classification models that identify
gender biases in the descriptions. Conducted with archival data,
the case study holds relevance across Galleries, Libraries, Archives,
and Museums (GLAM), particularly for institutions with
catalog descriptions in English. In addition to bringing Natural
Language Processing (NLP) methods to Archives, we identified
opportunities to bring Archival Science methods, such as Cultural
Humility (Tai, 2021) and Feminist Standpoint Appraisal (Caswell,
2022), to NLP. Through this two-way disciplinary exchange,
we demonstrate how Humanistic approaches to bias and uncertainty
can upend legacies of gender-based oppression that most
computational approaches to date uphold when working with data
at scale.
Language Processing, a field that applies computational methods
such as Machine Learning to human-written texts, can support
the measurement and evaluation of gender biased language
in archival catalogs. Working with English descriptions from the
catalog metadata of the University of Edinburgh’s Archives, we
created an annotated dataset and classification models that identify
gender biases in the descriptions. Conducted with archival data,
the case study holds relevance across Galleries, Libraries, Archives,
and Museums (GLAM), particularly for institutions with
catalog descriptions in English. In addition to bringing Natural
Language Processing (NLP) methods to Archives, we identified
opportunities to bring Archival Science methods, such as Cultural
Humility (Tai, 2021) and Feminist Standpoint Appraisal (Caswell,
2022), to NLP. Through this two-way disciplinary exchange,
we demonstrate how Humanistic approaches to bias and uncertainty
can upend legacies of gender-based oppression that most
computational approaches to date uphold when working with data
at scale.
Original language | English |
---|---|
Pages | 267-270 |
Publication status | Published - 1 Jul 2023 |