No Word Embedding Model Is Perfect: Evaluating the Representation Accuracy for Social Bias in the Media

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

External Research Organisations

  • Universität Hamburg
View graph of relations

Details

Original languageEnglish
Title of host publicationProceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022)
Pages2081-2093
Number of pages13
Publication statusPublished - Dec 2022
Event2022 Findings of the Association for Computational Linguistics: EMNLP 2022 - Abu Dhabi, United Arab Emirates
Duration: 7 Dec 202211 Dec 2022

Abstract

News articles both shape and reflect public opinion across the political spectrum. Analyzing them for social bias can thus provide valuable insights, such as prevailing stereotypes in society and the media, which are often adopted by NLP models trained on respective data. Recent work has relied on word embedding bias measures, such as WEAT. However, several representation issues of embeddings can harm the measures’ accuracy, including low-resource settings and token frequency differences. In this work, we study what kind of embedding algorithm serves best to accurately measure types of social bias known to exist in US online news articles. To cover the whole spectrum of political bias in the US, we collect 500k articles and review psychology literature with respect to expected social bias. We then quantify social bias using WEAT along with embedding algorithms that account for the aforementioned issues. We compare how models trained with the algorithms on news articles represent the expected social bias. Our results suggest that the standard way to quantify bias does not align well with knowledge from psychology. While the proposed algorithms reduce the gap, they still do not fully match the literature.

ASJC Scopus subject areas

Sustainable Development Goals

Cite this

No Word Embedding Model Is Perfect: Evaluating the Representation Accuracy for Social Bias in the Media. / Spliethöver, Maximilian; Keiff, Maximilian; Wachsmuth, Henning.
Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022). 2022. p. 2081-2093.

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Spliethöver, M, Keiff, M & Wachsmuth, H 2022, No Word Embedding Model Is Perfect: Evaluating the Representation Accuracy for Social Bias in the Media. in Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022). pp. 2081-2093, 2022 Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, 7 Dec 2022. https://doi.org/10.18653/v1/2022.findings-emnlp.152
Spliethöver, M., Keiff, M., & Wachsmuth, H. (2022). No Word Embedding Model Is Perfect: Evaluating the Representation Accuracy for Social Bias in the Media. In Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022) (pp. 2081-2093) https://doi.org/10.18653/v1/2022.findings-emnlp.152
Spliethöver M, Keiff M, Wachsmuth H. No Word Embedding Model Is Perfect: Evaluating the Representation Accuracy for Social Bias in the Media. In Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022). 2022. p. 2081-2093 Epub 2022 Nov 7. doi: 10.18653/v1/2022.findings-emnlp.152
Spliethöver, Maximilian ; Keiff, Maximilian ; Wachsmuth, Henning. / No Word Embedding Model Is Perfect : Evaluating the Representation Accuracy for Social Bias in the Media. Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022). 2022. pp. 2081-2093
Download
@inproceedings{106c31318af7407e94ca8519a2c22cb4,
title = "No Word Embedding Model Is Perfect: Evaluating the Representation Accuracy for Social Bias in the Media",
abstract = "News articles both shape and reflect public opinion across the political spectrum. Analyzing them for social bias can thus provide valuable insights, such as prevailing stereotypes in society and the media, which are often adopted by NLP models trained on respective data. Recent work has relied on word embedding bias measures, such as WEAT. However, several representation issues of embeddings can harm the measures{\textquoteright} accuracy, including low-resource settings and token frequency differences. In this work, we study what kind of embedding algorithm serves best to accurately measure types of social bias known to exist in US online news articles. To cover the whole spectrum of political bias in the US, we collect 500k articles and review psychology literature with respect to expected social bias. We then quantify social bias using WEAT along with embedding algorithms that account for the aforementioned issues. We compare how models trained with the algorithms on news articles represent the expected social bias. Our results suggest that the standard way to quantify bias does not align well with knowledge from psychology. While the proposed algorithms reduce the gap, they still do not fully match the literature.",
author = "Maximilian Splieth{\"o}ver and Maximilian Keiff and Henning Wachsmuth",
note = "Publisher Copyright: {\textcopyright} 2022 Association for Computational Linguistics.; 2022 Findings of the Association for Computational Linguistics: EMNLP 2022 ; Conference date: 07-12-2022 Through 11-12-2022",
year = "2022",
month = dec,
doi = "10.18653/v1/2022.findings-emnlp.152",
language = "English",
pages = "2081--2093",
booktitle = "Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022)",

}

Download

TY - GEN

T1 - No Word Embedding Model Is Perfect

T2 - 2022 Findings of the Association for Computational Linguistics: EMNLP 2022

AU - Spliethöver, Maximilian

AU - Keiff, Maximilian

AU - Wachsmuth, Henning

N1 - Publisher Copyright: © 2022 Association for Computational Linguistics.

PY - 2022/12

Y1 - 2022/12

N2 - News articles both shape and reflect public opinion across the political spectrum. Analyzing them for social bias can thus provide valuable insights, such as prevailing stereotypes in society and the media, which are often adopted by NLP models trained on respective data. Recent work has relied on word embedding bias measures, such as WEAT. However, several representation issues of embeddings can harm the measures’ accuracy, including low-resource settings and token frequency differences. In this work, we study what kind of embedding algorithm serves best to accurately measure types of social bias known to exist in US online news articles. To cover the whole spectrum of political bias in the US, we collect 500k articles and review psychology literature with respect to expected social bias. We then quantify social bias using WEAT along with embedding algorithms that account for the aforementioned issues. We compare how models trained with the algorithms on news articles represent the expected social bias. Our results suggest that the standard way to quantify bias does not align well with knowledge from psychology. While the proposed algorithms reduce the gap, they still do not fully match the literature.

AB - News articles both shape and reflect public opinion across the political spectrum. Analyzing them for social bias can thus provide valuable insights, such as prevailing stereotypes in society and the media, which are often adopted by NLP models trained on respective data. Recent work has relied on word embedding bias measures, such as WEAT. However, several representation issues of embeddings can harm the measures’ accuracy, including low-resource settings and token frequency differences. In this work, we study what kind of embedding algorithm serves best to accurately measure types of social bias known to exist in US online news articles. To cover the whole spectrum of political bias in the US, we collect 500k articles and review psychology literature with respect to expected social bias. We then quantify social bias using WEAT along with embedding algorithms that account for the aforementioned issues. We compare how models trained with the algorithms on news articles represent the expected social bias. Our results suggest that the standard way to quantify bias does not align well with knowledge from psychology. While the proposed algorithms reduce the gap, they still do not fully match the literature.

UR - http://www.scopus.com/inward/record.url?scp=85149869212&partnerID=8YFLogxK

U2 - 10.18653/v1/2022.findings-emnlp.152

DO - 10.18653/v1/2022.findings-emnlp.152

M3 - Conference contribution

SP - 2081

EP - 2093

BT - Proceedings of The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022)

Y2 - 7 December 2022 through 11 December 2022

ER -

By the same author(s)