@jaeschke

Debiasing Word Embeddings from Sentiment Associations in Names

, , and . Proceedings of the 13th International Conference on Web Search and Data Mining, ACM, (January 2020)
DOI: 10.1145/3336191.3371779

Abstract

Word embeddings, trained through models like skip-gram, have shown to be prone to capturing the biases from the training corpus, e.g. gender bias. Such biases are unwanted as they spill in downstream tasks, thus, leading to discriminatory behavior. In this work, we address the problem of prior sentiment associated with names in word embeddings where for a given name representation (e.g. "Smith"), a sentiment classifier will categorize it as either positive or negative. We propose DebiasEmb, a skip-gram based word embedding approach that, for a given oracle sentiment classification model, will debias the name representations, such that they cannot be associated with either positive or negative sentiment. Evaluation on standard word embedding benchmarks and a downstream analysis show that our approach is able to maintain a high quality of embeddings and at the same time mitigate sentiment bias in name embeddings.

Description

Debiasing Word Embeddings from Sentiment Associations in Names | Proceedings of the 13th International Conference on Web Search and Data Mining

Links and resources

Tags

community

  • @jaeschke
  • @sgottschalk
  • @dblp
@jaeschke's tags highlighted