On the Interpretability and Significance of Bias Metrics in Texts: a PMI-based Approach
Francisco Valentini, Germán Federico Rosati, Damián Blasi, Diego Fernandez Slezak, Edgar Altszyler
Main: Computational Social Science and Cultural Analytics Main-poster Paper
Poster Session 7: Computational Social Science and Cultural Analytics (Poster)
Conference Room: Frontenac Ballroom and Queen's Quay
Conference Time: July 12, 11:00-12:30 (EDT) (America/Toronto)
Global Time: July 12, Poster Session 7 (15:00-16:30 UTC)
Keywords:
nlp tools for social analysis
TLDR:
In recent years, word embeddings have been widely used to measure biases in texts. Even if they have proven to be effective in detecting a wide variety of biases, metrics based on word embeddings lack transparency and interpretability. We analyze an alternative PMI-based metric to quantify biases in...
You can open the
#paper-P5770
channel in a separate window.
Abstract:
In recent years, word embeddings have been widely used to measure biases in texts. Even if they have proven to be effective in detecting a wide variety of biases, metrics based on word embeddings lack transparency and interpretability. We analyze an alternative PMI-based metric to quantify biases in texts. It can be expressed as a function of conditional probabilities, which provides a simple interpretation in terms of word co-occurrences. We also prove that it can be approximated by an odds ratio, which allows estimating confidence intervals and statistical significance of textual biases. This approach produces similar results to metrics based on word embeddings when capturing gender gaps of the real world embedded in large corpora.