%0 Journal Article %D 2020 %T Causal inference for social discrimination reasoning %A Qureshi, Bilal %A Kamiran, Faisal %A Karim, Asim %A Salvatore Ruggieri %A Dino Pedreschi %X The discovery of discriminatory bias in human or automated decision making is a task of increasing importance and difficulty, exacerbated by the pervasive use of machine learning and data mining. Currently, discrimination discovery largely relies upon correlation analysis of decisions records, disregarding the impact of confounding biases. We present a method for causal discrimination discovery based on propensity score analysis, a statistical tool for filtering out the effect of confounding variables. We introduce causal measures of discrimination which quantify the effect of group membership on the decisions, and highlight causal discrimination/favoritism patterns by learning regression trees over the novel measures. We validate our approach on two real world datasets. Our proposed framework for causal discrimination has the potential to enhance the transparency of machine learning with tools for detecting discriminatory bias both in the training data and in the learning algorithms. %V 54 %P 425 - 437 %8 2020/04/01 %@ 1573-7675 %G eng %U https://link.springer.com/article/10.1007/s10844-019-00580-x %! Journal of Intelligent Information Systems %R https://doi.org/10.1007/s10844-019-00580-x %0 Journal Article %J Journal of Intelligent Information Systems %D 2019 %T Causal inference for social discrimination reasoning %A Qureshi, Bilal %A Kamiran, Faisal %A Karim, Asim %A Salvatore Ruggieri %A Dino Pedreschi %X The discovery of discriminatory bias in human or automated decision making is a task of increasing importance and difficulty, exacerbated by the pervasive use of machine learning and data mining. Currently, discrimination discovery largely relies upon correlation analysis of decisions records, disregarding the impact of confounding biases. We present a method for causal discrimination discovery based on propensity score analysis, a statistical tool for filtering out the effect of confounding variables. We introduce causal measures of discrimination which quantify the effect of group membership on the decisions, and highlight causal discrimination/favoritism patterns by learning regression trees over the novel measures. We validate our approach on two real world datasets. Our proposed framework for causal discrimination has the potential to enhance the transparency of machine learning with tools for detecting discriminatory bias both in the training data and in the learning algorithms. %B Journal of Intelligent Information Systems %P 1–13 %G eng %U https://link.springer.com/article/10.1007/s10844-019-00580-x %R 10.1007/s10844-019-00580-x %0 Journal Article %J arXiv preprint arXiv:1608.03735 %D 2016 %T Causal Discrimination Discovery Through Propensity Score Analysis %A Qureshi, Bilal %A Kamiran, Faisal %A Karim, Asim %A Salvatore Ruggieri %X Social discrimination is considered illegal and unethical in the modern world. Such discrimination is often implicit in observed decisions' datasets, and anti-discrimination organizations seek to discover cases of discrimination and to understand the reasons behind them. Previous work in this direction adopted simple observational data analysis; however, this can produce biased results due to the effect of confounding variables. In this paper, we propose a causal discrimination discovery and understanding approach based on propensity score analysis. The propensity score is an effective statistical tool for filtering out the effect of confounding variables. We employ propensity score weighting to balance the distribution of individuals from protected and unprotected groups w.r.t. the confounding variables. For each individual in the dataset, we quantify its causal discrimination or favoritism with a neighborhood-based measure calculated on the balanced distributions. Subsequently, the causal discrimination/favoritism patterns are understood by learning a regression tree. Our approach avoids common pitfalls in observational data analysis and make its results legally admissible. We demonstrate the results of our approach on two discrimination datasets. %B arXiv preprint arXiv:1608.03735 %G eng %U https://arxiv.org/abs/1608.03735