This repository is the implementation of the paper Fairness Aware Counterfactuals for Subgroups (FACTS). FACTS is a framework for auditing subgroup fairness through counterfactual explanations. We aim to (a) formulate different aspects of the difficulty of individuals in certain subgroups to achieve recourse, i.e. receive the desired outcome, either at the micro level, considering members of the subgroup individually, or at the macro level, considering the subgroup as a whole, and (b) introduce notions of subgroup fairness that are robust, if not totally oblivious, to the cost of achieving recourse. Below, appears a subgroup audited by one of our fairness metrics.
In our work, we call the above representation "Comparative Subgroup Counterfactuals". These if-then rules, along with the effectiveness that each action manages to achieve (micro or micro, see Section 2.2 of our paper) give, in our opinion, a very clear and intuitive insight into certain type of bias that a machine learning model may exhibit.
All experiments were run on the Anaconda platform using python version 3.9.16. To avoid bugs due to incompatible package versions, we have exported the yaml file for the conda environment on which we worked.
To create a conda environment with the same configuration, run:
conda env create --name facts --file conda_env.yml
and then activate it with
conda activate facts
A model is needed for auditing purposes. We trained, in our paper, a logistic regression classifier. Any other classification model could be used.
Specifically, our method expects a model with the form of facts.models.ModelAPI
, which means simply any python object with a predict
method which takes as input a DataFrame containing the dataset and outputs a 1-dimensional prediction array of 0s and 1s.
In the scope of precomputed results, we provide:
- full sets of precomputed rules for each dataset (with a frequent itemset minimum support of 1%).
- only for the Adult dataset with 'race' as the protected attribute, we have also provided a file which, in addition to the rules, contains the actual model and the test data we used, which are the main inputs required by our framework. Note: for reading this file correctly, you will probably need to have
scikit-learn
version 1.0.2 installed. - finally, all experiments we ran use 131313 as the value of the random_state parameter, wherever applicable (notice that, for example, the
LogisticRegression
model with default parameters is deterministic).
The above assets can be found in the directory Pre-computed Results. The commands required to read them are displayed in the Jupyter notebooks we provide (more on this in the following sections).
We provide our main experimental results, together with the code that produces them, in the Jupyter notebooks located in the Notebooks directory.
For each case study we have two notebooks. One that shows how our framework can be used for the generation of a large set of candidate counterfactuals, and another which showcases how one can get aggregate statistics of the raw results.
Specifically, for the calculation of the ground set of candidate counterfactuals we have the following notebooks:
and the following notebooks include the aggregate ranking statistics:
The respective results can be seen in our main paper and in the Appendices C and D. More specifically:
- For the Adult dataset with the sex as protected attribute, we show some Comparative Subgroup Counterfactuals in the main part of the paper, in Figures 2 and 3, their ranking statistics in Table 1 of the main paper and some aggregate statistics in Tables 10 and 11 of the Appendix D.
- For the Adult dataset with the race as protected attribute, we show some Comparative Subgroup Counterfactuals in Figure 4 and their ranking statistics in Table 6 of the Appendix C and some aggregate statistics in Tables 12 and 13 of the Appendix D.
- For the COMPAS dataset (with race as the protected attribute), we show some Comparative Subgroup Counterfactuals in Figure 5 and their ranking statistics in Table 7 of the Appendix C and some aggregate statistics in Tables 14 and 15 of the Appendix D.
- For the SSL dataset, we show some Comparative Subgroup Counterfactuals in Figure 6 and their ranking statistics in Table 8 of the Appendix C and some aggregate statistics in Tables 16 and 17 of the Appendix D.
- For the Ad Campaign dataset, we show some Comparative Subgroup Counterfactuals in Figure 7 and their ranking statistics in Table 9 of the Appendix C and some aggregate statistics in Tables 18 and 19 of the Appendix D.
Both notebooks, for each dataset, also contain a series of examples on the generation of Comparative Subgroup Counterfactuals, which is a central representation of the output of our method.