In this tutorial you will get to know when, why and how to use
fairmodels is a tool for bias detection, visualization and mitigation. It is compatible with
DALEX (@dalex) and
DALEXtra which are model agnostic explainers. Some knowledge of how to use those explainers will be needed but in this tutorial you should grasp the idea.
For this tutorial we will use
compas data (@propublica) to see if someone will become recidivist in next 2 years.
knitr::opts_chunk$set( collapse = FALSE, comment = "#>", warning = FALSE, message = FALSE, cache = FALSE, fig.align = 'center', dpi = 72, fig.width = 10, fig.height = 7 )
Let's say you are building algorithm for court system that predicts if someone will become recidivist in the future. First you gather information then you build a model and predict outcomes. You get satisfying accuracy. It is pretty good, but it appears that the model is more likely to say that African Americans will become recidivists. It was trained on data that where discrimination of certain ethnic groups was visible. Racist algorithms are no good for the court system. So now we have some options. First one is to gather more reliable data, and the second one is trying to understand bias and choose model with least amount of it. We will choose the second option.
We will use modified ProPublica's compas data to represent our problem.
library(fairmodels) data("compas") head(compas)
fairmodels package to work properly we want to flip factor levels in target variable, so positive outcome (not being a recidivist) is being predicted by models. It is only needed for one specific function but more on it later.
compas$Two_yr_Recidivism <- as.factor(ifelse(compas$Two_yr_Recidivism == '1', '0', '1'))
We train a
ranger (@ranger) model and create an explainer with
library(DALEX) library(ranger) # train rf_compas <- ranger(Two_yr_Recidivism ~., data = compas, probability = TRUE) # numeric target values y_numeric <- as.numeric(compas$Two_yr_Recidivism)-1 # explainer rf_explainer <- explain(rf_compas, data = compas[,-1], y = y_numeric, colorize = FALSE)
Than we create call function
This function aggregates many explainers so you may compare many models.
We assign object to name
fobject which is short version of
fairness_object - object returned by
fobject <- fairness_check(rf_explainer, # explainer protected = compas$Ethnicity, # protected variable as factor privileged = "Caucasian", # level in protected variable, potentially more privileged cutoff = 0.5, # cutoff - optional, default = 0.5 colorize = FALSE)
Let's see if our ranger model has bias.
print(fobject, colorize = FALSE)
Of course the
protected parameter can be changed, for example to (here binary)
In many metrics
ranger exceeds fairness threshold (which can be changed by
epsilon parameter). If bars reach red field on the left it means that there is bias towards certain unprivileged subgroup. If they reach one on the right it means bias towards privileged (Caucasian - in all metrics here Caucasian subgroup is referenced as base - 1) subgroup. This values mean the proportion (ratio) of certain unprivileged subgroup to privileged subgroup (eg. Asian/Caucasian). By default it represents 80% rule (for eg. women should get credit at rate at least 80% of that of men) (@adverseImpact) . Metrics measure how equal treatment and mistreatment among subgroups is. More on those metrics: wikipedia and in @rubin
Why do we have this bias? Model did learn from biased data. We can see it on plot below
As we can see it is more likely that model will categorize African Americans as being recidivists than for example Asians.
Sometimes it is worth it to see what the raw (unscaled) metrics are. User could use
fobject$groups_data but the easier way is to plot
To really see what
fairness_object is about, we need to make some more models and explainers.
library(gbm) rf_compas_1 <- ranger(Two_yr_Recidivism ~Number_of_Priors+Age_Below_TwentyFive, data = compas, probability = TRUE) lr_compas_1 <- glm(Two_yr_Recidivism~., data=compas, family=binomial(link="logit")) rf_compas_2 <- ranger(Two_yr_Recidivism ~., data = compas, probability = TRUE) rf_compas_3 <- ranger(Two_yr_Recidivism ~ Age_Above_FourtyFive+Misdemeanor, data = compas, probability = TRUE) df <- compas df$Two_yr_Recidivism <- as.numeric(compas$Two_yr_Recidivism)-1 gbm_compas_1<- gbm(Two_yr_Recidivism~., data = df) explainer_1 <- explain(rf_compas_1, data = compas[,-1], y = y_numeric) explainer_2 <- explain(lr_compas_1, data = compas[,-1], y = y_numeric) explainer_3 <- explain(rf_compas_2, data = compas[,-1], y = y_numeric, label = "ranger_2") explainer_4 <- explain(rf_compas_3, data = compas[,-1], y = y_numeric, label = "ranger_3") explainer_5 <- explain(gbm_compas_1, data = compas[,-1], y = y_numeric)
Now we create one object with all explainers
fobject <- fairness_check(explainer_1, explainer_2, explainer_3, explainer_4, explainer_5, protected = compas$Ethnicity, privileged = "Caucasian", verbose = FALSE)
As we can see there is some parameters in fairness_check such as:
1. x, ... - list of
DALEX explainers, and other
2. protected - factor, containing subgroups as levels. Protected stands for protected variable (or sensitive attribute)
3. privileged - character, level in protected, it is subgroup suspected of having better results
4. cutoff - numeric, vector of cutoffs values matching the order of levels in protected variable. It affects only explainers so if
fairness_object is passed it's cutoff vector won't be changed.
5. label - character, vector of labels for explainers only. Very convenient for
fairness_check() iterative approach - having explainer, checking for bias, mitigating bias, passing both explainer and fairness object and comparing fairness.
6. epsilon - numeric, boundary position in
fairness_check(). Fairness metrics are satisfied if parity loss values are between (-epsilon, epsilon)
fairness_object is output value of
It is S3 object consisting of:
* Parity loss metrics
Popular confusion matrix metrics with parity loss - sum of absolute logarithms of ratios of unprivileged subgroups and privileged one (more on that in documentation
?fairness_check. If model would have 0 in certain parity loss metric it would mean that it treats all subgroups equally.
Fairness object gets metrics based on confusion matrix and checks them over the groups.
# for the first model fobject$groups_data$ranger$TPR
It is simply metrics for certain subgroup.
What is relation between
If we were going only to take score from certain metric (Let's say fpr and 0.3) we wouldn't know if it is good or bad. But we are aiming for equal treatment over all groups so if this metric score would be the same in all groups it would be very good. But the metrics wouldn't be comparable between each others (fpr - 0.3 in all groups and accuracy - 0.9 in all groups, both are good in terms of parity). That is why we use privileged - to set benchmark. And for example Caucasian in fpr had score of 0.3 and African American 0.6. After setting
privilieged = Caucasian Caucasian would have score 0, and African American
log(0.6/0.3). It enables comparing and adding scores.
Note: When dealing with plots we use formula sum(abs(log(Mu/Mp))) for all unprivileged metrics (Mu - metric unprivileged, Mp - metric privileged) to represent aggregated score in subgroups. In short is how much it differs from ideal scores. The closer to zero the better
group_confusion_matrices Similarly to groups data it stores confusion matrix for each subgroup
# for first model fobject$cutoff$ranger
list of cutoff values for each model
data used in print and plot of
fairness_object. It is already processed data and ready to plot. If someone were to use
abs() metrics there would be equal to particular metrics in
$parity_loss_metric_data. It means that it allows negative values. So when value is negative it means that score of privileged group in this metric was better.
... - other parameters passed to
We now have a few models in our
Let's see how they perform in different metrics.
sm <- stack_metrics(fobject) plot(sm)
It displays accumulated (Stacked) metric scores for each model. The least metric score the better.
cm <- choose_metric(fobject, "TPR") plot(cm)
With this task we should use
PCA. We call
create_fairness_pca() to create fairness pca object.
fair_pca <- fairness_pca(fobject) print(fair_pca)
Another way to deal with grouped data is using heatmap.
fheatmap <- fairness_heatmap(fobject) plot(fheatmap, text_size = 3)
For both models and metrics dendograms are created. This way through hierarchical clustering we can look on similarities between models/metrics. It should give similar but more detailed information than PCA Now we know what those scores are and how "similar" models are to each other
Sometimes we would like to know how good are models in performance metrics and in fairness metrics at the same time, to see the tradeoff between them.
fap <- performance_and_fairness(fobject, fairness_metric = "STP") plot(fap)
When we have narrowed down our search for the best model we can use
group_metric to check once again metrics within groups and decide which model to use.
fobject2 <- fairness_check(explainer_1,explainer_2, protected = compas$Ethnicity, privileged = "Caucasian", verbose = FALSE) gm <- group_metric(fobject2, fairness_metric = "FPR") plot(gm)
fradar <- fairness_radar(fobject2) plot(fradar)
We may see how cutoff affects parity loss of metrics
All cutoffs measures where metrics exist (are not NA) and how they change if we modify cutoffs in all subgroups.
In this plot
NA values are natural, so warnings are to be expected.
ac <- all_cutoffs(fobject2) plot(ac)
This function shows how parity loss metrics would change if we modified cutoff only for one subgroup (here African American) with other cutoffs fixed.
cpc <- ceteris_paribus_cutoff(fobject2, subgroup = "African_American") plot(cpc)
fairmodels is powerful and flexible tool for detecting bias and asserting fairness in models. Many metrics are available (there are tools to create own metrics based on confusion matrix). Besides bias detection
fairmodels offers also bias mitigation. It is set of pre-processing and post-processing algorithms working either on data or on explainer which function is to lower parity loss in certain metrics. If you are interested in this and adult data case study, please check Advanced Tutorial. If you encountered a bug or you have cool idea for a new feature please write and issue here.
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.