Prevent Discriminatory Data Attributes Used As Model Features

6 / 45 Data This practice was not ranked.
Click to read more.
This practice addresses requirements
from the EU guidelines for trustworthy ML.
Click to read more.


Intent

Avoid building discriminatory practices into machine learning applications.

Motivation

Using personal data attributes, such as gender or ethnicity, as features of machine learning algorithms introduces discriminatory bias, and ultimately leads to models with a negative impact on society.

Applicability

Discriminatory attributes should not be used in applications with a direct or indirect impact on human lives, society or the environment.

Description

When processing personal data – for example the customer data in a banking application – it is important to prevent discriminatory attributes from being used as model features, because the resulting models may base their decisions on these attributes, and ultimately discriminate users.

A widely known example for misusing this practice is the COMPAS case; where discriminatory attributes were used to predict if a perpetrator was likely to recidivate.

An intuitive step for preventing discriminatory attributes from being used as features is to remove them from the training data. However, removing sensitive attributes (or not including them in the first place) is not a cure for fair machine learning, and can even exacerbate fairness issues, if used improperly.

Always be aware that there may be latent sensitive attributes. For instance, in some cases a combination of features that are not considered discriminatory by themselves can be used by a machine learning algorithm to reconstruct a discriminatory attribute. Ultimately, this would have the same effect as using discriminatory attributes directly.

In order to prevent the use of discriminatory attributes, a hybrid approach is needed; consisting of removing the attributes from the training data, testing for latent factors that may uncover them, and continuously testing for other biases such as social or subgroup bias.

Related

Read more



6 / 45 Data This practice was not ranked.
Click to read more.
This practice addresses requirements
from the EU guidelines for trustworthy ML.
Click to read more.