Fairness in Machine Learning

Just some notes

(Work in progress)

Varieties of Fairness

One of the hardest problems in fairness is that there is no consensus on the definition of fairness or what does it mean to be fair. Depending on the context or culture, the definition of fairness can be different .

Researchers and designers at Google’s PAIR (People and AI Research) , initiative created the What-If visualization tool as a pragmatic resource for developers of machine learning systems. The tool provides a set of fairness metrics that can be used to evaluate the fairness of a model. The metrics are grouped into five categories:

It can be seen that, these proposed metrics are already complex and hard to understand. For example, in my opinion, the “group unware” and “equal opportunity” are quite similar to each other where both of them aim to ensure that the model does not discriminate based on “protected characteristics” like gender, age, race, etc. Overall, these metrics can be grouped into two categories: group fairness and individual fairness which are also the two main categories of fairness in machine learning.

Learning Fair Representations

One of the milestone work in fairness is the paper “Learning Fair Representations” by Zemel et al. (2013) . The authors proposed a method to learn fair representations by learning a latent representation that encodes the data well but obfuscates information about protected attributes. The method is based on the intuition that if the learned representation does not contain any information about the protected attribute, then any classifier based on these representation cannot use the protected attribute to make predictions.

The authors formulated this using the notion of statistical parity, which requires that the probability that a random element from \(X^+\) maps to a particular prototype is equal to the probability that a random element from \(X^-\) maps to the same prototype

\[P(Z = k \mid x^+ \in X^+) = P(Z = k \mid x^- \in X^-) \; \forall k\]

Where \(X^+\) and \(X^-\) are the sets of protected and unprotected examples, respectively, and \(Z\) is the latent representation with \(K\) prototypes.

Fairness in Deep Learning?

Fairness in Generative Models

Fairness Machine Learning is mostly considered in the context of decision making models such as classification models. However, fairness is also an important issue in generative models, which is not well studied yet. Recently, the central problem of fairness in generative models is how to ensure diversity in the generated outputs. For example, a response to a question about famous musicians should not only include names or images of people of the same gender identity or skin tone , .

Some of the following attributes will be highly considered when talking about fairness in generative models:

When evaluating the fairness of generative models, the authors of suggest to consider the following metrics:

In summary, we can think about two scenarios when evaluating the fairness of generative models: same input - diverse output and diverse input - same output.