fairness through unawareness procedural fairness, or fairness through unawareness is a fairness system If you have no idea about the demographics of protected groups, you will make better decisions. exclude sensitive features from datasets exclude proxies of protected groups Problem: deeply correlated information (such as stuff that people like) is hard to get rid of—individual features does nothing with respect to predicting gender, but taken in groups it can recover protected group information. fairness through awareness we only care about the outcome fairness through parity that the prediction for different groups
fairness through calibration We want the CORRECTNESS of the algorithm to be similar between protected groups. disparate impact \begin{equation} \frac{P(G=G^{}|D=0)}{P(G=G^{}|D=1)} \leq \epsilon \end{equation} where, by US law, disparate impact states \epsilon must be 0.2 or smaller for protected groups D. where G^{*} is the correct prediction.