Jill-Jênn Vie

Researcher at Inria

% FARE: Provably Fair Representation Learning with Practical Certificates % Jill-Jênn Vie % MILLE CILS 2023 — aspectratio: 169 institute: \includegraphics[height=1cm]{figures/inria.png} \includegraphics[height=1cm]{figures/soda.png} header-includes:

Context

Learning fair representations such that \alert{any} classifier using these representations cannot discriminate even if they are trying to.

We need \alert{practical} certificates

Requirements

  1. High-probability: Bound the fairness metric with high probability
  2. Finite sample bound and not asymptotic i.e. $n \to \infty$
  3. Distribution-free bound
  4. Model-free: Should hold for any model trained on representations
  5. Non-vacuous bound (i.e. $< 1$)

(Bounds should be explicitly computable on real datasets)

Strong points

Background

Metrics

\pause

\begin{definition} Given small $\varepsilon$, finite dataset $D$, encoder $f : x \mapsto z$ creating representations, a \alert{practical DP distance certificate} is a value $T^*(n, D) \in \R$ such that \(\sup_{g \in \mathcal{G}} \Delta(g) \leq T^*(n, D)\) holds with probability $1 - \varepsilon$. \end{definition}

Trick

Representations should have finite support
i.e. $f : \R^d \to {z_1, \ldots, z_k}$ one of $k$ possible values (whaaat?)

\pause

But actually this includes decision trees (each leaf has same encoding)

Examples of criteria

Results

\centering

Personal thoughts

Nikola Jovanović, Mislav Balunovic, Dimitar Iliev Dimitrov, Martin Vechev. \alert{FARE: Provably Fair Representation Learning with Practical Certificates.} Proceedings of the 40th International Conference on Machine Learning, PMLR 202:15401-15420, 2023. \url{https://arxiv.org/abs/2210.07213}

Thanks! jill-jenn.vie@inria.fr