Robust ML Auditing using Prior Knowledge

Jade Garcia Bourrée*
Inria, Université de Rennes
Augustin Godinot*
Université de Rennes, Inria, IRISA/CNRS, PEReN
Gilles Tredan
LAAS, CNRS

ICML25 arXiv Code

Do you remember Dieselgate? The car computer would detect when it was on a test-bench and reduce the engine power to fake environmental compliance. Well, this can happen with AI regulation too.

An audit is pretty straightforward.

  1. I, the auditor 🕵️ come up with questions to ask your model.
  2. You, the platform 😈 answer my questions.
  3. I look at your answers and decide whether your system abides by the law by computing a series of aggregate metrics.
Interactions between the platform, the auditor and the users
Interactions between the platform, the auditor and the users

Now, you know the metric, you know the questions, and I don't have access to your model. Thus, nothing prevents you from manipulating the answers of your model to pass the audit. And this is very easy! In fact, any fairness mitigation method can be transformed into an audit manipulation attack.

There are two main approaches to avoid manipulations.

In this paper, we formalize the second approach as a search for efficient "audit priors". We instantiate our framework with a simple idea: just look at the accuracy of the platform's answers. Our experiments show that this can help reduce the amount of unfairness a platform could hide.

The amount of unfairness a platform can hide as a function of the auditor query budget
The amount of unfairness a platform can hide as a function of the auditor query budget

If you want to read more about this, I encourage you to read the paper, but not only! Recently, there has been a lot of exciting works on robust audits, here are a few I enjoyed: