Reduced Products of Abstract Domains for Fairness Certification of Neural Networks
We present Libra, an open-source abstract interpretation-based static analyzer for certifying fairness of ReLU neural network classifiers for tabular data. Libra combines a sound forward pre-analysis with an exact backward analysis that leverages the polyhedra abstract domain to provide definite fairness guarantees when possible, and to otherwise quantify and describe the biased input space regions. The analysis is configurable in terms of scalability and precision. We equipped Libra with new abstract domains to use in the pre-analysis, including a generic reduced product domain construction, as well as search heuristics to find the best analysis configuration. We additionally set up the backward analysis to allow further parallelization. Our experimental evaluation demonstrates the effectiveness of the approach on neural networks trained on a popular dataset in the fairness literature.
Wed 7 DecDisplayed time zone: Auckland, Wellington change
13:30 - 15:00 | SAS Papers 1COVID Time Papers In Person at Lecture Theatre 2 Chair(s): Roberto Giacobazzi University of Verona | ||
13:30 30mTalk | Abstract Neural Networks COVID Time Papers In Person Link to publication DOI | ||
14:00 30mTalk | Reduced Products of Abstract Domains for Fairness Certification of Neural Networks COVID Time Papers In Person Denis Mazzucato INRIA & École Normale Supérieure, Caterina Urban Inria & École Normale Supérieure | Université PSL Link to publication DOI | ||
14:30 30mTalk | Static analysis of ReLU neural networks with tropical polyhedra COVID Time Papers In Person Eric Goubault Ecole Polytechnique, Sebastien Palumby Ecole Polytechnique, Sylvie Putot École Polytechnique, Louis Rustenholz Universidad Politécnica de Madrid (UPM) and IMDEA Software Institute, Sriram Sankaranarayanan University of Colorado, Boulder Link to publication DOI |