Trustworthy machine learning (ML) encompasses multiple fields of research, including (but not limited to) robustness, algorithmic fairness, interpretability and privacy. Recently, relationships between techniques and metrics used across different fields of trustworthy ML have emerged, leading to interesting work at the intersection of algorithmic fairness, robustness, and causality.
On one hand, causality has been proposed as a powerful tool to address the limitations of initial statistical definitions of fairness. However, questions have emerged regarding the applicability of such approaches in practice and the suitability of a causal framing for studies of bias and discrimination. On the other hand, the Robustness literature has surfaced promising approaches to improve fairness in ML models. For instance, parallels can be shown between individual fairness and local robustness guarantees. In addition, the interactions between fairness and robustness can help us understand how fairness guarantees hold under distribution shift or adversarial/poisoning attacks.
After a first edition of this workshop that focused on causality and interpretability, we will turn to the intersectionality between algorithmic fairness and recent techniques in causality and robustness. In this context, we will investigate how these different topics relate, but also how they can augment each other to provide better or more suited definitions and mitigation strategies for algorithmic fairness. We are particularly interested in addressing open questions in the field, such as:
- How can causally grounded fairness methods help develop more robust and fair algorithms in practice?
- What is an appropriate causal framing in studies of discrimination?
- How do approaches for adversarial/poisoning attacks target algorithmic fairness?
- How do fairness guarantees hold under distribution shift?
Mon 1:00 a.m. - 1:10 a.m.
|
Opening remarks
(
opening remarks by organizers
)
SlidesLive Video » |
Awa Dieng 🔗 |
Mon 1:10 a.m. - 1:12 a.m.
|
Speaker Intro
(
live intro
)
|
🔗 |
Mon 1:12 a.m. - 1:34 a.m.
|
Invited Talk: Generalizability, robustness and fairness in machine learning risk prediction models
(
Invited Talk
)
SlidesLive Video » By leveraging principles of health equity, I will discuss the use of causal models and machine learning to address realistic challenges of data collection and model use across environments. Examples include a domain adaptation approach that improves prediction in under-represented population sub-groups by leveraging invariant information across groups when possible, and an algorithmic fairness method which specifically incorporates structural factors to better account for and address sources of bias and disparities. |
Rumi Chunara 🔗 |
Mon 1:34 a.m. - 1:45 a.m.
|
Q&A for Rumi Chunara
(
live questions
)
|
🔗 |
Mon 1:45 a.m. - 1:50 a.m.
|
Short break
(
short break
)
|
🔗 |
Mon 1:50 a.m. - 1:52 a.m.
|
Speaker Intro
(
live intro
)
|
🔗 |
Mon 1:52 a.m. - 2:20 a.m.
|
Invited Talk: Path-specific effects and ML fairness
(
live talk
)
SlidesLive Video » Arguably, the consideration of causal effects along subsets of causal paths is required for understanding and addressing unfairness in most real-world scenarios. In this talk I will share some of my thoughts on this goal and on the challenges to achieve it. |
Silvia Chiappa 🔗 |
Mon 2:20 a.m. - 2:30 a.m.
|
Q&A for Silvia Chiappa
(
live questions
)
|
🔗 |
Mon 2:30 a.m. - 2:40 a.m.
|
Short break
(
short break
)
|
🔗 |
Mon 2:40 a.m. - 2:42 a.m.
|
Speaker Intro
(
live intro
)
|
🔗 |
Mon 2:42 a.m. - 3:10 a.m.
|
Invited Talk: Causality and fairness in ML: promises, challenges & open questions
(
live talk
)
SlidesLive Video » In the recent years, we have observed an explosion of research approaches at the intersection of causality and fairness in machine learning (ML). These works are often motivated by the promise that causality allows us to reason about the causes of unfairness both in the data and in the ML algorithm. However, the promises of existing causal fair approaches require strong assumptions, which hinder their practical application. In this talk, I will provide a quick overview of both the promises and the technical challenges of causal fair ML frameworks from a theoretical perspective. Finally, I will show how to leverage probabilistic ML to partially relax causal assumptions in order to develop more practical solutions to causal fair ML. |
Isabel Valera 🔗 |
Mon 3:10 a.m. - 3:18 a.m.
|
Q&A for Isabel Valera
(
live questions
)
|
🔗 |
Mon 3:18 a.m. - 3:20 a.m.
|
Intro to Contributed Talks
(
live intro
)
|
🔗 |
Mon 3:20 a.m. - 3:30 a.m.
|
Algorithmic Bias and Data Bias: Understanding the Relation between Distributionally Robust Optimization and Data Curation
(
Oral
)
SlidesLive Video » Machine learning systems based on minimizing average error have been shown to perform inconsistently across notable subsets of the data, which is not exposed by a low average error for the entire dataset. In consequential social and economic applications, where data represent people, this can lead to discrimination of underrepresented gender and ethnic groups. Distributionally Robust Optimization (DRO) seemingly addresses this problem by minimizing the worst expected risk across subpopulations. We establish theoretical results that clarify the relation between DRO and the optimization of the same loss averaged on an adequately weighted training dataset. A practical implication of our results is that neither DRO nor curating the training set should be construed as a complete solution for bias mitigation. |
Agnieszka Słowik · Leon Bottou 🔗 |
Mon 3:30 a.m. - 3:40 a.m.
|
On the Impossibility of Fairness-Aware Learning from Corrupted Data
(
Oral
)
SlidesLive Video » Addressing fairness concerns about machine learning models is a crucial step towards their long-term adoption in real-world automated systems. Many approaches for training fair models from data have been developed and an implicit assumption about such algorithms is that they are able to recover a fair model, despite potential historical biases in the data. In this work we show a number of impossibility results that indicate that there is no learning algorithm that can recover a fair model when a proportion of the dataset is subject to arbitrary manipulations. Specifically, we prove that there are situations in which an adversary can force any learner to return a biased classifier, with or without degrading accuracy, and that the strength of this bias increases for learning problems with underrepresented protected groups in the data. Our results emphasize on the importance of studying further data corruption models of various strength and of establishing stricter data collection practices for fairness-aware learning. |
Nikola Konstantinov · Christoph Lampert 🔗 |
Mon 3:40 a.m. - 3:50 a.m.
|
Achieving Counterfactual Fairness for Causal Bandit
(
Oral
)
SlidesLive Video » In online recommendation, customers arrive in a sequential and stochastic manner from an underlying distribution and the online decision model recommends a chosen item for each arriving individual based on some strategy. We study how to recommend an item at each step to maximize the expected reward while achieving user-side fairness for customers, i.e., customers who share similar profiles will receive a similar reward regardless of their sensitive attributes and items being recommended. By incorporating causal inference into bandits and adopting soft intervention to model the arm selection strategy, we first propose the d-separation based UCB algorithm (D-UCB) to explore the utilization of the d-separation set in reducing the amount of exploration needed to achieve low cumulative regret. Based on that, we then propose the fair causal bandit (F-UCB) for achieving the counterfactual individual fairness. Both theoretical analysis and empirical evaluation demonstrate effectiveness of our algorithms. |
Wen Huang · Lu Zhang · Xintao Wu 🔗 |
Mon 3:50 a.m. - 4:00 a.m.
|
Q&A for Contributed talks 1,2,3
(
live questions
)
SlidesLive Video » |
🔗 |
Mon 4:00 a.m. - 5:00 a.m.
|
Poster session 1 ( Poster session: join us on gathertown ) link » | 🔗 |
Mon 5:00 a.m. - 5:05 a.m.
|
Intro to Roundtables
(
live intro
)
|
🔗 |
Mon 5:05 a.m. - 6:00 a.m.
|
Roundtables ( Roundtable discussions: join us on gathertown ) link » | 🔗 |
Mon 6:00 a.m. - 6:10 a.m.
|
Short break ( Short break: join us on gather.town ) link » | 🔗 |
Mon 6:10 a.m. - 6:13 a.m.
|
Fairness for Robust Learning to Rank
(
Poster
)
SlidesLive Video » While conventional ranking systems focus solely on maximizing the utility of the ranked items to users, fairness-aware ranking systems additionally try to balance the exposure for different protected attributes such as gender or race. To achieve this type of group fairness for ranking, we derive a new ranking system based on the first principles of distributional robustness. We formulate a minimax game between a player choosing a distribution over rankings to maximize utility while satisfying fairness constraints against an adversary seeking to minimize utility while matching statistics of the training data. We show that our approach provides better utility for highly fair rankings than existing baseline methods. |
Omid Memarrast · Ashkan Rezaei · Rizal Fathony · Brian Ziebart 🔗 |
Mon 6:13 a.m. - 6:16 a.m.
|
Cooperative Multi-Agent Fairness and Equivariant Policies
(
Poster
)
SlidesLive Video » We study fairness through the lens of cooperative multi-agent learning. Our work is motivated by empirical evidence that naive maximization of team reward yields unfair outcomes for individual team members. To address fairness in multi-agent contexts, we introduce team fairness, a group-based fairness measure for multi-agent learning. We then prove that it is possible to enforce team fairness during policy optimization by transforming the team's joint policy into an equivariant map. We refer to our multi-agent learning strategy as Fairness through Equivariance (Fair-E) and demonstrate its effectiveness empirically. We then introduce Fairness through Equivariance Regularization (Fair-ER) as a soft-constraint version of Fair-E and show that it reaches higher levels of utility than Fair-E and fairer outcomes than non-equivariant policies. Finally, we present novel findings regarding the fairness-utility trade-off in multi-agent settings; showing that the magnitude of the trade-off is dependent on agent skill level. |
Niko Grupen · Bart Selman · Daniel Lee 🔗 |
Mon 6:16 a.m. - 6:19 a.m.
|
Fair SA: Sensitivity Analysis for Fairness in Face Recognition
(
Poster
)
SlidesLive Video » As the use of deep learning in high impact domains becomes ubiquitous, it is increasingly important to assess the resilience of models. One such high impact domain is that of face recognition, with real world applications involving images affected by various degradations, such as motion blur or high exposure. Moreover, images captured across different attributes, such as gender and race, can also challenge the robustness of a face recognition algorithm. While summary statistics suggest that the aggregate performance of face recognition models has continued to improve, these metrics do not directly measure the robustness or fairness of the models. Visual Psychophysics Sensitivity Analysis (VPSA) [1] provides a way to pinpoint the individual causes of failure by way of introducing incremental perturbations in the data. However, perturbations may affect subgroups differently. In this paper, we propose a new fairness evaluation based on robustness in the form of a generic framework that extends VPSA. With this framework, we can analyze the ability of a model to perform fairly for different subgroups of a population affected by perturbations, and pinpoint the exact failure modes for a subgroup by measuring targeted robustness. With the increasing focus on the fairness of face recognition algorithms, we use face recognition as an example application of our framework and propose to represent the fairness of a model via AUC matrices. We analyze the performance of common face recognition models and empirically show that certain subgroups may be at a disadvantage when images are perturbed. |
Aparna Joshi · Xavier Suau Cuadros · Nivedha Sivakumar · Luca Zappella · Nicholas Apostoloff 🔗 |
Mon 6:19 a.m. - 6:22 a.m.
|
Measure Twice, Cut Once: Quantifying Bias and Fairness in Deep Networks
(
Poster
)
SlidesLive Video » AI systems come with serious concerns of bias and fairness. Algorithmic bias is more abstract and unintuitive than traditional forms of discrimination and can be more difficult to detect and mitigate. A clear gap exists in the current literature on evaluating the relative bias in the performance of multi-class classifiers. In this work, we propose two simple yet effective metrics, Combined Error Variance (CEV) and Symmetric Distance Error (SDE), to quantitatively evaluate the class-wise bias of two models in comparison to one another. We evaluate the performance of these new metrics by demonstrating practical use cases with pre-trained models and show that they can be used to measure fairness as well as bias. |
Ziliang Zong · Cody Blakeney · Gentry Atkinson · Nathaniel Huish · · Vangelis Metsis 🔗 |
Mon 6:22 a.m. - 6:25 a.m.
|
Bounded Fairness Transferability subject to Distribution Shift
(
Poster
)
SlidesLive Video » We study the \emph{transferability of fair predictors} (i.e., classifiers or regressors) assuming domain adaptation. Given a predictor that is “fair” on some \emph{source} distribution (of features and labels), is it still fair on a \emph{realized} distribution that differs? We first generalize common notions of static, statistical group-level fairness to a family of premetric functions that measure “induced disparity.” We quantify domain adaptation by bounding group-specific statistical divergences between the source and realized distributions. Next, we explore cases of simplifying assumptions for which bounds on domain adaptation imply bounds on changes to induced disparity. We provide worked examples for two commonly used fairness definitions (i.e., demographic parity and equalized odds) and models of domain adaptation (i.e., covariate shift and label shift) that prove to be special cases of our general method. Finally, we validate our theoretical results with synthetic data. |
Reilly Raab · Yatong Chen · Yang Liu 🔗 |
Mon 6:25 a.m. - 6:28 a.m.
|
Counterfactual Fairness in Mortgage Lending via Matching and Randomization
(
Poster
)
SlidesLive Video » Unfairness in mortgage lending has created generational inequality among racial and ethnic groups in the US. Many studies address this problem, but most existing work focuses on correlation-based techniques. In our work, we use the framework of counterfactual fairness to train fair machine learning models. We propose a new causal graph for the variables available in the Home Mortgage Disclosure Act (HMDA) data. We use a matching-based approach instead of the latent variable modeling approach, because the former approach does not rely on any modeling assumptions. Furthermore, matching provides us with counterfactual pairs in which the race variable is isolated. We first demonstrate the unfairness in mortgage approval and interest rates between African-American and non-Hispanic White sub-populations. Then, we show that having balanced data using matching does not guarantee perfect counterfactual fairness of the machine learning models. |
Sama Ghoba · Nathan Colaner 🔗 |
Mon 6:28 a.m. - 6:31 a.m.
|
Structural Interventions on Automated Decision Making Systems
(
Poster
)
SlidesLive Video » To address discrimination and inequality in automated decision making systems it is standard practice to implement so-called ``fairness" metrics during algorithm design. These measures, although useful to enforce and diagnose fairness at the decision stage, are not sufficient to capture forms of discrimination arising throughout and from structural properties of the system as a whole. To complement the standard approach, we propose a systemic analysis, aided by structural causal models, through which social interventions can be compared to algorithmic interventions. This framework allows us to identify bias outside the algorithmic stage, and propose joint interventions on social dynamics and algorithm design. We show how, for a model of financial lending, structural interventions can drive the system towards equality even when algorithmic interventions are not able to do so. This means the responsibility of decision makers does not stop when local fairness metrics are satisfied, they must ensure a whole ecosystem that fosters equity for all. |
efren cruz · Sarah Rajtmajer · Debashis Ghosh 🔗 |
Mon 6:31 a.m. - 6:34 a.m.
|
Balancing Robustness and Fairness via Partial Invariance
(
Poster
)
SlidesLive Video » The Invariant Risk Minimization (IRM) framework aims to learn invariant features for out-of-distribution generalization with the assumption that the underlying causal mechanisms remain constant. In other words, environments should sufficiently |
Moulik Choraria · Ibtihal Ferwana · Ankur Mani · Lav Varshney 🔗 |
Mon 6:34 a.m. - 6:37 a.m.
|
Implications of Modeled Beliefs for Algorithmic Fairness in Machine Learning
(
Poster
)
SlidesLive Video » Ethics and societal implications of automated decision making have become a major theme in Machine Learning research. Conclusions from theoretical studies in this area are often stated in general terms (such as affirmative action possibly hurting all groups or fairness measures being incompatible with a decision maker being rational). Our work aims to highlight the degree to which such conclusions are in fact relying on modeled beliefs as well as on the technicalities of a chosen framework of analysis (eg. statistical learning theory, game theory, dynamics etc). We carefully discuss prior work through this lens and then highlight the effect of modeled beliefs by means of a simple statistical model where an observed score X is the result of two unobserved hidden variables ("talent" T and "environment" E). We assume that variable T is identically distributed for two subgroups of a population while E models the disparities between an advantaged and a disadvantaged group. We analyze (Bayes-)optimal decision making under a variety of distributional assumptions and show that even the simple model under consideration exhibits some counterintuitive effects. |
Ruth Urner · Jeff Edmonds · Karan Singh 🔗 |
Mon 6:37 a.m. - 6:40 a.m.
|
Fairness Degrading Adversarial Attacks Against Clustering Algorithms
(
Poster
)
SlidesLive Video » Clustering algorithms are ubiquitous in modern data science pipelines, and are utilized in numerous fields ranging from biology to facility location. Due to their widespread use, especially in societal resource allocation problems, recent research has aimed at making clustering algorithms fair, with great success. Furthermore, it has also been shown that clustering algorithms, much like other machine learning algorithms, are susceptible to adversarial attacks where a malicious entity seeks to subvert the performance of the learning algorithm. However, despite these known vulnerabilities, there has been no research undertaken that investigates fairness degrading adversarial attacks for clustering. We seek to bridge this gap by formulating a generalized attack optimization problem aimed at worsening the group-level fairness of centroid-based clustering algorithms. As a first step, we propose a fairness degrading attack algorithm for k-median clustering that operates under a whitebox threat model-- where the clustering algorithm, fairness notion, and the input dataset are known to the adversary. We provide empirical results as well as theoretical analysis for our simple attack algorithm, and find that the addition of the generated adversarial samples can lead to significantly lower fairness values. In this manner, we aim to motivate fairness degrading adversarial attacks as a direction for future research in fair clustering. |
Anshuman Chhabra · Adish Singla · Prasant Mohapatra 🔗 |
Mon 6:40 a.m. - 7:55 a.m.
|
Long break ( Long break: join us on gather.town ) link » | 🔗 |
Mon 7:55 a.m. - 8:00 a.m.
|
Intro to Afternoon session
(
live intro
)
|
🔗 |
Mon 8:00 a.m. - 8:02 a.m.
|
Speaker Intro
(
live intro
)
|
🔗 |
Mon 8:02 a.m. - 8:30 a.m.
|
Invited Talk: Causality and Fairness
(
live talk
)
SlidesLive Video » |
Elias Bareinboim 🔗 |
Mon 8:30 a.m. - 8:40 a.m.
|
Q&A for Elias Bareinboim
(
live questions
)
|
🔗 |
Mon 8:40 a.m. - 8:42 a.m.
|
Speaker Intro
(
live intro
)
|
🔗 |
Mon 8:42 a.m. - 9:20 a.m.
|
Invited Talk: Towards Reliable and Robust Model Explanations
(
Invited talk
)
SlidesLive Video » As machine learning black boxes are increasingly being deployed in domains such as healthcare and criminal justice, there is growing emphasis on building tools and techniques for explaining these black boxes in an interpretable manner. Such explanations are being leveraged by domain experts to diagnose systematic errors and underlying biases of black boxes. In this talk, I will present some of our recent research that sheds light on the vulnerabilities of popular post hoc explanation techniques such as LIME and SHAP, and also introduce novel methods to address some of these vulnerabilities. More specifically, I will first demonstrate that these methods are brittle, unstable, and are vulnerable to a variety of adversarial attacks. Then, I will discuss two solutions to address some of the aforementioned vulnerabilities–(i) a Bayesian framework that captures the uncertainty associated with post hoc explanations and in turn allows us to generate explanations with user specified levels of confidence, and (ii) a framework based on adversarial training that is designed to make post hoc explanationsmore stable and robust to shifts in the underlying data; I will conclude the talk by discussing our recent theoretical results which shed light on the equivalence and robustness of state-of-the-art explanation methods. |
Himabindu Lakkaraju 🔗 |
Mon 9:20 a.m. - 9:28 a.m.
|
Q&A for Hima Lakkaraju
(
live questions
)
|
🔗 |
Mon 9:28 a.m. - 9:30 a.m.
|
Intro to Contributed Talks
(
live intro
)
|
🔗 |
Mon 9:30 a.m. - 9:40 a.m.
|
The Many Roles that Causal Reasoning Plays in Reasoning about Fairness in Machine Learning
(
Oral
)
SlidesLive Video » We survey the many roles that causal reasoning plays in reasoning about fairness in machine learning. While the existing scholarship on causal approaches to fairness in machine learning has focused on the degree to which features in a model might have been causally affected by (discrimination on the basis of) sensitive features, causal reasoning also plays an important---if more implicit---role in other ways of assessing the fairness of models. This paper therefore tries to distinguish and disentangle the many roles that causal reasoning plays in reasoning about fairness, with the additional goal of asking how causality is thought to help achieve these normative goals and to what extent this is possible or necessary. |
Irene Y Chen · Hal Daumé III · Solon Barocas 🔗 |
Mon 9:40 a.m. - 9:50 a.m.
|
Detecting Bias in the Presence of Spatial Autocorrelation
(
Oral
)
SlidesLive Video » In spite of considerable practical importance, current algorithmic fairness literature lacks technical methods to account for underlying geographic dependency while evaluating or mitigating bias issues for spatial data. We initiate the study of bias in spatial applications in this paper, taking the first step towards formalizing this line of quantitative methods. Bias in spatial data applications often gets confounded by underlying spatial autocorrelation. We propose hypothesis testing methodology to detect the presence and strength of this effect, then account for it by using a spatial filtering-based approach---in order to enable application of existing bias detection metrics. We evaluate our proposed methodology through numerical experiments on real and synthetic datasets, demonstrating that in the presence of several types of confounding effects due to the underlying spatial structure our testing methods perform well in maintaining low type-II errors and nominal type-I errors. |
Subhabrata Majumdar · Cheryl Flynn · Ritwik Mitra 🔗 |
Mon 9:50 a.m. - 10:00 a.m.
|
Fair Clustering Using Antidote Data
(
Oral
)
SlidesLive Video » Clustering algorithms are widely utilized for many modern data science applications. This motivates the need to make outputs of clustering algorithms fair. Traditionally, new fair algorithmic variants to clustering algorithms are developed for specific notions of fairness. However, depending on the application context, different definitions of fairness might need to be employed. As a result, new algorithms and analysis need to be proposed for each combination of clustering algorithm and fairness definition. Additionally, each new algorithm would need to be reimplemented for deployment in a real-world system. Hence, we propose an alternate approach to group-level fairness in center-based clustering inspired by research on data poisoning attacks. We seek to augment the original dataset with a small number of data points, called antidote data. When clustering is undertaken on this new dataset, the output is fair, for the chosen clustering algorithm and fairness definition. We formulate this as a general bi-level optimization problem which can accommodate any center-based clustering algorithms and fairness notions. We then categorize approaches for solving this bi-level optimization for two different problem settings. Extensive experiments on different clustering algorithms and fairness notions show that our algorithms can achieve desired levels of fairness on many real-world datasets with a very small percentage of antidote data added. We also find that our algorithms achieve lower fairness costs and competitive clustering performance compared to other state-of-the-art fair clustering algorithms. |
Anshuman Chhabra · Adish Singla · Prasant Mohapatra 🔗 |
Mon 10:00 a.m. - 10:10 a.m.
|
Q&A for Contributed talks 4,5,6
(
live questions
)
SlidesLive Video » |
🔗 |
Mon 10:10 a.m. - 10:15 a.m.
|
Short Break
(
short break
)
|
🔗 |
Mon 10:15 a.m. - 10:17 a.m.
|
Speaker Intro
(
live intro
)
|
🔗 |
Mon 10:17 a.m. - 10:47 a.m.
|
Invited Talk: Lessons from robust machine learning
(
live talk
)
SlidesLive Video » Current machine learning (ML) methods are primarily centered around improving in-distribution generalization where models are evaluated on new points drawn from nearly the same distribution as the training data. On the other hand, robustness and fairness involve reasoning about out-of-distribution performance such as accuracy on protected groups or perturbed inputs, and reliability even in the presence of spurious correlations. In this talk, I will describe an important lesson from robustness: in order to improve out-of-distribution performance, we often need to question the common assumptions in ML. In particular, we will see that ‘more data’, ‘bigger models’, or ‘fine-tuning pretrained features’ which improve in-distribution generalization often fail out-of-distribution. |
Aditi Raghunathan 🔗 |
Mon 10:47 a.m. - 10:55 a.m.
|
Q&A for Aditi Raghunathan
(
live questions
)
|
🔗 |
Mon 10:55 a.m. - 11:00 a.m.
|
Short break
(
short break
)
|
🔗 |
Mon 11:00 a.m. - 11:40 a.m.
|
Panel: Been Kim (Google Brain), Solon Barocas (Microsoft Research), Ricardo Silva (UCL), Rich Zemel (U. of Toronto)
(
Live Discussion
)
SlidesLive Video » |
🔗 |
Mon 11:40 a.m. - 12:20 p.m.
|
Poster session 2 ( Poster session: join us on gathertown ) link » | 🔗 |
Mon 12:20 p.m. - 12:30 p.m.
|
Closing remarks
(
Closing remarks by organizers
)
SlidesLive Video » |
Miriam Rateike 🔗 |