# Fair Inference On Outcomes

###### Abstract

Many data analysis tasks, such as solving prediction problems or inferring cause effect relationships, can be framed as statistical inference on models with outcome variables. This type of inference has been very successful in a variety of applications, including image and video analysis, speech recognition, machine translation, autonomous vehicle control, game playing, and validating hypotheses in the empirical sciences. As statistical and machine learning models become an increasingly ubiquitous part of our lives, policymakers, regulators, and advocates have expressed fears about the harmful impact of deployment of such models that encode harmful and discriminatory biases of their creators. A growing community is now addressing issues of fairness and transparency in data analysis in part by defining, analyzing, and mitigating harmful effects of algorithmic bias from a variety of perspectives and frameworks [18, 4, 6, 8, 3, 7].

In this paper, we consider the problem of fair statistical inference involving outcome variables. Examples include classification and regression problems, and estimating treatment effects in randomized trials or observational data. The issue of fairness arises in such problems where some covariates or treatments are “sensitive,” in the sense of having potential of creating discrimination. In this paper, we argue that the presence of discrimination in our setting can be formalized in a sensible way as the presence of an effect of a sensitive covariate on the outcome along certain causal pathways, a view which generalizes [16]. We discuss a number of complications that arise in classical statistical inference due to this view, and suggest workarounds, based on recent work in causal and semi-parametric inference.

vskip=0pt

## 1 Introduction

Modern technology, such as the ever present cell phones and cameras, the internet and social media has resulted in an explosion of data, and as a result, the importance and complexity of data analysis tasks has grown. Many of these tasks, such as event prediction, object recognition and tracking, and hypothesis validation can be framed as statistical inference involving an outcome variable of interest. Unfortunately, data analysis is based on statistical models that do not, by default, encode human intuitions about fairness and bias. For instance, it is well-known that recidivism is predicted at higher rates among certain minorities in the United States. To what extent are these predictions discriminatory? What is a sensible framework for thinking about these issues? A variety of approaches have been suggested in the literature [18, 4, 6, 8, 3, 7].

In this paper, we propose to model discrimination based on a “sensitive feature,” such as race or gender, with respect to an outcome as the presence of an effect of the feature on the outcome along certain “disallowed” causal pathways. As a simple example, discussed in [16], job applicants’ gender should not *directly* influence the hiring decision, but may influence the hiring decision indirectly, via secondary applicant characteristics important for the job, and correlated with gender. We argue that this view captures a number of intuitive properties of discrimination, and generalizes existing formal [16] and informal proposals.

We discuss a number of difficulties this view poses for classical statistical inference. In particular, because discrimination is represented by a causal quantity defined on counterfactual random variables, model selection, parameter identification, and assessing out of sample performance become more difficult. In addition, sensitive variables like race and gender cannot in general be conceptualized as treatments in a hypothetical randomized controlled trial, which makes formulating causal targets of inference important for assessing discrimination, such as direct and indirect effects, challenging [30]. We propose workarounds for these difficulties based on causal and semi-parametric methodology [23, 28, 10, 30]. Finally, we illustrate how fair inference on outcomes, which can be viewed as maximizing a constrained likelihood or calculating a constrained posterior distribution, can be performed in practice on a set of simulated datasets.

The paper is organized as follows. We fix our notation in Section 2. In Sections 3 and 4, we give a brief introduction to causal inference and mediation analysis. In Section 5, we review the argument that discrimination can be formalized as the presence of a path-specific effect. In Section 6, we define fair inference based on our definition as a certain type of constrained optimization problem. In Section 7, we discuss a number of extensions to the basic framework of fair inference. In Section 8, we illustrate our framework via a simulation study. Additional discussion and our conclusions are found in Section 9.

## 2 Notation And Preliminaries

Variables will be denoted by uppercase Roman letters, , and sets of variables as uppercase bold Roman letters, . Values will be lowercase Roman letters, , and sets of values by lowercase bold Roman letters, . A state space of a variable will be denoted by , so if is a value of , . A state space of a set of variables is the usual Cartesian state space: . We will represent datasets by by matrices , where rows are realizations, and columns are features , and an outcome . We denote by and the th realization of the th feature and the outcome in the dataset. Similarly, is the th realization of the entire feature vector. Given , and , denote by the restriction of to values in .

We will consider three types of outcome based inference tasks, classification, regression, and establishing causal effects.
The first two tasks can be framed as using an existing dataset drawn from to learn a model that can predict likely values of given a new instance .
The last task is to quantify the causal influence of a distinguished feature named the *exposure* or *treatment* on the outcome as a contrast of *potential outcome* random variables in two arms of a hypothetical or actually realized randomized controlled trial. If we denote the active treatment value by , and the placebo treatment value by , a mean difference comparison of the two arms, termed the *average causal effect*, would be .

A popular approach to solving inference problems with outcome variables of the type we described is based on maximizing the conditional likelihood function of the data: , with respect to parameter values .

Given the likelihood maximizing values , a classification decision for a new instance is .
Similarly, regression problems can be solved by predicting the expected value of given a new instance , that is . Finally, given a treatment ,
and assumptions of *consistency* and *conditional ignorability*, described in more detail in the next section,
the expected counterfactual response can be estimated from observed data using the *parametric g-formula* [20]:

(1) |

where is the empirical average function.

In this paper, we consider outcome inference problems where some of the features are *sensitive*, in the sense that
making inferences on the outcome based on features carelessly may result in discrimination. For simplicity, we will mostly restrict our attention
to . There are many examples of pairs that have this property. Examples include hiring discrimination ( is a hiring decision, and
is gender), or recidivism prediction in parole hearings (where is a parole decision, and is race).

## 3 Causal Inference

We will represent discrimination formally using the language of causal inference. In causal inference, in addition to the outcome , we distinguish a *treatment variable* , and sometimes also one or more *mediator variables* , or . The primary object of interest in causal inference is the potential outcome, [13], which represents the outcome of the th row (unit) if, possibly contrary to fact, were set to value . Just as the rows of the observed data matrix are assumed to be independent realizations of , potential outcomes are assumed to be realizations of an underlying counterfactual random variable .

Given , comparison of and would allow us quantify the *individual causal effect* of on . Comparison of and in expectation: would allow us to quantify the average causal effect (ACE) of on . In general, the average causal effect is not computed using the conditional expectation , since association of and may be causal or spurious (or some combination of the two).

Causal inference uses assumptions in *causal models* to link observed data with counterfactual contrasts of interest. One such assumption, known as *consistency*, states that the mechanism that determines the value of the outcome does not distinguish the method by which the treatment was assigned, as long as the treatment value assigned was invariant. This is expressed
as . Here reads “the random variable , had been intervened on to whatever value would have naturally attained.” Without assumptions like consistency, a link between observed and counterfactual distributions cannot be established, and thus causal inferences from observed data are impossible.

Aside from consistency, additional assumptions in the causal model are needed to express a counterfactual parameter as a functional of the observed data distribution. When such a functional exists, we say the parameter is *identified* from the observed data under the causal model. The simplest such assumption is known as *ignorability* and states that the probability distribution of treatment assignment does not depend on any counterfactual outcome. In other words, or for any , where denotes marginal independence. Under ignorability and consistency, conditional and counterfactual distributions coincide, since . Ignorability holds by design in studies where treatment is randomly assigned. This forms the basis for the causal validity of randomized controlled trials.

A less stringent assumption which allows causal inferences to be drawn from data where treatment was not randomly assigned is known as *conditional ignorability*. This assumption states that conditional on a set of factors , is independent of any counterfactual outcome. In other words, or for any , where represents conditional independence. Given these assumptions, we have
,
known as the adjustment formula, the backdoor formula, or stratification.
Intuitively, the set acts as a set of *observed confounders*, such that adjusting for their influence suffices to remove all non-causal dependence of and , leaving only the part of the dependence that corresponds to the causal effect. A general characterization of identifiable functionals of causal effects, many of them significantly more complex than the adjustment formula, exists [29, 25].
A popular approach for estimating the ACE identified by this formula is via a parametric model and (1).

### 3.1 Causal Diagrams

Cause effect relationships in causal inference problems are often represented by graphical causal models [27, 16]. Such models generalize independence models on directed acyclic graphs, also known as Bayesian networks [14], to also encode conditional independence statements on counterfactual random variables [19]. In such graphs, vertices represent observed random variables, and absence of directed edges represents absence of direct causal relationships (similarly to how in a Bayesian network it is the absence of edges that is significant and encodes conditional independence). As an example, in Fig. 1 (a), is potentially a direct cause of and , while is not a direct of cause of . Note that in this diagram lies on a causal pathway , and thus mediates a part of the causal influence of on , represented by all directed paths which start at and end at .