Goto

Collaborating Authors

 Ritov, Ya'acov


Algorithmic Fairness in Performative Policy Learning: Escaping the Impossibility of Group Fairness

arXiv.org Machine Learning

In many prediction problems, the predictive model affects the distribution of the prediction target. This phenomenon is known as performativity and is often caused by the behavior of individuals with vested interests in the outcome of the predictive model. Although performativity is generally problematic because it manifests as distribution shifts, we develop algorithmic fairness practices that leverage performativity to achieve stronger group fairness guarantees in social classification problems (compared to what is achievable in non-performative settings). In particular, we leverage the policymaker's ability to steer the population to remedy inequities in the long term. A crucial benefit of this approach is that it is possible to resolve the incompatibilities between conflicting group fairness definitions.


A statistical framework for weak-to-strong generalization

arXiv.org Machine Learning

Modern large language model (LLM) alignment techniques rely on human feedback, but it is unclear whether the techniques fundamentally limit the capabilities of aligned LLMs. In particular, it is unclear whether it is possible to align (stronger) LLMs with superhuman capabilities with (weaker) human feedback without degrading their capabilities. This is an instance of the weak-to-strong generalization problem: using weaker (less capable) feedback to train a stronger (more capable) model. We prove that weak-to-strong generalization is possible by eliciting latent knowledge from pre-trained LLMs. In particular, we cast the weak-to-strong generalization problem as a transfer learning problem in which we wish to transfer a latent concept from a weak model to a strong pre-trained model. We prove that a naive fine-tuning approach suffers from fundamental limitations, but an alternative refinement-based approach suggested by the problem structure provably overcomes the limitations of fine-tuning. Finally, we demonstrate the practical applicability of the refinement approach with three LLM alignment tasks.


Learning In Reverse Causal Strategic Environments With Ramifications on Two Sided Markets

arXiv.org Machine Learning

Motivated by equilibrium models of labor markets, we develop a formulation of causal strategic classification in which strategic agents can directly manipulate their outcomes. As an application, we compare employers that anticipate the strategic response of a labor force with employers that do not. We show through a combination of theory and experiment that employers with performatively optimal hiring policies improve employer reward, labor force skill level, and in some cases labor force equity. On the other hand, we demonstrate that performative employers harm labor force utility and fail to prevent discrimination in other cases. In many applications of predictive modeling, the model itself may affect the distribution of samples on which it has to make predictions; this problem is known as strategic classification (Hardt et al., 2015; Brückner et al., 2012) or performative prediction (Perdomo et al., 2020). For example, traffic predictions affect route decisions, which ultimately impact traffic. Such situations can arise in a variety of applications; a common theme is that the samples correspond to strategic agents with an incentive to "game the system" and elicit a desired outcome from the model. In the standard strategic classification setup, the agents are allowed to modify their features, but they do not modify the outcome that the predictive model targets. An example of this is spam classification: spammers craft their messages (e.g. There is a line of work on causal strategic classification that seeks to generalize this setup by allowing the agents to change both their features and outcomes, usually by incorporating a causal model between the two (Miller et al., 2020; Kleinberg and Raghavan, 2020; Haghtalab et al., 2023; Horowitz and Rosenfeld, 2023).


On conditional parity as a notion of non-discrimination in machine learning

arXiv.org Machine Learning

We identify conditional parity as a general notion of non-discrimination in machine learning. In fact, several recently proposed notions of non-discrimination, including a few counterfactual notions, are instances of conditional parity. We show that conditional parity is amenable to statistical analysis by studying randomization as a general mechanism for achieving conditional parity and a kernel-based test of conditional parity.


LLE with low-dimensional neighborhood representation

arXiv.org Machine Learning

The local linear embedding algorithm (LLE) is a non-linear dimension-reducing technique, widely used due to its computational simplicity and intuitive approach. LLE first linearly reconstructs each input point from its nearest neighbors and then preserves these neighborhood relations in the low-dimensional embedding. We show that the reconstruction weights computed by LLE capture the high-dimensional structure of the neighborhoods, and not the low-dimensional manifold structure. Consequently, the weight vectors are highly sensitive to noise. Moreover, this causes LLE to converge to a linear projection of the input, as opposed to its non-linear embedding goal. To overcome both of these problems, we propose to compute the weight vectors using a low-dimensional neighborhood representation. We prove theoretically that this straightforward and computationally simple modification of LLE reduces LLE's sensitivity to noise. This modification also removes the need for regularization when the number of neighbors is larger than the dimension of the input. We present numerical examples demonstrating both the perturbation and linear projection problems, and the improved outputs using the low-dimensional neighborhood representation.