Andrii Babii (UNC Chapel Hill),
for
"Binary Choice with Asymmetric Loss and Fairness in
Machine Learning Classification, with an Application to Racial Justice,"
Asymmetric loss is crucially relevant in some situations; consider, for example, classification as "guilty" or "non guilty". Traditional classification methods have a hard time with it, however, as they ultimately treat type I and II errors symmetrically. (See, e.g., here.) This paper makes impressive progress.
Abstract:
The importance of asymmetries in prediction problems arising in economics has been recognized for a long time. In this paper, we focus on binary choice problems in a data-rich environment with general loss functions. In contrast to the asymmetric regression problems, the binary choice with general loss functions and high-dimensional datasets is challenging and not well understood. Econometricians have studied binary choice problems for a long time, but the literature does not offer computationally attractive solutions in data-rich environments. In contrast, the machine learning literature has many computationally attractive algorithms that form the basis for much of the automated procedures that are implemented in practice, but it is focused on symmetric loss functions that are independent of individual characteristics. One of the main contributions of our paper is to show that the theoretically valid predictions of binary outcomes with arbitrary loss functions can be achieved via a very simple reweighting of the logistic regression, or other state-of-the-art machine learning techniques, such as boosting or (deep) neural networks. We apply our analysis to racial justice in pretrial detention.
See also here.
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.