Report written by Jon Kleinberg, Jens Ludwig, Sendhil Mullainathan, and Cass R. Sunstein. 45 pages.
The law forbids discrimination. But the ambiguity of human decision-making often makes it extraordinarily hard for the legal system to know whether anyone has actually discriminated. To understand how algorithms affect discrimination, we must therefore also understand how they affect the problem of detecting discrimination. By one measure, algorithms are fundamentally opaque, not just cognitively but even mathematically. Yet for the task of proving discrimination, processes involving algorithms can provide crucial forms of transparency that are otherwise unavailable. These benefits do not happen automatically. But with appropriate requirements in place, the use of algorithms will make it possible to more easily examine and interrogate the entire decision process, thereby making it far easier to know whether discrimination has occurred. By forcing a new level of specificity, the use of algorithms also highlights, and makes transparent, central tradeoffs among competing values. Algorithms are not only a threat to be regulated; with the right safeguards in place, they have the potential to be a positive force for equity.
- Jon Kleinberg — Cornell University – Department of Computer Science
- Jens Ludwig — University of Chicago; National Bureau of Economic Research (NBER)
- Sendhil Mullainathan — Harvard University – Department of Economics; National Bureau of Economic Research (NBER)
- Cass R. Sunstein — Harvard Law School; Harvard University – Harvard Kennedy School (HKS)