Podcast Banner

Podcasts

Paul, Weiss Waking Up With AI

Accuracy vs. Fairness in AI

In this week’s episode, Katherine introduces the infamous accuracy versus fairness problem in AI, while pointing to some recent research developments that could show promise in identifying bias in models.

Stream here or subscribe on your
preferred podcast app:

Episode Transcript

Katherine Forrest: Hey, good morning, everyone and welcome to another episode of “Waking Up With AI,” a Paul, Weiss podcast. I'm Katherine Forrest, and today you've only got Katherine Forrest. We have sort of an unusual situation where Anna is actually in Abu Dhabi right now where she's doing what she likes to do best, as we all do, which is talking about AI. So, I'm going to be talking this morning about a really important area in machine learning and all forms of artificial intelligence, which is the trade-off between accuracy and fairness in AI. And this is really an important area that we introduced in one of our very first episodes about how models can actually be accurate — they can be accurate and do what they're supposed to do, but they can actually result in some unfair outcomes. So, we're going to go through that a little bit today.

So, it's been a notorious problem because originally when models were developed, the initial sort of push was to try to get predictive outcomes that would accurately reflect the data set. And that was an enormous effort, and it was an effort that then eventually succeeded. And researchers realized that when they maximized for accuracy off of a data set, sometimes if the data set had issues, there would be impacts on fairness.

Let's go through this a little bit. First, let me give you an example. And so, the example that I want to use is a run-of-the-mill credit card application. And I want to talk about the different kinds of characteristics that can go into that, which is you can look at somebody's where they live and their marital status and educational status and their gender. And potentially racial categories might find their way into it based upon any number of differential characteristics such as zip code or educational institution. You might have a loan loss history. You might have all kinds of indicia in terms of prior debt history, et cetera, et cetera, et cetera.

So, you've got a variety of characteristics that are part of a data set. That data set is the normative world that the AI tool understands as the world that it is to try and make its predictions off of. If that normative world has got embedded biases in it, structural inequalities from the history of a particular organization, the history of a particular set of loan files, a history of the United States of America, for whatever reason, then that data set is all that the AI tool understands for its world. So, it will make predictions off of that data set.