Predict responsibly: Fairness needed in algorithmic decision-making, U of T experts say
University of Toronto PhD student David Madras says many of today鈥檚 algorithms are good at making accurate predictions, but don't know how to handle uncertainty well. If a badly calibrated algorithm makes the wrong decision, it鈥檚 usually very wrong.
鈥淎 human user might understand a situation better than a computer, whether it鈥檚 information not available because it's very qualitative in nature, or something happening in the real world that didn't get inputted into the algorithm,鈥 says Madras, a machine learning researcher in the department of computer science who is also affiliated with the Vector Institute for Artificial Intelligence.
鈥淏oth can be very important and can have an effect on what predictions should be [made].鈥
Madras is presenting his research, , at the International Conference on Learning Representations (ICLR), in Vancouver this week.The conference is focused on the methods and performance of machine learning and brings together leaders in the field.
Madras says he and Toniann Pitassi, a professor in U of T鈥檚 departments of computer science and mathematics and a computational theory expert who also explores computational fairness, as well as Richard Zemel, a U of T professor of computer science and research director of the Vector Institute, have developed their model with fairness in mind. Where there is a degree of uncertainty, an algorithm must have the option to answer, 鈥淚 don鈥檛 know鈥 and defer its decision to a human user.
Madras explains if Facebook were to use an algorithm to automatically tag people in images, it's maybe not so important if the tagging is done wrong. But when individual outcomes are high-impact, the risk can be greater. He says the model hasn鈥檛 been applied yet to any specific application, but rather the researchers are thinking of the types of ways it could be used in real world cases.
鈥淚n medical settings, it can be important to output something that can be interpreted 鈥 there's some amount of uncertainty surrounding its prediction 鈥 and a doctor should decide if treatment should be given.鈥
Madras鈥 graduate supervisor Zemel, who will take up an this summer, is also examining how machine learning can be made to be more expressive, controllable and fair.
Zemel says machine learning based on historical data, such as whether a bank loan was approved or the length of prison sentences, will naturally pick up on biases. And the biases in the data set can play out in a machine鈥檚 predictions, he says.
鈥淚n this paper, we think a lot about an external decision-maker. In order to train up our model, we have to use historical decisions that are made by decision-makers. The outcomes of those decisions, created by existing decision-makers, can be themselves biased or in a sense incomplete.鈥
Madras believes the increased focus on algorithmic fairness alongside issues of privacy, security and safety, will help make machine learning more applicable to high-stakes applications.
鈥淚t鈥檚 raising important questions about the role of an automated system that's making important decisions, and how to get [them] to predict in ways that we want them to.鈥
Madras says he鈥檚 continuing to think about issues of fairness and related areas, like causality: Two things can be correlated 鈥 because they occur together frequently 鈥 but that doesn't mean that one causes the other.
鈥淚f an algorithm is deciding when to give someone a loan, it might learn that people who live in a certain postal code are less likely to pay back loans. But that can be a source of unfairness. It鈥檚 not like living in a certain postal code makes you less likely to pay back a loan,鈥 he says.
鈥淚t鈥檚 an interesting and important set of problems to be working on."