
Should we be concerned that the decisions of AIs are inscrutable?
Machine learning is a black box – even when the decision is correct, how the algorithm arrived at it can be a mystery
by John Zerilli
John Zerilli is a philosopher with particular interests in cognitive science, artificial intelligence and the law. He is currently a Leverhulme Trust Fellow at the University of Oxford and was previously a Research Fellow at the Leverhulme Centre for the Future of Intelligence at the University of Cambridge. His books include The Adaptable Mind (Oxford University Press, 2020) and A Citizen’s Guide to Artificial Intelligence (MIT Press, 2021).