SVG Image
< Back to news

July 4, 2024

Developing a method to make AI explainable to humans

AI can take over many of our tasks, creating endless possibilities. But how can we ensure that AI models are understandable and explainable to humans? In a new, interdisciplinary research project, UvA researchers are developing a method for this. ‘We accept more easily what makes sense to us – and that can lead us to trust systems that are not trustworthy.’

AI models can solve many tasks, but they are also becoming increasingly complex. The field of Explainable AI (XAI) is concerned with unpacking the complex behaviour of these models in a way that humans can understand. In the project HUE: bridging AI representations to Human-Understandable Explanations, researchers Giovanni Cinà (Faculty of Medicine) and Sandro Pezzelle (Faculty of Science) are developing a method that will make it possible to ‘x-ray’ AI models and make them more transparent.

 

'Many AI models are black boxes,' explains Pezzelle. 'We can feed them with a lot of data and they can make a prediction – which may or may not be correct – but we do not know what goes on internally.' This is problematic, because we tend to interpret the output according to our own expectations, also known as confirmation bias.

 

Cinà: 'We are more likely to believe explanations that match our prior beliefs. We accept more easily what makes sense to us, and that can lead us to trust models that are not really trustworthy. This is a big problem, for instance when we use AI models to interpret medical data in order to detect disease. Unreliable models may start to influence doctors and lead them to misdiagnose results.'

 

The researchers are developing a method to mitigate this confirmation bias. 'We want to align what we think the model is doing with what it is actually doing', Cinà says. 'To make a model more transparent, we need to examine some explanations for why it came up with a certain prediction.' To do this, the researchers are creating a formal framework that allows them to formulate human-understandable hypotheses about what the model has learned, and to test these more precisely.

 

Pezzelle: 'Our method can be applied to any machine learning or deep learning model, as long as we can inspect it. For that reason, a model like ChatGPT is not a good candidate, because we cannot look into it: we only get its final output. The model has to be open source for our method to work.'

 

Cinà and Pezzelle, who come from different academic backgrounds – medical AI and natural language processing (NLP), respectively – have joined forces in order to develop a method that can be applied to various domains. Pezzelle: 'Currently, solutions that are proposed in one of these disciplines do not necessarily reach the other field. So our aim is to create a more unified approach.'

 

Cinà: 'There is a technical challenge to that, and also a challenge in terms of expertise: we talk about systems that are roughly similar, but we have very different terminology. But at the same time, it is very valuable to be able to use each other’s expertise.'


Source: UvA.nl 

Vergelijkbaar >

Similar news items

>View all news items >
 A Look Ahead at Cardiovascular Care: A Triple Inaugural Lecture from Amsterdam UMC Heart Center

13 January 2025

A Look Ahead at Cardiovascular Care: A Triple Inaugural Lecture from Amsterdam UMC Heart Center >

In a unique joint inaugural lecture, Professors Folkert Asselbergs, Steven Chamuleau, and Robert Klautz explored the future of cardiovascular care. Their vision focuses on the year 2040, a time when technological innovations, network medicine, and artificial intelligence (AI) will fundamentally transform the roles of cardiologists and cardiac surgeons.

read more >

Spring 2025 round of NWO’s Take-off programme now open

January 9

Spring 2025 round of NWO’s Take-off programme now open >

The Netherlands Organisation for Scientific Research (NWO) has opened the Spring 2025 round of its Take-off programme. As of January 6, 2025, researchers can apply for funding for feasibility studies and early-phase projects to bring innovative scientific ideas to the market.

read more >

CWI Research Semester Programme: Control Theory and Reinforcement Learning

January 10

CWI Research Semester Programme: Control Theory and Reinforcement Learning >

In Spring 2025, the Centrum Wiskunde & Informatica (CWI) will host a research programme themed "Control Theory and Reinforcement Learning: Connections and Challenges." The programme brings together researchers and students to explore the intersections and challenges of these two fields.

read more >