< Back to news

July 4, 2024

Developing a method to make AI explainable to humans

AI can take over many of our tasks, creating endless possibilities. But how can we ensure that AI models are understandable and explainable to humans? In a new, interdisciplinary research project, UvA researchers are developing a method for this. ‘We accept more easily what makes sense to us – and that can lead us to trust systems that are not trustworthy.’

AI models can solve many tasks, but they are also becoming increasingly complex. The field of Explainable AI (XAI) is concerned with unpacking the complex behaviour of these models in a way that humans can understand. In the project HUE: bridging AI representations to Human-Understandable Explanations, researchers Giovanni Cinà (Faculty of Medicine) and Sandro Pezzelle (Faculty of Science) are developing a method that will make it possible to ‘x-ray’ AI models and make them more transparent.

 

'Many AI models are black boxes,' explains Pezzelle. 'We can feed them with a lot of data and they can make a prediction – which may or may not be correct – but we do not know what goes on internally.' This is problematic, because we tend to interpret the output according to our own expectations, also known as confirmation bias.

 

Cinà: 'We are more likely to believe explanations that match our prior beliefs. We accept more easily what makes sense to us, and that can lead us to trust models that are not really trustworthy. This is a big problem, for instance when we use AI models to interpret medical data in order to detect disease. Unreliable models may start to influence doctors and lead them to misdiagnose results.'

 

The researchers are developing a method to mitigate this confirmation bias. 'We want to align what we think the model is doing with what it is actually doing', Cinà says. 'To make a model more transparent, we need to examine some explanations for why it came up with a certain prediction.' To do this, the researchers are creating a formal framework that allows them to formulate human-understandable hypotheses about what the model has learned, and to test these more precisely.

 

Pezzelle: 'Our method can be applied to any machine learning or deep learning model, as long as we can inspect it. For that reason, a model like ChatGPT is not a good candidate, because we cannot look into it: we only get its final output. The model has to be open source for our method to work.'

 

Cinà and Pezzelle, who come from different academic backgrounds – medical AI and natural language processing (NLP), respectively – have joined forces in order to develop a method that can be applied to various domains. Pezzelle: 'Currently, solutions that are proposed in one of these disciplines do not necessarily reach the other field. So our aim is to create a more unified approach.'

 

Cinà: 'There is a technical challenge to that, and also a challenge in terms of expertise: we talk about systems that are roughly similar, but we have very different terminology. But at the same time, it is very valuable to be able to use each other’s expertise.'


Source: UvA.nl 

Vergelijkbaar >

Similar news items

>View all news items >
TikTok turns to AI, cuts hundreds of jobs worldwide

October 14

TikTok turns to AI, cuts hundreds of jobs worldwide >

TikTok, the popular social media platform, has cut hundreds of jobs globally, particularly in the content moderation department. The company aims to rely more on artificial intelligence (AI) for moderating content. This decision has a significant impact, especially in Malaysia and the Netherlands, according to various sources such as NOS , de Volkskrant , and NRC .

read more >

Analysis: OpenAI moves from idealism to profit-driven future

October 14

Analysis: OpenAI moves from idealism to profit-driven future >

OpenAI, the company behind ChatGPT, has raised $6.6 billion from investors, bringing its valuation to $157 billion. Once established as a nonprofit organization with idealistic goals, OpenAI is now transitioning into a profit-driven enterprise. These developments and the financial restructuring mark a new phase for OpenAI, as analyzed by de Volkskrant .

read more >

Responsible AI: A Collaborative Effort

October 14

Responsible AI: A Collaborative Effort >

How can AI be developed in a way that avoids discrimination and excessive energy consumption, but instead benefits society? This is the central question of a large national AI initiative housed at the Informatics Institute of the University of Amsterdam (UvA). In collaboration with companies and (semi-)government institutions, the initiative focuses on responsible AI solutions for societal challenges such as food waste, cancer treatments, and combating discrimination.

read more >