< Back to news
For computational linguists like Fernández, large language models suddenly offer a new tool to quantify all kinds of properties of human dialogues and test whether certain hypotheses about human language use are correct. For example, one of the theories from psycholinguistics is that people unconsciously adapt their language use in such a way that their conversation partner can understand them with as little effort as possible. For example, by making a sentence shorter, or by using simpler words or simpler constructions. Fernández: 'With these powerful language models we can to a certain extent quantify how people use language. Then we see that people do indeed try to speak in such a way that the other person understands them with minimal effort. But we also see that for some sentences and for some language use the models underestimate that effort. That's because large language models are trained on many more texts than you and I can ever read.'
Although large language models are great at generating language, it is difficult to get them to do a specific task, such as booking a restaurant or a ticket. Fernández: 'Language models generate what is most likely and they are not trained to work with you to achieve the goal you have in mind. To achieve this, the system must know what the goal is and how it can achieve it. That is a major challenge at the moment.'
Read the entire article here.
Published by the University of Amsterdam.
14 February 2024
Raquel Fernández tries to make chatbots more human
Last year, AI systems that write human texts made their breakthrough worldwide. Yet many scientific questions about how exactly they work remain unanswered. The UvA asked three UvA researchers how they try to make the underlying language models more transparent, reliable and human. This week, colleague Raquel Fernández talks about her research.
UvA professor of Computational Linguistics & Dialogue Systems Raquel Fernández, affiliated with the ILLC, tries to bridge the gap between large language models and the way people use language. Fernández, who leads the Dialogue Modeling research group: 'I am interested in how people talk to each other and how we can naturally transfer this ability to machines.'
For computational linguists like Fernández, large language models suddenly offer a new tool to quantify all kinds of properties of human dialogues and test whether certain hypotheses about human language use are correct. For example, one of the theories from psycholinguistics is that people unconsciously adapt their language use in such a way that their conversation partner can understand them with as little effort as possible. For example, by making a sentence shorter, or by using simpler words or simpler constructions. Fernández: 'With these powerful language models we can to a certain extent quantify how people use language. Then we see that people do indeed try to speak in such a way that the other person understands them with minimal effort. But we also see that for some sentences and for some language use the models underestimate that effort. That's because large language models are trained on many more texts than you and I can ever read.'
Although large language models are great at generating language, it is difficult to get them to do a specific task, such as booking a restaurant or a ticket. Fernández: 'Language models generate what is most likely and they are not trained to work with you to achieve the goal you have in mind. To achieve this, the system must know what the goal is and how it can achieve it. That is a major challenge at the moment.'
Read the entire article here.
Published by the University of Amsterdam.
Vergelijkbaar >
Similar news items
14 November 2024
The Amsterdam Vision on AI: A Realistic View on Artificial Intelligence
In its new policy, The Amsterdam Vision on AI , the city outlines how artificial intelligence (AI) should be integrated into urban life and how it should influence the city according to its residents. This vision was developed through months of conversations and dialogues with a wide range of Amsterdammers—from festival-goers to schoolchildren, experts to novices—who shared their thoughts on the future role of AI in Amsterdam.
read more >
14 November 2024
Interview: KPN Responsible AI Lab with Gianluigi Bardelloni and Eric Postma
ICAI's interview appeared this time with Gianluigi Bardelloni and Eric Postma, they talk about the developments in their ICAI Lab.
read more >
November 14
AI pilots TLC Science: generative AI in academic education
The University of Amsterdam has launched a new project through its Teaching & Learning Centre Science, exploring how Generative AI, like ChatGPT, can enhance academic education. This pilot program at the Faculty of Science tests and evaluates various applications of GenAI in higher education.
read more >