
Join us for an insightful discussion with Elif Akata, a PhD researcher at Helmholtz Munich and the University of Tübingen. She will discuss her recent paper, “Playing Repeated Games with Large Language Models.”
Join us for an insightful discussion with Elif Akata, a PhD researcher at Helmholtz Munich and the University of Tübingen.
She will discuss her recent paper, “Playing Repeated Games with Large Language Models.”
Abstract
Large language models (LLMs) are increasingly used in applications where they interact with humans and other agents. We propose to use behavioural game theory to study LLMs’ cooperation and coordination behaviour. Here we let different LLMs play finitely repeated 2 × 2 games with each other, with human-like strategies, and actual human players. Our results show that LLMs perform particularly well at self-interested games such as the iterated Prisoner’s Dilemma family. However, they behave suboptimally in games that require coordination, such as the Battle of the Sexes. We verify that these behavioural signatures are stable across robustness checks. We also show how GPT-4’s behaviour can be modulated by providing additional information about its opponent and by using a ‘social chain-of-thought’ strategy. This also leads to better scores and more successful coordination when interacting with human players. These results enrich our understanding of LLMs’ social behaviour and pave the way for a behavioural game theory for machines.
About speaker
Elif Akata is a PhD student in machine learning and cognitive science. Her research focuses on understanding how LLMs behave as social, collaborative agents and how we can design systems that effectively interact, adapt, and communicate with humans and each other in dynamic environments.