figshare
Browse

Indirect Reciprocity in Hybrid Human-AI Populations

Download (563.67 kB)
presentation
posted on 2025-05-09, 09:18 authored by Alexandre S. Pires, Fernando P. Santos

Indirect reciprocity (IR) is a key mechanism to explain cooperation in human populations. With IR, individuals are associated with reputations, which can be used by others when deciding to cooperate or defect: the costs of cooperation can therefore be outweighed by the long-term benefits of keeping a specific reputation. IR has been studied assuming human populations. However, social interactions involve nowa- days artificial agents (AAs) such as social bots, conversational agents, or even collaborative robots. It remains unclear how IR dynamics will be affected once artificial agents co-exist with humans. In this project we aim to develop game-theoretical models to investigate the potential effect of AAs in the dynamics of human cooperation. We study settings where artificial agents are potentially subject to the different reputation update rules as the remaining individuals in the population. Furthermore, we consider both settings where reputations are public and setting where reputations are privately held. We show that introducing a small fraction of AAs, with a strategy discriminating based on reputation, increases the cooperation rate in the whole population. Our theoretical work contributes to identify the settings where artificial agents, even with simple hard-coded strategies, can help humans solve social dilemmas of cooperation. At this workshop, we hope to discuss future research avenues where citizens preferences, incentives, and strategic adaptation are considered when designing artificial agents to leverage cooperation in hybrid systems.

History

Usage metrics

    Licence

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC