Jennings_Accountability_Ascription_in_Multiagent_Teams.pdf (311.62 kB)
Download file

Applying strategic reasoning for accountability ascription in multiagent teams

Download (311.62 kB)
conference contribution
posted on 05.11.2021, 11:19 by Vahid Yazdanpanah, Sebastian Stein, EH Gerding, Nick JenningsNick Jennings
For developing human-centred trustworthy autonomous systems and ensuring their safe and effective integration with the society, it is crucial to enrich autonomous agents with the capacity to represent and reason about their accountability. This is, on one hand, about their accountability as collaborative teams and, on the other hand, their individual degree of accountability in a team. In this context, accountability is understood as being responsible for failing to deliver a task that a team was allocated and able to fulfil. To that end, the semantic (strategic reasoning) machinery of the Alternating-time Temporal Logic (ATL) is a natural modelling approach as it captures the temporal, strategic, and coalitional dynamics of the notion of accountability. This allows focusing on the main problem on: “Who is accountable for an unfulfilled task in multiagent teams: when, why, and to what extent?” We apply ATL-based semantics to define accountability in multiagent teams and develop a fair and computationally feasible procedure for ascribing a degree of accountability to involved agents in accountable teams. Our main results are on decidability, fairness properties, and computational complexity of the presented accountability ascription methods in multiagent teams.


UKRI Trustworthy Autonomous Systems Hub

Engineering and Physical Sciences Research Council

Find out more...

AutoTrust: Designing a Human-Centered Trusted, Secure, Intelligent and Usable Internet of Vehicles

Engineering and Physical Sciences Research Council

Find out more...

Turing AI Fellowship: Citizen-Centric AI Systems

Engineering and Physical Sciences Research Council

Find out more...


Published in

Proceedings of the Workshop on Artificial Intelligence Safety 2021 co-located with the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI 2021)




AISafety 2021. Artificial Intelligence Safety 2021



VoR (Version of Record)

Rights holder

Copyright © 2021 for the individual papers by the papers' authors. Copyright © 2021 for the volume as a collection by its editors

Publisher statement

This is an Open Access Article. It is published by under the Creative Commons Attribution 4.0 Unported Licence (CC BY). Full details of this licence are available at:

Publication date


Copyright date




Book series

CEUR Workshop Proceedings vol. 2916




Huáscar Espinoza; John McDermid; Xiaowei Huang; Mauricio Castillo-Effen; Xin Cynthia Chen; José Hernández-Orallo; Seán Ó hÉigeartaigh; Richard Mallah; Gabriel Pedroza


Deposit date: 4 November 2021

Usage metrics