posted on 2021-11-05, 11:19authored byVahid Yazdanpanah, Sebastian Stein, EH Gerding, Nick JenningsNick Jennings
For developing human-centred trustworthy autonomous systems and ensuring their safe and effective integration with the society, it is crucial to enrich autonomous agents with the capacity to represent and reason about their accountability. This is, on one hand, about their accountability as collaborative teams and, on the other hand, their individual degree of accountability in a team. In this context, accountability is understood as being responsible for failing to deliver a task that a team was allocated and able to fulfil. To that end, the semantic (strategic reasoning) machinery of the Alternating-time Temporal Logic (ATL) is a natural modelling approach as it captures the temporal, strategic, and coalitional dynamics of the notion of accountability. This allows focusing on the main problem on: “Who is accountable for an unfulfilled task in multiagent teams: when, why, and to what extent?” We apply ATL-based semantics to define accountability in multiagent teams and develop a fair and computationally feasible procedure for ascribing a degree of accountability to involved agents in accountable teams. Our main results are on decidability, fairness properties, and computational complexity of the presented accountability ascription methods in multiagent teams.
Proceedings of the Workshop on Artificial Intelligence Safety 2021
co-located with the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI 2021)
This is an Open Access Article. It is published by CEUR-WS.org under the Creative Commons Attribution 4.0 Unported Licence (CC BY). Full details of this licence are available at: http://creativecommons.org/licenses/by/4.0/
Huáscar Espinoza; John McDermid; Xiaowei Huang; Mauricio Castillo-Effen; Xin Cynthia Chen; José Hernández-Orallo; Seán Ó hÉigeartaigh; Richard Mallah; Gabriel Pedroza