Back to Search Start Over

Explainable agents adapt to human behaviour

Authors :
Universitat Politècnica de Catalunya. Doctorat en Intel·ligència Artificial
Universitat Politècnica de Catalunya. Departament de Ciències de la Computació
Barcelona Supercomputing Center
Universitat Politècnica de Catalunya. IDEAI-UPC - Intelligent Data sciEnce and Artificial Intelligence Research Group
Tormos Llorente, Adrián
Giménez Ábalos, Víctor
Domènech Vila, Marc
Gnatyshak, Dmitry
Álvarez Napagao, Sergio
Vázquez Salceda, Javier
Universitat Politècnica de Catalunya. Doctorat en Intel·ligència Artificial
Universitat Politècnica de Catalunya. Departament de Ciències de la Computació
Barcelona Supercomputing Center
Universitat Politècnica de Catalunya. IDEAI-UPC - Intelligent Data sciEnce and Artificial Intelligence Research Group
Tormos Llorente, Adrián
Giménez Ábalos, Víctor
Domènech Vila, Marc
Gnatyshak, Dmitry
Álvarez Napagao, Sergio
Vázquez Salceda, Javier
Publication Year :
2023

Abstract

When integrating artificial agents into physical or digital environments that are shared with humans, agents are often equipped with opaque Machine Learning methods to enable adapting their behaviour to dynamic human needs and environment. This brings about agents that are also opaque and therefore hard to explain. In previous work, we show that we can reduce an opaque agent into an explainable Policy Graph (PG) which works accurately in multi-agent environments. Policy Graphs are based on a discretisation of the world into propositional logic to identify states, and the choice of which discretiser to apply is key to the performance of the reduced agent. In this work, we explore this further by 1) reducing a single agent into an explainable PG, and 2) enforcing collaboration between this agent and an agent trained from human behaviour. The human agent is computed by using GAIL from a series of human-played episodes, and kept unchanged. We show that an opaque agent created and trained to collaborate with the human agent can be reduced to an explainable, non-opaque PG, so long as predicates regarding collaboration are included in the state representation, by showing the difference in reward between the agent and its PG. Code is available at https://github.com/HPAI-BSC/explainable-agents-with-humans<br />This work has been partially supported by EU Horizon 2020 Project StairwAI (grant agreement No. 101017142).<br />Peer Reviewed<br />Postprint (published version)

Details

Database :
OAIster
Notes :
7 p., application/pdf, English
Publication Type :
Electronic Resource
Accession number :
edsoai.on1390666359
Document Type :
Electronic Resource