Treffer: Evolution of Inherently Interpretable Visual Control Policies

Title:
Evolution of Inherently Interpretable Visual Control Policies
Contributors:
Université Toulouse Capitole (UT Capitole), Communauté d'universités et établissements de Toulouse (Comue de Toulouse), Institut de recherche en informatique de Toulouse (IRIT), Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Université Toulouse - Jean Jaurès (UT2J), Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Université Toulouse III - Paul Sabatier (UT3), Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Centre National de la Recherche Scientifique (CNRS)-Institut National Polytechnique (Toulouse) (Toulouse INP), Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Toulouse Mind & Brain Institut (TMBI), Université Toulouse - Jean Jaurès (UT2J), Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Université Toulouse III - Paul Sabatier (UT3), Università degli studi di Trieste = University of Trieste, Real Expression Artificial Life (IRIT-REVA), Communauté d'universités et établissements de Toulouse (Comue de Toulouse)-Université Toulouse Capitole (UT Capitole), Institut Supérieur de l'Aéronautique et de l'Espace (ISAE-SUPAERO), Institut universitaire de France (IUF), Ministère de l'Education nationale, de l’Enseignement supérieur et de la Recherche (M.E.N.E.S.R.)
Source:
GECCO '25: Proceedings of the Genetic and Evolutionary Computation Conference ; GECCO '25: Genetic and Evolutionary Computation Conference ; https://hal.science/hal-05178539 ; GECCO '25: Genetic and Evolutionary Computation Conference, Jul 2025, Malaga, Spain. pp.358-367, ⟨10.1145/3712256.3726332⟩
Publisher Information:
CCSD
ACM
Publication Year:
2025
Collection:
Université Toulouse 2 - Jean Jaurès: HAL
Subject Geographic:
Document Type:
Konferenz conference object
Language:
English
DOI:
10.1145/3712256.3726332
Rights:
http://creativecommons.org/licenses/by/ ; info:eu-repo/semantics/OpenAccess
Accession Number:
edsbas.30B3F60E
Database:
BASE

Weitere Informationen

International audience ; Vision-based decision-making tasks encompass a wide range of applications, including safety-critical domains where trustworthiness is as key as performance. These tasks are often addressed using Deep Reinforcement Learning (DRL) techniques, based on Artificial Neural Networks (ANNs), to automate sequential decision making. However, the "black-box" nature of ANNs limits their applicability in these settings, where transparency and accountability are essential. To address this, various explanation methods have been proposed; however, they often fall short in fully elucidating the decision-making pipeline of ANNs, a critical aspect for ensuring reliability in safety-critical applications. To bridge this gap, we propose an approach based on Graph-based Genetic Programming (GGP) to generate transparent policies for vision-based control tasks. Our evolved policies are constrained in size and composed of simple and well-understood operational modules, enabling inherent interpretability. We evaluate our method on three Atari games, comparing explanations derived from common explainability techniques to those derived from interpreting the agent's true computational graph. We demonstrate that interpretable policies offer a more complete view of the decision process than explainability methods, enabling a full comprehension of competitive game-playing policies.