Actionable Explainable AI (AxAI) 2023

Special Session: in the Cross Domain Conference for Machine Learning and Knowledge Extraction (CD-MAKE 2023)

to be held in conjunction with the 18th International Conference on Availability, Reliability and Security (ARES 2022 –

August 29 – September 01, 2023

Methods of Explainable Artificial Intelligence (XAI) are developed especially with the goal to make decisions of opaque machine-learned models (e.g., Deep Learning) transparent, interpretable and comprehensible. However, merely establishing transparency, interpretability, and comprehensibility is not enough to derive value from explanations. In addition, it is important to improve models by creating opportunities to act on and learn from explanations. This can be achieved through so-called actionable concepts, methods, measures, and metrics for explainable learning and reasoning (Gunning and Aha, 2019). An important aspect of actionable XAI is the incorporation of psychological insights into the design of explanations and interactive interfaces for the purpose of model understandability, validation, and correctability. Similarly, it is important to develop evaluation criteria that enable meaningful and generalizable comparability of explanations from a user and application perspective. The goal is to find the best possible explanatory approaches for the respective context of use. In this special session, we want to bring together interdisciplinary researchers who are working on exactly these aspects of Explainable Artificial Intelligence and who want to present and discuss new, groundbreaking research that goes beyond testing existing work in new application areas.
Invited contributions: full research papers and short research papers. Extended versions of the accepted papers will be solicited for a special issue to be published on the Machine Learning and Knowledge Extraction (MAKE) journal

Topics of interest include, but are not limited to

  • Approaches for context-aware explainability
  • Concepts and methods for human-centered explainable artificial intelligence
  • Experimental and empirical studies on ML systems that show the suitability of novel explanatory approaches in different application domains (e.g., environmental studies/agriculture/forestry, medicine, autonomous driving, etc.)
  • Methods that allow for corrective feedback by means of a human-in-the-loop to improve explanatory approaches and ML models
  • Methods to integrate human knowledge into automated decision-making in ML systems
  • Novel data set benchmarks to validate ML systems from a user and application domain perspective
  • Novel techniques for the evaluation of ML systems based on the aspect of fidelity and robustness of an explanation
  • Human-centered user interfaces that integrate or compare various XAI methods to facilitate and improve the evaluation of ML systems
  • XAI methods that support safer and more effective human/AI-decision making (e.g., action correction and choice of actions that are more beneficial in a given context)
Important Dates
Extended Submission Deadline March 27, 2023 (AoE)  April 17, 2023
Author Notification June 01, 2023
Proceedings Version June 22, 2023 (AoE)
Conference August 29 – September 01, 2023
Special Session Chairs

Bettina FINZEL, University of Bamberg, Germany
Anna SARANTI, Human-Centered AI Lab, University of Natural Resources and Life Sciences Vienna, Austria

Program Committee 2023

Tim Miller, The University of Melbourne, Australia
Fabrizio Silvestri, University of Rome, Italy
Christian Geißler, Technische Universität Berlin, Germany
Giovanna Nicora, University of Pavia, Italy
Lukas-Valentin Herm, Julius-Maximilians-Universität Würzburg, Germany
Kary Främling, Umeå University, Sweden
Andreas Hinterreiter, Johannes Kepler University Linz, Austria
Hadi Khorshidi, The University of Melbourne, Australia
Andrew Silva, Georgia Institute of Technology, USA
Przemek Biecek, Polish Academy of Science, University of Wroclaw, Poland
Mohan Sridharan, University of Birmingham, England
Satyapriya Krishna, Harvard University, USA
Britta Wrede, Bielefeld University, Germany
Jan Rauch, Prague University of Economics and Business, Czechia
Matthew Gombolay, Massachusetts Institute of Technology, USA
Ute Schmid, University of Bamberg, Germany
Tomas Kliegr, Prague University of Economics and Business, Czechia
Wojciech Samek, Fraunhofer Heinrich-Hertz Institute, Germany
Enea Parimbelli, University of Pavia, Italy
Jan Kretinsky, Technical University of Munich, Germany
Stefano Teso, University of Trento, Italy
Roland Roller, German Research Center for Artificial Intelligence, Germany
Wolf-Tilo Balke, Technical University of Braunschweig, Germany
Diego Hernán Peluffo-Ordoñez, Yachay Tech, Ecuador
Fan Yang, Rice University, USA
Nijat Mehdiyev, German Research Center for Artifical Intelligence and Saarland University, Germany
Julien Aligon, Université de Toulouse, France
Tanya Braun, University of Münster, Germany
Udo Schlegel, University of Konstanz, Germany

Submission Guidelines

The submission guidelines valid for this special track are the same as for the CD-MAKE conference, the submissions will be held to the same quality criteria as all other CD-MAKE submissions and will be published in the CD-MAKE conference proceedings. The guidelines can be found at .