The IJCAI 2024 Workshop on Explainable Artificial Intelligence (XAI) is a wrap!
Below, we give some high-level information and our reflections about the 2024 state of the workshop and of XAI in general.
Overall, we had 37 submissions, with 25 articles accepted.
Thanks to our 56-person program committee. The workshop wouldn’t happen without you!
The workshop was split into two days: an in-person workshop in Jeju on 5 August, and a virtual day on 15 August for those authors who could not make it to Jeju. There were 13 papers at the in-person day, 12 online.
The in-person workshop was well attended, with over 100 people registered, and more than 80 people in each session throughout the day.
For those interested, Tim live-Tweeted the in-person workshop, with a short summary of every presentation
At the in-person day, Professor Elisabeth André (left) gave an excellent talk titled Presenting Users with Alternative Realities: Counterfactual, Semi-Factual, and Alterfactual Explanations in XAI, which demonstrated the benefits of combining disciplines, in this case, AI, HCI, and cognitive science.
Our poster session was also a success again – so much lively discussion that gave attendees a chance to talk in detail with authors about their work. As well as workshop authors, we invited authors of papers related to the workshop to come along and present.
Some high-level thoughts on the current states of the field based on observations from the workshop papers:
It is clear that the field has matured a lot in the last 5 years. At the first instance of the workshop in 2017, many of the presentations were on algorithms that could highlight feature importance or some other information, with minimal consideration of whether this would be understandable, there were very few (if any?) human studies, and little focus on what the goal of the algorithms were: understandability, assistance on a downstream task, etc. In the 2024 edition, there were studies using co-design approaches with stakeholders, much more interest in the goals of methods, and several applied studies.
Similar to the above, there is a much greater focus on evaluation (see right some survey results from the online day). Not just papers proposing evaluation, but many authors had clearly thought hard about their evaluation, what it means, and how it relates to their goals. New metrics were being discussed, with in-depth ideas about how they relate to other metrics.
There is much less focus on perturbation-based feature attribution than previously – this having been identified as a source of error and instability, so should be used carefully.
Counterfactuals (and other *-actuals, such as alter-factuals and semi-factuals) are still having success in human behavioural studies, providing further evidence that these techniques do have some cognitive merit.
Until next year!
Tim, Mor, Tobias, and Hendrik
(And big thanks to Marija Slavkovik for giving us the idea of 'publishing' a reflection with our proceedings).
The full proceedings can be found here.
Generating Global Policy Summaries for Reinforcement Learning Agents Using Large Language Models
Sahar Admoni, Ofra Amir
Analyzing the Influence of Training Samples on Explanations
André Artelt, Barbara Hammer
Class-Discriminative Attention Maps for Vision Transformers
Lennart Brocki, Neo Christopher Chung
Enhancing XAI with LLMs: A Banking Marketing Recommendation Case Study
Alessandro Castelnovo, Roberto Depalmas, Fabio Mercorio, Nicolò Mombelli, Daniele Potertì, Antonio Serino,
Andrea Seveso, Salvatore Sorrentino, Laura Viola
The Extrapolation Power of Implicit Models
Juliette Decugis, Alicia Y. Tsai, Max Emerling, Ashwin Ganesh, Laurent El Ghaoui
Contextual Importance and Utility in Python: New Functionality and Insights with the Py-Ciu Package
Kary Främling
Mengisti Berihu Girmay , Felix Möhrle
Explaining Graph Classification with Language-Augmented Graph Concept Bottleneck Model
Lijie Hu, Huanyi Xie, Lu Yu, Tianhao Huang, Muhammad Asif Ali, Di Wang
A Harmonic Metric for LLM Trustworthiness
Nicholas S. Kersting, Mohammad Rahman, Suchismitha Vedala, Yang Wang
Towards Gradient-Based Time-Series Explanations Through a SpatioTemporal Attention Network
Min Hun Lee
“You just can’t go around killing people” Explaining Agent Behavior to a Human Terminator
Uri Menkes, Assaf Hallak, Ofra Amir
Learning Symbolic Expressions to Solve Multi-Period Time Slot Pricing Vehicle Routing Problems
Fábio Neves-Moreira, Daniela Fernandes, Miguel Lunet, Pedro Amorim
Yasunobu Nohara
I^2 AM: Interpreting Image-to-Image Latent Diffusion Models via Attribution Maps
Junseo Park, Hyeryung Jang
Generating Part-Based Global Explanations Via Correspondence
Kunal Rathore , Prasad Tadepalli
On the Feasibility of Fidelity- for Graph Pruning
Yong-Min Shin, Won-Yong Shin
Improving Explainability of Softmax Classifiers Using a Prototype-Based Joint Embedding Method
Hilarie Sit, Brendan Keith, Karianne Bergen
Can Unfairness in ML Decision-Making Processes be Assessed Through the Lens of Formal Explanations?
Belona Sonna, Alban Grastien
Can Explanations Increase Teachers’ Trust and Satisfaction? An Empirical Study
Deliang Wang, Gaowei Chen
The Impact of an XAI-Augmented Approach on Binary Classification with Scarce Data
Ximing Wen, Rosina O. Weber, Anik Sen, Darryl Hannan, Steven C. Nesbit, Vincent Chan, Alberto Goffi, Michael Morris, John C. Hunninghake, Nicholas E. Villalobos, Edward Kim, Christopher J. MacLellan
XGeoS-AI: An Interpretable Learning Framework for Deciphering Geoscience Image Segmentation
Jin-Jian Xu, Hao Zhang, Chao-Sheng Tang, Lin Li, Dian-Long Wang, Bo Liu, Bin Shi
Hyeonggeun Yun
Challenges in Interpretability of Additive Models
Xinyu Zhang, Julien Martinelli, S.T. John
ONNXExplainer: an ONNX Based Generic Framework to Explain Neural Networks Using Shapley Values
Yong Zhao, Runxin He, Nicholas Kersting, Can Liu, Shubham Agrawal, Chiranjeet Chetia, Yu Gu
Ziwei Zhao, David Leake, Xiaomeng Ye, David Crandall