This talk presents advancements in Explainable AI, spanning from classical deep learning to large language models, with contributions that enhance both the usability and usefulness of interpretability methods to improve trust, performance, and safety in AI systems.

Overview

Explainable AI (XAI) plays a pivotal role in enabling understanding and trust in AI systems, especially as their complexity continues to grow. In the era of classical deep learning, research efforts have focused on providing human-understandable explanations for model outputs and behaviors, often delivered through explanation-as-a-service, i.e., usable XAI. More recently, with the advent of large language models, researchers have paid additional attention to understanding the internal representations and mechanisms of these models, leveraging this knowledge to address broader challenges such as safety issues, forming the foundation of useful XAI.

In this talk, I will present my recent contributions to both directions. First, I will discuss the faithfulness issues of attention-based interpretation techniques, a hallmark of usable XAI in the deep learning era, which are unstable against randomness and perturbations during training or testing. I will then introduce editable concept bottleneck models that support removing or inserting some training data or new concepts to address issues like privacy concerns, data mislabelling, spurious concepts, or concept annotation errors. Next, I will discuss Med-MICN, a multi-modal interpretability alignment framework that aligns for various angles, including neural symbolic reasoning, concept semantics, and saliency maps, achieving strong results in the medical domain. Finally, I will share my vision and recent work on usable and useful XAI for LLMs and MLLMs. These works underscore the practical and theoretical advancements required to build XAI systems that are both usable and useful. They highlight the transformative potential of XAI in real-world applications, offering tools to enhance trust, performance, and safety in AI-driven systems.

Presenters

Brief Biography

Lijie Hu is a Ph.D. candidate in the Computer Science program at King Abdullah University of Science and Technology (KAUST), with a Master’s degree in Mathematics from Renmin University of China. Her research focuses on responsible AI, particularly in explainable AI (XAI) and privacy-preserving machine learning. Lijie’s recent research emphasizes making XAI more accessible and practical. Her work centers on developing Usable XAI-as-a-Service systems (Usable XAI) and Useful Explainable AI toolkits (Useful XAI), bridging the gap between theoretical innovation and real-world application. Her research was recognized as “Best of PODS 2022”. She has received several prestigious honors, including the KAUST Dean’s List Award in 2022, 2024, and 2025, and was recognized as a Top Reviewer at AISTATS 2023. Beyond her research, Lijie actively contributes to the academic community as a member of the AAAI Student Committee.