Structured Representation Learning

Interpretability, Robustness, and Transferability for LLMs

AAAI 2026 Tutorial

Time: Tuesday, January 20, 2:00–6:00 PM   |   Place: Peridot 205,Singapore EXPO

Hanqi Yan

Hanqi Yan

King's College London

Guangyi Chen

Guangyi Chen

CMU & MBZUAI

Jonathan Richard Schwarz

Jonathan Richard Schwarz

ICL & Thomson Reuters

Abstract. As LLMs move from research labs to real-world applications, understanding and controlling their behavior has become critical, given their rapid evolution and opaque internal mechanisms. This tutorial explores principled representation learning as a foundation for controllability, interpretability, and transferability. Participants will learn how to build interpretable, modular representations that guide model behavior, improve reasoning efficiency, and extend capabilities to new tasks through recombination. Grounded in human-centered measures and careful data design, this tutorial offers a roadmap for more robust, transparent, and trustworthy LLM-assisted systems.

Schedule

Session 1 Introduction
Session 2 The Principles of Representation Learning
Session 3 Representations for Latent Reasoning
Coffee Break
Session 4 Understand and Model Edit via Representation Learning
Session 5 Integrate Models Internals for Self-Improvements
Session 6 Conclusion and Future Work