Joe’s talk will focus on atomic inference, an approach to model interpretability that involves decomposing a task into discrete atoms, before making predictions for each individual atom and combining these atom-level predictions using interpretable rules. Unlike most interpretability methods, which identify influential features without any guarantees of faithfulness, atomic inference identifies exactly which components of an input are responsible for each model decision. While early work applied atomic inference to simple datasets such as SNLI, these methods have now proven to be successful on challenging datasets such as ANLI. Joe will share why he’s excited by these methods, and will discuss some open challenges that remain for future work. Joe will also discuss some of his recent work on model robustness, including his 2024 ACL paper about creating more robust NLI models by generating synthetic data with LLMs. Joe will highlight some of the data quality issues with LLM-generated data, and will discuss ideas for further research in this area.
Invited Speaker: Joe Stacey (Apple/Imperial College London)
Bio: Joe is an Apple AI/ML scholar in the 4th year of his PhD at Imperial College London, supervised by Marek Rei. Joe’s research involves creating more robust and interpretable NLP models, focusing on the task of Natural Language Inference. Prior to his PhD, Joe worked at PwC for 6 years as an analytics consultant, and also previously taught maths in a challenging secondary school.