Skip to yearly menu bar Skip to main content


Poster

Understanding Inter-Concept Relationships in Concept-Based Models

Naveen Raman · Mateo Espinosa Zarlenga · Mateja Jamnik


Abstract:

Concept-based explainability methods provide insight into deep learning systems by constructing explanations using human-understandable concepts. While the literature on human reasoning demonstrates that we exploit relationships between concepts when solving tasks, it is unclear whether concept-based methods incorporate the rich structure of inter-concept relationships. We analyse the concept representations learnt by concept-based models to understand to what extent these models correctly capture inter-concept relationships. First, we empirically demonstrate that state-of-the-art concept-based models produce representations that lack stability and robustness, and find that widely used methods fail to capture known inter-concept relationships. Then, we develop a novel algorithm which leverages inter-concept relationships to improve concept intervention accuracy, which demonstrates how correctly capturing inter-concept relationships can improve downstream tasks.

Live content is unavailable. Log in and register to view live content