Resources
Here, we include some useful resources for reasearch in concept representation learning.
Relevant Papers and Materials
Below we include a list of works in Concept representation learning, particularly in the areas of Interpretability/Explainability, that are relevant to concept-based interpretable deep learning. We will discuss several of these papers in our tutorial, however we thought that it may be benefitial to write them down in list format for people to access these works more easily. Please keep in mind that this is in no way an exhaustive list of important works within concept learning as this is a fast moving field and we have only so much space we can use here. Nevertheless, we still hope you may find this list helpful if you want to get a sense of where the field is and where it is heading.
Concept Learning Surveys
These are some of the surveys that touch on concept representation learning and its use in interpretable/explainable AI:
2023
- Concept-based Explainable Artificial Intelligence: A SurveyarXiv preprint arXiv:2312.12936, 2023
2022
2020
Various Aspects of XAI
Similarly, there are several key surveys/works that discuss formalisms, definitons, and limitatons of key ideas in the general field of XAI. These works touch upon definitions of what it means to explain a model and on some of the issues of so-called “traditional” XAI approaches (e.g., saliency methods):
2023
- Dear XAI community, we need to talk! Fundamental misconceptions in current XAI researchIn World Conference on Explainable Artificial Intelligence , 2023
2022
- The Disagreement Problem in Explainable Machine Learning: A Practitioner’s PerspectiveTransactions on Machine Learning Research, 2022
- How cognitive biases affect XAI-assisted decision-making: A systematic reviewIn Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society , 2022
2021
- A historical perspective of explainable artificial intelligenceWiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 2021
- Notions of explainability and evaluation approaches for explainable artificial intelligenceInformation Fusion, 2021
2020
- Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AIInformation fusion, 2020
2019
- Stop explaining black box machine learning models for high stakes decisions and use interpretable models insteadNature machine intelligence, 2019
- Explanations can be manipulated and geometry is to blameAdvances in neural information processing systems, 2019
- Interpretation of neural networks is fragileIn Proceedings of the AAAI conference on artificial intelligence , 2019
2018
- Explaining explanations: An overview of interpretability of machine learningIn 2018 IEEE 5th International Conference on data science and advanced analytics (DSAA) , 2018
- Explainable AI: the new 42?In International cross-domain conference for machine learning and knowledge extraction , 2018
-
Supervised Concept Learning
Here we include some relevant works in concept representation learning that assume concept-labels are provided in some manner to learn concept representations from which explanations can be then constructed:
2024
-
- Understanding inter-concept relationships in concept-based modelsIn Proceedings of the 41st International Conference on Machine Learning , 2024
- A Neuro-Symbolic Benchmark Suite for Concept Quality and Reasoning ShortcutsIn The Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track , 2024
- Energy-Based Concept Bottleneck Models: Unifying Prediction, Concept Intervention, and Probabilistic InterpretationsIn The Twelfth International Conference on Learning Representations , 2024
- Concept-Based Interpretable Reinforcement Learning with Limited to No Human LabelsIn Workshop on Interpretable Policies in Reinforcement Learning at RLC-2024 , 2024
- Learning to Intervene on Concept BottlenecksIn Forty-first International Conference on Machine Learning , 2024
- Stochastic Concept Bottleneck ModelsIn The Thirty-eighth Annual Conference on Neural Information Processing Systems , 2024
- Beyond concept bottleneck models: How to make black boxes intervenable?In 38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024), Vancouver, Canada, December 10-15, 2024 , 2024
2023
-
-
-
-
-
-
- Understanding and enhancing robustness of concept-based modelsIn Proceedings of the AAAI Conference on Artificial Intelligence , 2023
- Concept correlation and its effects on concept-based modelsIn Proceedings of the ieee/cvf winter conference on applications of computer vision , 2023
- Towards robust metrics for concept representation evaluationIn Proceedings of the AAAI Conference on Artificial Intelligence , 2023
- Interpretability is in the mind of the beholder: A causal framework for human-interpretable representation learningEntropy, 2023
- A closer look at the intervention procedure of concept bottleneck modelsIn International Conference on Machine Learning , 2023
- Interactive concept bottleneck modelsIn Proceedings of the AAAI Conference on Artificial Intelligence , 2023
2022
- Concept activation regions: A generalized framework for concept-based explanationsAdvances in Neural Information Processing Systems, 2022
- Glancenets: Interpretable, leak-proof concept-based modelsAdvances in Neural Information Processing Systems, 2022
- Concept embedding models: Beyond the accuracy-explainability trade-offAdvances in Neural Information Processing Systems, 2022
- Addressing leakage in concept bottleneck modelsAdvances in Neural Information Processing Systems, 2022
- Learning from uncertain concepts via test time interventionsIn Workshop on Trustworthy and Socially Responsible Machine Learning, NeurIPS 2022 , 2022
2021
-
- Promises and pitfalls of black-box concept learning modelsICML Workshop on Theoretic Foundation, Criticism, and Application Trend of Explainable AI, 2021
2020
2018
- Net2vec: Quantifying and explaining how concepts are encoded by filters in deep neural networksIn Proceedings of the IEEE conference on computer vision and pattern recognition , 2018
- Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (TCAV)In International conference on machine learning , 2018
2017
- Network dissection: Quantifying interpretability of deep visual representationsIn Proceedings of the IEEE conference on computer vision and pattern recognition , 2017
-
Unsupervised Concept Learning
In contrast to the works above, the following papers attempt to learn concept representations without implicit or explicit concept labels. This is done by the means of concept discovery and represents a particularly active are of reasearch in this field:
2025
2023
-
- Tabcbm: Concept-based interpretable neural networks for tabular dataTransactions on Machine Learning Research, 2023
- Bridging the human-ai knowledge gap: Concept discovery and transfer in alphazeroarXiv preprint arXiv:2310.16410, 2023
- Global concept-based interpretability for graph neural networks via neuron analysisIn Proceedings of the AAAI conference on artificial intelligence , 2023
2021
- Gcexplainer: Human-in-the-loop concept-based explanations for graph neural networks3rd ICML Workshop on Human in the Loop Learning,, 2021
2020
- On completeness-aware concept-based explanations in deep neural networksAdvances in neural information processing systems, 2020
2019
- Towards automatic concept-based explanationsAdvances in neural information processing systems, 2019
2018
- Towards robust interpretability with self-explaining neural networksAdvances in neural information processing systems, 2018
Reasoning with Concepts
Finally, we include some papers that describe very interesting things one can do once one has learnt some concept representations (regardless of whether these representations were learnt with or without concept supervision). These works are highly related to the field of neuro-symbolic reasoning and we discuss them in more detail in our presentation:
2024
- DiConStruct: Causal Concept-based Explanations through Black-Box DistillationarXiv preprint arXiv:2401.08534, 2024
2023
2022
-
- Entropy-based logic explanations of neural networksIn Proceedings of the AAAI Conference on Artificial Intelligence , 2022
- Algorithmic concept-based explainable reasoningIn Proceedings of the AAAI Conference on Artificial Intelligence , 2022
2021
2020
2019
- Explaining classifiers with causal concept effect (cace)arXiv preprint arXiv:1907.07165, 2019
2018
-
- Deepproblog: Neural probabilistic logic programmingAdvances in neural information processing systems, 2018
2016
2014
Concept-Learning Public Codebases
Below we list some concept-based open-sourced libraries. As with our reference material, this is by no means an exaustive list but rather one that contains libraries we have had the chance to interact with in the past. If you wish to include your library here, and it is related to concept-learning, please do not hesitate to contact us and we will include it here.