Types of Explainability in AI Systems: Post-hoc and Intrinsic
๐งฉ Categories of Explainability in AI
Explainability in AI can be divided into two main categories:
๐ง 1. Post-Hoc Explainability
Generates explanations after a model makes a decision.
Applicable to complex, black-box models.
๐ง Techniques:
- ๐ Feature importance
- ๐งช Surrogate models
- ๐งฌ Example-based explanations
๐งฐ Common Tools:
- ๐งฎ SHAP (SHapley Additive Explanations)
- ๐ LIME (Local Interpretable Model-agnostic Explanations)
๐ Example:
Explaining a deep neural network using SHAP or LIME.
๐ข 2. Intrinsic Interpretability
Models that are inherently understandable and transparent by design.
๐ง Examples:
- ๐ณ Decision Trees
- โ Linear Regression / Logistic Regression
๐ Example:
A decision tree clearly shows the decision path for each prediction.
๐ฏ Choosing the Right Approach
Understanding these categories helps in selecting suitable techniques based on:
- โ๏ธ Application complexity
- ๐ Performance needs
- ๐ Transparency requirements