Evaluation metrics like accuracy, precision, and recall don't tell the full story. Stakeholders need to understand how and why models make predictions for debugging, regulatory compliance, building user trust, and making business decisions.
Explainability methods help answer "Why did the model predict X?" Most methods work by assigning attribution values to features—numbers showing how much each feature contributed to the prediction. This section covers the foundational concepts and five major explainability approaches.
What You'll Learn:
All these techniques produce attribution values—numerical indicators of feature importance:
But to interpret "cholesterol = +0.4" or "blood pressure = -0.2," you need to know: relative to what? That's where baselines come in.