SHAP is a unified approach to explain machine learning model outputs using Shapley values from cooperative game theory. This skill provides comprehensive guidance for:
SHAP works with all model types: tree-based models (XGBoost, LightGBM, CatBoost, Random Forest), deep learning models (TensorFlow, PyTorch, Keras), linear models, and black-box models.
Trigger this skill when users ask about:
Decision Tree:
Tree-based model? (XGBoost, LightGBM, CatBoost, Random Forest, Gradient Boosting)
shap.TreeExplainer (fast, exact)Deep neural network? (TensorFlow, PyTorch, Keras, CNNs, RNNs, Transformers)
shap.DeepExplainer or shap.GradientExplainer
Linear model? (Linear/Logistic Regression, GLMs)
shap.LinearExplainer (extremely fast)Any other model? (SVMs, custom functions, black-box models)
shap.KernelExplainer (model-agnostic but slower)Unsure?
shap.Explainer (automatically selects best algorithm)See references/explainers.md for detailed information on all explainer types.
import shap
# Example with tree-based model (XGBoost)
import xgboost as xgb
# Train model
model = xgb.XGBClassifier().fit(X_train, y_train)
# Create explainer
explainer = shap.TreeExplainer(model)
# Compute SHAP values
shap_values = explainer(X_test)
# The shap_values object contains:
# - values: SHAP values (feature attributions)
# - base_values: Expected model output (baseline)
# - data: Original feature values
For Global Understanding (entire dataset):
# Beeswarm plot - shows feature importance with value distributions
shap.plots.beeswarm(shap_values, max_display=15)
# Bar plot - clean summary of feature importance
shap.plots.bar(shap_values)
For Individual Predictions:
# Waterfall plot - detailed breakdown of single prediction
shap.plots.waterfall(shap_values[0])
# Force plot - additive force visualization
shap.plots.force(shap_values[0])
For Feature Relationships:
# Scatter plot - feature-prediction relationship
shap.plots.scatter(shap_values[:, "Feature_Name"])
# Colored by another feature to show interactions
shap.plots.scatter(shap_values[:, "Age"], color=shap_values[:, "Education"])
See references/plots.md for comprehensive guide on all plot types.
This skill supports several common workflows. Choose the workflow that matches the current task.
Goal: Understand what drives model predictions
Steps:
Example:
# Step 1-2: Setup
explainer = shap.TreeExplainer(model)
shap_values = explainer(X_test)
# Step 3: Global importance
shap.plots.beeswarm(shap_values)
# Step 4: Feature relationships
shap.plots.scatter(shap_values[:, "Most_Important_Feature"])
# Step 5: Individual explanation
shap.plots.waterfall(shap_values[0])
Goal: Identify and fix model issues
Steps:
See references/workflows.md for detailed debugging workflow.
Goal: Use SHAP insights to improve features
Steps:
See references/workflows.md for detailed feature engineering workflow.
Goal: Compare multiple models to select best interpretable option
Steps:
See references/workflows.md for detailed model comparison workflow.
Goal: Detect and analyze model bias across demographic groups
Steps:
See references/workflows.md for detailed fairness analysis workflow.
Goal: Integrate SHAP explanations into production systems
Steps:
See references/workflows.md for detailed production deployment workflow.
Definition: SHAP values quantify each feature's contribution to a prediction, measured as the deviation from the expected model output (baseline).
Properties:
Interpretation:
Example:
Baseline (expected value): 0.30
Feature contributions (SHAP values):
Age: +0.15
Income: +0.10
Education: -0.05
Final prediction: 0.30 + 0.15 + 0.10 - 0.05 = 0.50
Purpose: Represents "typical" input to establish baseline expectations
Selection:
Impact: Baseline affects SHAP value magnitudes but not relative importance
Critical Consideration: Understand what your model outputs
Example: XGBoost classifiers explain margin output (log-odds) by default. To explain probabilities, use model_output="probability" in TreeExplainer.
# 1. Setup
explainer = shap.TreeExplainer(model)
shap_values = explainer(X_test)
# 2. Global importance
shap.plots.beeswarm(shap_values)
shap.plots.bar(shap_values)
# 3. Top feature relationships
top_features = X_test.columns[np.abs(shap_values.values).mean(0).argsort()[-5:]]
for feature in top_features:
shap.plots.scatter(shap_values[:, feature])
# 4. Example predictions
for i in range(5):
shap.plots.waterfall(shap_values[i])
# Define cohorts
cohort1_mask = X_test['Group'] == 'A'
cohort2_mask = X_test['Group'] == 'B'
# Compare feature importance
shap.plots.bar({
"Group A": shap_values[cohort1_mask],
"Group B": shap_values[cohort2_mask]
})
# Find errors
errors = model.predict(X_test) != y_test
error_indices = np.where(errors)[0]
# Explain errors
for idx in error_indices[:5]:
print(f"Sample {idx}:")
shap.plots.waterfall(shap_values[idx])
# Investigate key features
shap.plots.scatter(shap_values[:, "Suspicious_Feature"])
Explainer Speed (fastest to slowest):
LinearExplainer - Nearly instantaneousTreeExplainer - Very fastDeepExplainer - Fast for neural networksGradientExplainer - Fast for neural networksKernelExplainer - Slow (use only when necessary)PermutationExplainer - Very slow but accurateFor Large Datasets:
# Compute SHAP for subset
shap_values = explainer(X_test[:1000])
# Or use batching
batch_size = 100
all_shap_values = []
for i in range(0, len(X_test), batch_size):
batch_shap = explainer(X_test[i:i+batch_size])
all_shap_values.append(batch_shap)
For Visualizations:
# Sample subset for plots
shap.plots.beeswarm(shap_values[:1000])
# Adjust transparency for dense plots
shap.plots.scatter(shap_values[:, "Feature"], alpha=0.3)
For Production:
# Cache explainer
import joblib
joblib.dump(explainer, 'explainer.pkl')
explainer = joblib.load('explainer.pkl')
# Pre-compute for batch predictions
# Only compute top N features for API responses
Problem: Using KernelExplainer for tree models (slow and unnecessary) Solution: Always use TreeExplainer for tree-based models
Problem: DeepExplainer/KernelExplainer with too few background samples Solution: Use 100-1000 representative samples
Problem: Interpreting log-odds as probabilities Solution: Check model output type; understand whether values are probabilities, log-odds, or raw outputs
Problem: Matplotlib backend issues
Solution: Ensure backend is set correctly; use plt.show() if needed
Problem: Default max_display=10 may be too many or too few
Solution: Adjust max_display parameter or use feature clustering
Problem: Computing SHAP for very large datasets Solution: Sample subset, use batching, or ensure using specialized explainer (not KernelExplainer)
show=True (default)import mlflow
with mlflow.start_run():
# Train model
model = train_model(X_train, y_train)
# Compute SHAP
explainer = shap.TreeExplainer(model)
shap_values = explainer(X_test)
# Log plots
shap.plots.beeswarm(shap_values, show=False)
mlflow.log_figure(plt.gcf(), "shap_beeswarm.png")
plt.close()
# Log feature importance metrics
mean_abs_shap = np.abs(shap_values.values).mean(axis=0)
for feature, importance in zip(X_test.columns, mean_abs_shap):
mlflow.log_metric(f"shap_{feature}", importance)
class ExplanationService:
def __init__(self, model_path, explainer_path):
self.model = joblib.load(model_path)
self.explainer = joblib.load(explainer_path)
def predict_with_explanation(self, X):
prediction = self.model.predict(X)
shap_values = self.explainer(X)
return {
'prediction': prediction[0],
'base_value': shap_values.base_values[0],
'feature_contributions': dict(zip(X.columns, shap_values.values[0]))
}
This skill includes comprehensive reference documentation organized by topic:
Complete guide to all explainer classes:
TreeExplainer - Fast, exact explanations for tree-based modelsDeepExplainer - Deep learning models (TensorFlow, PyTorch)KernelExplainer - Model-agnostic (works with any model)LinearExplainer - Fast explanations for linear modelsGradientExplainer - Gradient-based for neural networksPermutationExplainer - Exact but slow for any modelIncludes: Constructor parameters, methods, supported models, when to use, examples, performance considerations.
Comprehensive visualization guide:
Includes: Parameters, use cases, examples, best practices, plot selection guide.
Detailed workflows and best practices:
Includes: Step-by-step instructions, code examples, decision criteria, troubleshooting.
Theoretical foundations:
Includes: Mathematical foundations, proofs, comparisons, advanced topics.
When to load reference files:
explainers.md when user needs detailed information about specific explainer types or parametersplots.md when user needs detailed visualization guidance or exploring plot optionsworkflows.md when user has complex multi-step tasks (debugging, fairness analysis, production deployment)theory.md when user asks about theoretical foundations, Shapley values, or mathematical detailsDefault approach (without loading references):
Loading references:
# To load reference files, use the Read tool with appropriate file path:
# /path/to/shap/references/explainers.md
# /path/to/shap/references/plots.md
# /path/to/shap/references/workflows.md
# /path/to/shap/references/theory.md
Choose the right explainer: Use specialized explainers (TreeExplainer, DeepExplainer, LinearExplainer) when possible; avoid KernelExplainer unless necessary
Start global, then go local: Begin with beeswarm/bar plots for overall understanding, then dive into waterfall/scatter plots for details
Use multiple visualizations: Different plots reveal different insights; combine global (beeswarm) + local (waterfall) + relationship (scatter) views
Select appropriate background data: Use 50-1000 representative samples from training data
Understand model output units: Know whether explaining probabilities, log-odds, or raw outputs
Validate with domain knowledge: SHAP shows model behavior; use domain expertise to interpret and validate
Optimize for performance: Sample subsets for visualization, batch for large datasets, cache explainers in production
Check for data leakage: Unexpectedly high feature importance may indicate data quality issues
Consider feature correlations: Use TreeExplainer's correlation-aware options or feature clustering for redundant features
Remember SHAP shows association, not causation: Use domain knowledge for causal interpretation
# Basic installation
uv pip install shap
# With visualization dependencies
uv pip install shap matplotlib
# Latest version
uv pip install -U shap
Dependencies: numpy, pandas, scikit-learn, matplotlib, scipy
Optional: xgboost, lightgbm, tensorflow, torch (depending on model types)
This skill provides comprehensive coverage of SHAP for model interpretability across all use cases and model types.