Papers
Topics
Authors
Recent
Search
2000 character limit reached

Local and Global Explanation Methods

Updated 26 January 2026
  • Local and global explanation methods are formal approaches that decompose model predictions into individual feature contributions and interaction effects for enhanced interpretability.
  • They link partial dependence plots and interventional SHAP values through an ANOVA-style functional decomposition, clarifying both local and global model behavior.
  • They support fairness interventions by enabling the precise removal of protected feature influences, thereby isolating and mitigating bias in model predictions.

Local and global explanation methods are formal approaches used in interpretable machine learning to elucidate how complex models make decisions. Local explanation methods characterize the prediction for an individual instance, typically by attributing contributions to features relevant for that point. Global explanation methods aim to summarize how a model behaves on average, across the entire input domain, by decomposing its prediction function or aggregating local mechanisms. The distinction, and the methodologies that bridge it, are central to modern explainable AI, spanning regression, classification, time-series, text, and vision domains.

1. Formal Decomposition and Marginal Identification

Functional decomposition is a rigorous approach to unify local and global explanation methods by expressing any prediction function f:Rd→Rf : \mathbb{R}^d \to \mathbb{R} as a sum over all feature subsets:

f(x)=∑S⊆{1,…,d}fS(xS)f(x) = \sum_{S \subseteq \{1,\ldots,d\}} f_S(x_S)

Here, f∅=f0f_\emptyset=f_0 (intercept), f{k}(xk)f_{\{k\}}(x_k) encodes the main effect of feature kk, and fS(xS)f_S(x_S) for ∣S∣≥2|S|\ge2 encodes pure interactions. This ANOVA-style expansion is not unique unless an additional identification constraint is imposed.

The marginal-identification constraint requires that for all SS,

∑T:T∩S≠∅∫fT(xT)pS(xS)dxS=0\sum_{T: T\cap S \neq \emptyset} \int f_T(x_T) p_S(x_S) dx_S = 0

where pSp_S is the marginal density of XSX_S. This uniquely pins down each fSf_S via Möbius inversion:

fS(xS)=∑V⊆S(−1)∣S∖V∣∫ffull(x)p−V(x−V)dx−Vf_S(x_S) = \sum_{V \subseteq S} (-1)^{|S \setminus V|} \int f_{\text{full}}(x) p_{-V}(x_{-V}) dx_{-V}

This identification recovers core local and global quantities. Specifically, for any singleton kk, the partial dependence plot (PDP) coincides with f0+fk(xk)f_0 + f_k(x_k), while interventional SHAP values use weighted sums over main effect and interaction components (Hiabu et al., 2022).

2. Relationships with Core Attribution Methods

2.1 Partial Dependence and Feature Effects

The partial dependence plot for kk is:

PDk(xk)=∫f(xk,x−k)p−k(x−k)dx−kPD_k(x_k) = \int f(x_k, x_{-k}) p_{-k}(x_{-k}) dx_{-k}

Under marginal identification, this simplifies to f0+fk(xk)f_0 + f_k(x_k). For any feature subset SS,

PDS(xS)=∫f(x)p−S(x−S)dx−S=∑U⊆SfU(xU)PD_S(x_S) = \int f(x) p_{-S}(x_{-S}) dx_{-S} = \sum_{U \subseteq S} f_U(x_U)

2.2 Interventional SHAP Values

The interventional SHAP value for kk:

ϕk(x)=∑S:k∈S1∣S∣fS(xS)\phi_k(x) = \sum_{S: k \in S} \frac{1}{|S|} f_S(x_S)

Each interaction term fSf_S that involves kk is apportioned $1/|S|$ of its value to feature kk (Hiabu et al., 2022). This recasting connects Shapley values, previously motivated through game theory, to a functional expansion, exposing the direct tie between local and global explanations.

2.3 Feature Importance Measures

Standard SHAP importance E[∣ϕk(X)∣]E[|\phi_k(X)|] mixes main and interaction effects. The decomposition enables refined importance measures:

Ik=E[∑S∋k1∣S∣∣fS(XS)∣]I_k = E \left[ \sum_{S \ni k} \frac{1}{|S|} |f_S(X_S)| \right]

For order separation,

Ik(1)=E[∣fk(Xk)∣],Ik(2)=∑j≠kE[∣fkj(Xk,j)∣],…I_k^{(1)} = E[|f_k(X_k)|],\quad I_k^{(2)} = \sum_{j\ne k} E[|f_{kj}(X_{k,j})|],\ldots

3. Algorithms for Exact and Approximate Functional Decomposition

The number of components fSf_S is generally exponential in dd, but many machine learning models admit low-dimensional base learners (e.g., tree ensembles with interaction order q≪dq \ll d). For such models, exact functional decomposition is feasible:

  • For each tree, identify the split set TtT_t.
  • For each U⊆TtU\subseteq T_t, recursively compute Mt(U,x)M_t(U, x)—the marginal tree value as an expectation over unselected features.
  • For each S⊆TtS\subseteq T_t, assemble fS(xS)f_S(x_S) from Mt(U,x)M_t(U, x) via signed sums.
  • For XGBoost, Algorithm 3 describes efficient aggregation via n×2qn\times 2^q tables; for random planted forests, grid-based marginals suffice.

Complexity is O(B2qn)O(B2^q n) for BB trees of interaction-order qq and nn points (Hiabu et al., 2022).

4. Post-hoc De-biasing and Fairness Interventions

Bias removal targeting protected features relies on the additive decomposition. Consider U⊆{1,…,d}U\subseteq \{1,\ldots,d\} as a block of features (e.g., gender) to eliminate from the model:

∫f(x)pU(xU)dxU=∑S:S∩U=∅fS(xS)\int f(x) p_U(x_U) dx_U = \sum_{S: S \cap U = \emptyset} f_S(x_S)

Thus, forming a debiased predictor entails simply dropping every component fSf_S with S∩U≠∅S\cap U\ne\emptyset from the sum. This approach removes both direct and indirect bias via protected features (Hiabu et al., 2022).

5. Key Experimental Results: Fidelity and Separation of Effects

Empirical studies verify the nuances of functional decomposition:

  • Toy problems: For m(x1,x2)=x1+x2+2x1x2m(x_1,x_2) = x_1 + x_2 + 2x_1x_2 with correlated features, standard interventional SHAP may mask main effects due to cancellation. Decomposition yields distinct main/interaction components f1(x1)f_1(x_1), f2(x2)f_2(x_2), f12(x1,x2)f_{12}(x_1,x_2), which can be inspected separately.
  • Real data: On bike-sharing, decomposing fitted XGBoost reveals strong interaction terms (e.g., hour×\timesworkingday) that standard SHAP summary plots cannot separate.
  • Feature importance simulation: When interactions involve specific features (e.g., x2x3x4x_2x_3x_4), the decomposed importances precisely attribute effects by order.
  • De-biasing: When retraining on a subset of unprotected features, standard refitting fails to remove indirect bias. Decomposition allows rigorous elimination, as shown on simulated salary data and UCI Adult.

6. Integration with Broader Explanation Frameworks

Functional decomposition under marginal identification creates a mathematical bridge between local and global explanations, encompassing:

7. Practical Implications and Trade-offs

This unification addresses critical needs in explainable AI:

  • Separates out main/interaction effects so practitioners can diagnose local anomalies, interaction-induced bias, or global functional structure.
  • Enables robust feature importance assessments not confounded by interaction-induced cancellation.
  • Provides actionable pipelines for post-hoc fairness interventions by component removal, rather than ambiguous retraining.
  • Offers scalable algorithms amenable to tree ensembles and structured models.

The key limitation remains the exponential scaling of decomposition with dimension dd for general black-box models absent low-order structure. Approximate methods, surrogates, and rule-based reductions remain essential for scalability (Seppäläinen et al., 14 Feb 2025, Setzu et al., 2021).


In summary, local and global explanation methods are formalized and unified by functional decomposition with marginal identification constraints. This framework rigorously subsumes interventional SHAP, partial dependence, feature effect curves, and enables principled algorithms for bias reduction and feature importance. Empirical findings demonstrate increased fidelity, interpretability, and robustness, particularly for models and domains with rich interaction structures (Hiabu et al., 2022).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Local and Global Explanation Methods.