Papers
Topics
Authors
Recent
Search
2000 character limit reached

Fisher Information Matrix Scores

Updated 17 November 2025
  • Fisher Information Matrix (FIM) scores are quantitative measures derived from the local curvature of the likelihood, guiding uncertainty assessment in parameter estimation.
  • They distinguish between observed information (sample-based Hessian) and expected information (average over the data distribution), affecting coverage properties.
  • The expected FIM is shown to yield more accurate mean-squared coverage error for confidence intervals compared to the observed FIM under regular conditions.

The Fisher Information Matrix (FIM) is a central object in asymptotic theory and interval estimation for parametric models, providing local curvature information about the likelihood and thus quantifying parameter uncertainty. "FIM scores" typically refer both to the quantitative entries of the FIM or its inverse—and, by extension, to the resulting measures of statistical precision or confidence in inference tasks. Two primary forms of the Fisher information matrix arise in practice: the observed FIM, computed directly from the sample via the Hessian of the log-likelihood at the maximum-likelihood estimate (MLE), and the expected FIM, computed as the expectation over the data-generating distribution of the observed information. A long-standing question in theory and practice concerns which form yields more accurate coverage properties for confidence regions or intervals constructed via asymptotic normality of the MLE.

1. Formal Definitions: Observed and Expected FIM

Let X=(X1,,Xn)X = (X_1, \dots, X_n) be independent random variables with joint likelihood L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta) and log-likelihood (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X). The multivariate parameter is θRp\theta \in \mathbb{R}^p.

The two principal forms of the Fisher information matrix are:

  • Observed information at θ\theta:

Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)

  • Expected information at θ\theta:

Iexp(θ)=EX[Iobs(θ;X)]I_{\mathrm{exp}}(\theta) = E_X \left[ I_{\mathrm{obs}}(\theta; X) \right]

At the MLE θ^n\hat{\theta}_n, one typically uses

Hˉn(θ^n)=n1θ2[(θ^n;X)](observed)\bar{H}_n(\hat{\theta}_n) = n^{-1} \nabla^2_\theta \left[ -\ell(\hat{\theta}_n; X) \right] \quad \text{(observed)}

and its expectation, L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)0, for the expected information.

The inverse of either matrix provides the standard asymptotic (plug-in) estimator for the covariance matrix of the MLE.

2. Asymptotic Normality and Confidence Region Construction

Classical large-sample theory, under standard regularity conditions (smoothness of likelihood, identifiability, and interchange of differentiation and expectation), yields the following expansion for the score at the MLE:

L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)1

for some L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)2 between L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)3 and the true parameter L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)4. Rearrangement and scaling by L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)5 gives

L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)6

By the central limit theorem and consistency, as L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)7:

L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)8

L(θ;X)=i=1npi(Xi;θ)L(\theta; X) = \prod_{i=1}^n p_i(X_i; \theta)9

So (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)0.

Approximate confidence intervals for a scalar parameter (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)1 (the (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)2th component of (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)3) take the form:

(θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)4

The variance term (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)5 is typically substituted by the (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)6 entry of either (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)7 or (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)8 at (θ;X)=logL(θ;X)\ell(\theta; X) = \log L(\theta; X)9.

3. MSE Criterion for FIM-Based Interval Coverage

Rather than comparing confidence interval lengths, Jiang & Spall focus on actual coverage probability—how closely the constructed interval attains the nominal confidence level, on average. For component θRp\theta \in \mathbb{R}^p0, let θRp\theta \in \mathbb{R}^p1 denote the true (unknown) asymptotic covariance, and use

θRp\theta \in \mathbb{R}^p2

to denote the realized coverage when θRp\theta \in \mathbb{R}^p3 is the plug-in variance estimate (either θRp\theta \in \mathbb{R}^p4 or θRp\theta \in \mathbb{R}^p5).

Define the mean-squared error (MSE) of coverage error for each estimator: θRp\theta \in \mathbb{R}^p6 A smaller MSE indicates that the approximate interval more closely (in the mean-squared sense) attains its nominal coverage on average.

4. Main Theorem: Superiority of the Expected FIM in Coverage MSE

Theorem (Jiang & Spall):

Under the paper's regularity assumptions (existence and boundedness of derivatives, LLN and CLT for i.n.i.d. data, etc.),

θRp\theta \in \mathbb{R}^p7

If θRp\theta \in \mathbb{R}^p8 and θRp\theta \in \mathbb{R}^p9 differ nontrivially in the limit, this inequality is strict.

Interpretation:

Asymptotically, confidence intervals constructed with θ\theta0 never have larger mean-squared coverage error than those constructed with θ\theta1, and typically perform strictly better component-wise.

Proof Sketch

  • MLE Error Expansion:

θ\theta2

where θ\theta3 are standardized normal scores and θ\theta4 are entries of θ\theta5.

  • Inverse Matrix Expansions:

θ\theta6

θ\theta7

with θ\theta8, zero mean, and controlled variance.

  • Taylor Expansion of the Coverage Function:

θ\theta9

  • Combine Terms to Compare Mean-Squared Errors:

The difference in MSEs is driven by the variance of Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)0 (vanishing in expectation for Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)1, but present for Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)2).

5. Practical Implications and Application Guidelines

  • Interval Estimation:

For constructing confidence intervals or regions based on the asymptotic normality of the MLE, the user must choose to "score" parameter uncertainty with either the observed Hessian or the expected Fisher information.

  • Empirical Recommendation:

Jiang & Spall's result establishes that, under regularity, using the expected FIM never worsens and generally improves the MSE of coverage. Whenever a closed-form or accurate numerical estimate of Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)3 is available, its use in the construction of confidence intervals or regions is strongly justified.

  • Exceptions and Variations:

In scalar cases with available ancillary statistics (Efron & Hinkley), observed information conditional on those statistics can sometimes yield even better coverage. However, for multivariate and i.n.i.d. settings, the expected FIM is at least as good as, and often strictly better than, the observed FIM for mean-squared coverage error.

  • Moderate and Finite Sample Sizes:

While the superiority of Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)4 manifests asymptotically as Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)5 increases, the observed FIM may fluctuate more in small samples due to data-induced noise. When Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)6 can be computed efficiently, it is typically preferred in practice for "scoring" uncertainty.

6. Broader Context and Impact

The findings of Jiang & Spall directly challenge a persistent heuristic preference for the observed FIM in finite-sample estimation without conditional ancillarity. The result is robust across multivariate, non-i.i.d. scenarios and is not limited to specific model classes. In application domains—such as precision forecasting, uncertainty quantification, and experimental design—adopting the expected FIM for uncertainty "scoring" in MLE-based inference yields intervals whose empirical coverage is at least as accurate as, and often more conservative/accurate than, those from observed-FIM-based procedures.

Concisely: In all regular settings where the plug-in asymptotic normal confidence region is used to quantify parameter uncertainty, the expected FIM should be used in place of the observed FIM—inverting Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)7 at Iobs(θ;X)=θ2(θ;X)I_{\mathrm{obs}}(\theta; X) = -\nabla_\theta^2 \ell(\theta; X)8 yields, per parameter, intervals that realize coverage probabilities no worse, and generally better, than their observed-FIM-based counterparts in mean-squared error. This conclusion now stands on rigorous asymptotic and componentwise grounds (Jiang, 2021).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Fisher Information Matrix (FIM) Scores.