Orthant Normal Distribution
- Orthant normal distribution is defined as the probability that all components of a multivariate Gaussian are nonnegative, with applications in Gaussian process theory and polyhedral geometry.
- Specialized techniques like Steck’s reduction for equicorrelated structures and asymptotic bounds provide efficient estimation methods for these multidimensional probabilities.
- The holonomic gradient method reformulates the problem as a system of ODEs, offering enhanced numerical stability and competitive accuracy for high-dimensional integration challenges.
An orthant normal distribution is a classical but highly nontrivial object in the study of multivariate Gaussian measures: it refers to the computation of the probability that a multivariate normal vector falls within a specified orthant, most often the non-negative orthant, i.e., the set . The study of such probabilities is essential in multivariate probability, Gaussian process theory, polyhedral geometry, and has nontrivial applications in random polynomials on the simplex. For equicorrelated covariance structures, this problem admits specialized reductions and asymptotic analysis, while modern computation exploits connections to holonomic systems.
1. Formal Definition and Integral Representation
Let , where every is real-valued and is a symmetric positive-definite covariance matrix. The orthant probability for the non-negative orthant is
with multivariate density
For the equicorrelated case for and for , the orthant probability admits a one-dimensional reduction due to Steck (1962): where , , and is the univariate normal CDF (Pinasco et al., 2020). More generally,
with extensions allowing arbitrary mean and covariance (Koyama et al., 2012).
2. Analytical Bounds for the Equicorrelated Orthant Probability
For the equicorrelated normal, displays distinct asymptotic regimes depending on :
- For and : There exists a universal constant such that
with the Beta function.
- For and suitably large :
and for all ,
For all fixed , up to polylogarithmic factors in (Pinasco et al., 2020).
Steck’s reduction, Markov inequalities, and Mill’s ratio-type bounds underpin these estimates, with critical refinements available for the logarithmic correction regimes in .
3. Holonomic Gradient Method and Pfaffian Systems
Direct computation of orthant probabilities for arbitrary covariance matrices is numerically challenging. The holonomic gradient method (HGM) establishes a system of linear PDEs—the holonomic system—annihilating the multivariate integral
where parameters are and (Koyama et al., 2012). The corresponding orthant probability is then
The HGM reformulates the problem as a first-order Pfaffian ODE for the vector of components, with recurrence relations for derivatives in and . This reduction ensures that the numerical procedure is well-conditioned as long as the covariance remains positive definite.
The algorithm iteratively integrates from a diagonal start (zero mean, diagonal covariance) to the desired along a straight path, solving the ODE with a method such as fourth-order Runge–Kutta. The principal computational cost stems from the exponential state scaling, , but practical calculations are reported up to under moderate accuracy (Koyama et al., 2012).
4. Plackett Recurrence and Comparative Numerical Strategies
Classical approaches, such as Plackett's 1954 recurrence, relate derivatives of orthant probabilities to lower-dimensional analogs through explicit recurrence relations. However, the recursion involves denominators , introducing numerical instability for near-singular correlation matrices or as .
The holonomic gradient method, by contrast, eliminates such problematic factors and operates on rational recurrences over an exponentially-sized state space, yielding improved stability on high-dimensional or ill-conditioned problems. Comparative studies indicate HGM’s competitive or superior computational timing for large and improved accuracy benchmarks up to (Koyama et al., 2012).
5. Asymptotics and Applications in Random Polynomials on the Simplex
A substantial application concerns random homogeneous polynomials evaluated on the standard -simplex . Specifically, for a random Bombieri-normalized -homogeneous polynomial , the event that attains a (relative) maximum at a vertex is equivalent to the event that an associated equicorrelated multivariate normal has all positive entries. The relevant correlation parameter for vertices is
For , the events become almost independent in total variation as , and the probability that has a maximum at some vertex converges to $1$ (Pinasco et al., 2020).
6. Numerical Implementation and Performance
The HGM algorithm operates through an initial parameter transformation to , initialization on a diagonal covariance, ODE setup via block-sparse recurrences, and numerical integration along a chosen path. Key implementation items include:
- Sparse matrix optimization given that each recurrence couples subsets differing by at most two elements.
- Precomputation of recurrence sparsity patterns for efficiency.
- Monitoring of covariance positivity to stay within the holonomic domain.
Numerical experiments reported by Koyama & Takemura show high-precision agreement with exact results (up to ), competitive runtimes for small compared to Miwa–Hayter–Kuriki recursion, and clear acceleration for large (e.g., HGM is faster for ). Limitations can occur when is large and nearly singular, causing ODE stiffness (Koyama et al., 2012).
7. Connections to Broader Contexts
Orthant normal probabilities underpin various problems in statistical inference, Gaussian process excursion sets, and the study of random fields, as well as in polyhedral geometry and stochastic optimization. Their computation, both by analytic reduction in symmetric (equicorrelated) cases and by holonomic systems in the general setting, remains an active area with substantial algorithmic and theoretical developments (Pinasco et al., 2020, Koyama et al., 2012).