Papers
Topics
Authors
Recent
Search
2000 character limit reached

Probabilistic Inclusion Depth (PID)

Updated 24 December 2025
  • PID is a data depth measure that quantifies the centrality of fuzzy and binary contours using a probabilistic inclusion operator.
  • It computes inclusion scores via average measures, ensuring binary consistency, coordinate-agnosticism, and robustness to perturbations.
  • The PID-mean approximation enables linear-time, GPU-accelerated computation, making it scalable for high-dimensional ensemble visualizations.

Probabilistic Inclusion Depth (PID) is a data depth measure designed for centrality ordering and ensemble visualization of fuzzy contours—that is, soft masks output by modern segmentation models as well as conventional binary contours—across arbitrary spatial domains. PID generalizes previous contour depth concepts by employing a probabilistic inclusion operator that supports both fuzzy and crisp representations, accommodates non-uniform grids, and enables scalable computation for large and high-dimensional ensembles (Wu et al., 17 Dec 2025).

1. Probabilistic Inclusion Operator

Let Ω\Omega denote a spatial domain (e.g., a 2D image or 3D volume) equipped with a measure %%%%1%%%% (counting for discrete grids, Lebesgue for continuous). A fuzzy contour, or soft mask, is described by a function

u:Ω[0,1]u : \Omega \to [0,1]

such that u(x)u(x) indicates the probability or degree of membership of xx in the contour. The total mass is m(u)=Ωu(x)dμ(x)m(u) = \int_\Omega u(x)d\mu(x). Provided m(u)>0m(u)>0, uu induces a measure πu(E)=1m(u)Eu(x)dμ(x)\pi_u(E) = \frac{1}{m(u)} \int_E u(x)d\mu(x) for measurable EΩE\subseteq \Omega.

Given two fuzzy masks uu and vv, the probabilistic inclusion operator is

upv:=EXπu[v(X)]=Ωu(x)v(x)dμ(x)Ωu(x)dμ(x)=1Ωu(x)(1v(x))dμ(x)m(u).u \subset_p v := \mathbb{E}_{X\sim\pi_u}[v(X)] = \frac{\int_\Omega u(x)v(x)d\mu(x)}{\int_\Omega u(x)d\mu(x)} = 1 - \frac{\int_\Omega u(x)(1-v(x))d\mu(x)}{m(u)}.

For indicator functions u=1A,v=1Bu=1_A, v=1_B, this reduces to AB/A|A\cap B| / |A|, recovering the continuous-subset operator ϵ\subset_\epsilon used in exact Inclusion Depth (eID).

2. Mathematical Formulation of Probabilistic Inclusion Depth (PID)

Given an ensemble of NN fuzzy contours {u1,,uN}\{u_1, \ldots, u_N\}, PID assigns to each uiu_i a scalar depth quantifying its centrality: ri:=1Nj=1N(uipuj),si:=1Nj=1N(ujpui).r_i := \frac{1}{N} \sum_{j=1}^N (u_i \subset_p u_j), \qquad s_i := \frac{1}{N} \sum_{j=1}^N (u_j \subset_p u_i). Here, rir_i measures the average degree to which uiu_i is contained in the rest, and sis_i measures how well uiu_i contains others. PID is then defined as

PID(ui):=min{ri,si}.\mathrm{PID}(u_i) := \min\{r_i, s_i\}.

This yields a robust, center-outward ordering. In expanded form: ri=1Nj=1Nuiujdμm(ui),si=1Nj=1Nuiujdμm(uj),PID(ui)=min{ri,si}.r_i = \frac{1}{N} \sum_{j=1}^N \frac{\int u_i u_j d\mu}{m(u_i)}, \qquad s_i = \frac{1}{N} \sum_{j=1}^N \frac{\int u_i u_j d\mu}{m(u_j)}, \qquad \mathrm{PID}(u_i) = \min\{r_i,s_i\}.

3. Theoretical Properties

PID and the p\subset_p operator exhibit several important properties:

  • Binary consistency: For masks ui{0,1}u_i\in\{0,1\}, PID coincides exactly with eID.
  • Monotonicity in vv: upvu\subset_p v is monotonic in vv.
  • Linearity in vv: up(av1+bv2)=a(upv1)+b(upv2)u\subset_p(av_1+bv_2)=a(u\subset_p v_1)+b(u\subset_p v_2).
  • Scale-invariance in uu: Multiplying uu by a positive scalar does not affect upvu\subset_p v.
  • Asymmetry: upvvpuu\subset_p v\neq v\subset_p u in general.
  • Lipschitz continuity: p\subset_p and PID are Lipschitz continuous in the L1L^1 norm, providing robustness to mask perturbations.
  • Coordinate-agnostic: PID depends solely on μ\mu and is invariant under measure-preserving transformations of Ω\Omega, supporting uniform/non-uniform grids and manifolds.

4. Efficient Computation and PID-mean Approximation

Exact PID (Quadratic Complexity)

For NN ensemble members of spatial size MM, full PID evaluation is O(N2M)O(N^2M):

  • For each pair (i,j)(i,j), compute xΩwxui(x)uj(x)\sum_{x\in\Omega} w_x u_i(x)u_j(x), along with denominators for uiu_i and uju_j.
  • PIDi\mathrm{PID}_i is obtained via min{ri,si}\min\{r_i, s_i\} as described above.

PID-mean (Linear Complexity)

The ensemble mean mask uˉ(x)=1Niui(x)\bar u(x) = \frac{1}{N}\sum_i u_i(x) enables a linear-time surrogate: rimean=uipuˉ=uiuˉdμm(ui),simean=uˉpui=uiuˉdμm(uˉ)r_i^{\text{mean}} = u_i \subset_p \bar u = \frac{\int u_i\bar u d\mu}{m(u_i)}, \qquad s_i^{\text{mean}} = \bar u \subset_p u_i = \frac{\int u_i\bar u d\mu}{m(\bar u)}

PID-meani=min{rimean,simean}\mathrm{PID}\text{-mean}_i = \min\{r_i^{\text{mean}}, s_i^{\text{mean}}\}

This involves O(NM)O(NM) operations since uˉ\bar u is precomputed. The approximation error PID(ui)PID-mean(ui)|\mathrm{PID}(u_i)-\mathrm{PID}\text{-mean}(u_i)| is bounded by the coefficient of variation of {m(uj)}\{m(u_j)\}; if mask masses are homogeneous, PID-mean closely approximates the exact PID.

5. GPU Implementation and Computational Scalability

PID-mean's linear structure allows efficient GPU parallelization:

  • For each uiu_i, a CUDA thread block computes numerators and denominators over Ω\Omega, exploiting memory coalescence for maximal bandwidth.
  • Within each block, voxels are processed in parallel, accumulating partial sums via shared-memory tree reduction; totals are transferred to the host for final computation of rimean,simeanr_i^{\text{mean}}, s_i^{\text{mean}}.

Empirical performance demonstrates:

Dataset eID (CPU) PID-mean (CPU) PID-mean (GPU)
3D synthetic, N=10N=10k–20k, 50350^3 vox 50–250 s 4–11 s 1.2–2.0 s (up to 125× eID)
3D, N=50N=50, 1283128^35123512^3 vox 22–216 s 14–29 s 2.3–5.5 s (up to 39× eID)

For large N>20,000N>20,000 or high resolutions, eID becomes intractable in memory, while PID-mean (especially on GPU) remains scalable (Wu et al., 17 Dec 2025).

6. Fuzzy Isovalue Modeling and Sensitivity Encoding

PID supports a probabilistic treatment of isovalue selection in scalar field ensembles by integrating over an isovalue distribution. For each scalar field Fi(x)F_i(x), instead of extracting a sharp isocontour at qq, define QQ with density pQ(t)p_Q(t) and set: ui(x)=P(Fi(x)<Q)=t:Fi(x)<tpQ(t)dt.u_i(x) = \mathbb{P}(F_i(x) < Q) = \int_{t: F_i(x)<t} p_Q(t)dt. This results in a fuzzy, "soft" isocontour capturing how small changes in qq affect membership, encoding isolevel sensitivity. Such modeling is comparable to interval volumes and probabilistic marching cubes but specifically adapts to depth-based ensemble visualization via PID. These fuzzy masks directly serve as PID/PID-mean inputs.

7. Experimental Results and Practical Applications

Rank Stability and Consistency

PID yields significantly more stable ensemble rankings than previous methods under isovalue shifts and member removal:

  • For weather-forecast ensembles (N=20), PID achieves a rank-rank Pearson correlation of r=0.965r=0.965 across isolevels q=5620q=5620 m and q=5640q=5640 m, outperforming eID (r=0.780r=0.780).
  • On synthetic contour sets (N=200N=200), PID-mean agrees strongly with eID (r=0.987r=0.987) and Prob-IoU (r=0.920r=0.920); CBD and ISM show weaker consistency.

Upon removing extreme outliers in 3D ellipsoid data, PID maintains high rank correlation (Pearson=0.997\text{Pearson}=0.997, τ=0.964\tau=0.964), exceeding Prob-IoU.

Application Domains

  • Medical segmentation: PID-mean ranks outputs from 31 SegResNet models (224×224×144) in the MSD Brain-Tumour dataset, producing 3D ensemble boxplots that delineate tumour shape uncertainty.
  • Large-scale binary masks: On IXI data (400 hippocampus, 400 ventricle masks), PID-mean extracts central envelopes and medians without the overplotting typical for spaghetti plots.
  • Scalar field ensembles: PID-mean applied to 50×150 ScalarFlow smoke plume reconstructions tracks uncertainty evolution in 3D time-series boxplots.
  • Flexible grids/manifolds: Weight factors wxw_x enable PID-mean on uniform/non-uniform grids, meshes, or manifolds; the method is coordinate-agnostic.

Scalability extends to tens of thousands of ensemble members and 3D grids of up to 5123512^3 voxels using GPU PID-mean. For spatiotemporal ensembles, time is incorporated as an extra spatial axis, enabling 4D centrality rankings (Wu et al., 17 Dec 2025).


PID generalizes contour-depth analysis to fuzzy masks by substituting set inclusion with an expectation under the reference mask, retaining core theoretical guarantees (monotonicity, stability, coordinate-agnosticism, binary specialization) and enabling scalable, high-resolution ensemble visualization via linear- and GPU-accelerated algorithms. Demonstrated across synthetic and real applications in meteorology, medical imaging, and volumetric data, PID provides robust rankings, increased stability to isovalue fluctuations, and substantial computational acceleration relative to prior approaches (Wu et al., 17 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Probabilistic Inclusion Depth (PID).