History-Conditioned Local Projector
- History-conditioned local projector is a data-driven operator that infers fine-grained transition statistics in observed Markov trajectories while accounting for hidden memory.
- The methodology empirically constructs and normalizes history-conditioned transition histograms, then uses spectral analysis of eigenvalues to quantify memory timescales.
- This approach enables the detection of deviations from the Markov property and guides state splitting to better represent hidden dynamics in complex systems.
A history-conditioned local projector is a data-driven operator on observed discrete-time Markov trajectories, which infers fine-grained transition statistics between observed (coarse-grained) states while accounting for the memory effects induced by hidden or unobservable degrees of freedom. Originally introduced in the context of Markov-state holography, this analysis allows one to empirically assess local deviations from the Markov property, detect evidence of hidden-state topology, and quantify the timescale of memory due to hidden-path decorrelations from projected high-dimensional dynamics (Zhao et al., 14 Mar 2025).
1. Definitions and Notational Framework
Let the full (microscopic) state space be , where is the space of directly observed (“lumped”) states, and denotes the set of microstates projected onto observed state : .
A length- observation history at time is . For each observed transition at times , and immediately preceding history , define
- : number of observed transitions with history ,
- ,
- Empirical history-conditioned transition probability:
These transition probabilities form the entries of the history-conditioned local projector.
2. Mathematical Construction
For each ordered transition and history , the local projector-matrix entry is defined as
Alternatively, in model-based terms (where are microstates reached upon entering ):
Here is a path-sum over hidden trajectories compatible with .
Given the microscopic splitting-probability matrix and the sub-block (within ), the probability for the first exit from is:
and the history-projected transition probability becomes:
Fixing , the matrix with entries can be analyzed spectrally. The eigenvalues satisfy , .
3. Data-Driven Estimation and Algorithm
Empirical construction proceeds by traversing a time series :
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
Input: trajectory s[1..T], max history K_max
Initialize: N[j][h_k][i] = 0 for all j, h_k, i
Loop:
for t = K_max..T-1:
for k=0..K_max:
h_k = (s[t-k+1], ..., s[t])
j = s[t]
i = s[t+1]
N[j][h_k][i] += 1
N_j[j][h_k] += 1
Compute probabilities:
for each j, h_k:
if N_j[j][h_k]>0:
for each i:
P_{i←j|h_k} = N[j][h_k][i] / N_j[j][h_k] |
Normalize each slice so that . The maximal is increased until transition histograms converge, measured by the total-variation distance
declaring convergence when (e.g., ).
4. Memory Quantification and Spectral Analysis
The subleading eigenvalue of determines the local memory timescale:
As increases, and , characterizing the number of steps over which hidden-path correlations decay. Large (i.e., long ) indicates slow decay and persistent hidden-path memory, signifying insufficient Markovianity at the observed level.
5. Canonical Example: Three-State System with Hidden Path
Consider observed states , , , where is actually two hidden microstates . The transition scheme is , , , , . The microscopic splitting probabilities are , , to each with $1/2$, .
For (no history), . For (history of one step), there are two histories, (from ) and (from ):
where are arrival fractions at from or . Thus, is a diagonal matrix with , and eigenvalues $1$, . The timescale is then .
As , path weights approach their stationary distribution, , and .
6. Testing Markov Property and Inferring Hidden Structure
The local history-conditioned projector provides a direct test for hidden memory. If, for state , becomes independent of for all , then is Markov of order locally; any persistent -dependence reflects hidden-path memory or missing microstates.
If shows multiple bars or more than one significant eigenvalue, one proposes splitting into , assigning rates so that match these bars. Recomputing with the relump validates the split if Markovianity improves.
7. Practical Considerations and Limitations
Building all histograms up to maximal history has computational complexity and storage cost in the worst case. Reliable estimation of generally requires per history; since history classes grow as , must be kept modest or histories coarsened by binning. Severe undersampling arises exponentially with . Convergence proofs require at least one observable Markov state; application to continuous observables depends on bin width choices.
For detailed derivations and further implementation guidance, see (Zhao et al., 14 Mar 2025).