Papers
Topics
Authors
Recent
Search
2000 character limit reached

Parallelized Hierarchical Connectome: A Spatiotemporal Recurrent Framework for Spiking State-Space Models

Published 1 Apr 2026 in q-bio.NC | (2604.01295v1)

Abstract: This work presents the Parallelized Hierarchical Connectome (PHC), a general framework that upgrades temporal-only State-Space Models (SSMs) into spatiotemporal recurrent networks. Conventional SSMs achieve high-speed sequence processing through parallel scans, yet are limited to temporal recurrence without lateral or feedback interactions within a single timestep. PHC maps the diagonal SSM core to a shared Neuron Layer and inter-neuronal communication to a shared Synapse Layer, where neurons are partitioned into hierarchical regions governed by the connectome topology. A Multi-Transmission Loop enables intra-slice spatial recurrence, allowing signals to propagate across the hierarchical connectome within each temporal window while preserving O(logT) parallelism. This framework enables integration of neuro-physical priors typically intractable for standard SSMs, including adaptive leaky integrate-and-fire dynamics, Dale's Law, short-term plasticity, and reward-modulated spike-timing-dependent plasticity. The framework is instantiated as PHCSSM, the first model to unify recurrent spiking neural network dynamics with diagonal SSM parallelism while enforcing all five biological constraints and learnable lateral connections within a fully parallelizable training pipeline. Empirical results on physiological benchmarks from the UEA multivariate time-series archive demonstrate that PHCSSM achieves performance competitive with state-of-the-art SSMs while reducing parameter complexity from Theta(D2 L) for L-layer stacked architectures to Theta(D2). These findings suggest that biologically grounded inductive biases offer a principled route to parameter-efficient sequence modeling, opening diagonal SSMs to spatiotemporal recurrence and enabling fully parallelizable recurrent spiking neural network training.

Authors (1)

Summary

  • The paper introduces a novel framework that decouples spatial and temporal dynamics, enabling fully parallelizable spiking state-space models with biologically inspired constraints.
  • It partitions state-space dynamics into orthogonal neuron and synapse layers using a Multi-Transmission Loop to emulate realistic neural microcircuits.
  • Empirical results on benchmark datasets demonstrate superior accuracy and parameter efficiency, highlighting the benefits of integrating biological regularization.

Parallelized Hierarchical Connectome: A Spatiotemporal Recurrent Framework for Spiking State-Space Models

Introduction and Motivation

The Parallelized Hierarchical Connectome (PHC) framework introduces a fundamentally novel approach to recurrent sequence modeling by reconciling two previously irreconcilable properties: spatiotemporal recurrence with lateral feedback and O(log T) temporal parallelizability. Conventional State-Space Models (SSMs), including S4, S5, LRU, and Mamba, leverage diagonal state transition structures to enable parallel associative scans; this design sacrifices intra-step spatial connectivity, limiting sequence models to purely temporal recurrence. Biologically inspired Spiking Neural Networks (SNNs) and RNNs support complex lateral and feedback communication but are non-parallelizable due to their dense recursive dependencies.

The PHC framework departs from both paradigms. It explicitly partitions state-space dynamics into orthogonal temporal and spatial components, enabling the integration of biologically grounded constraints such as Dale's Law, adaptive spiking (ALIF), short-term plasticity (STP), strict connectome topology, and reward-modulated spike-timing-dependent plasticity (R-STDP) within a scalable, parallelizable architecture suitable for mainstream sequence modeling benchmarks.

Architectural Framework: Spatiotemporal Decoupling and the Multi-Transmission Loop

At the core of PHC is the Intra-Step Spatiotemporal Decoupling principle, which separates conventional state-space evolution into a strictly diagonal, cell-autonomous Neuron Layer (NL) and a sparse, structured Synapse Layer (SL). Whereas conventional SSMs replicate layerwise depth (L) via stacking independent blocks and MLPs (Figure 1A), PHC collapses this into a hierarchical, region-partitioned neuron population where all communication—feedforward, lateral, and feedback—takes place via a shared, biologically constrained connectome. Figure 1

Figure 1: Structural comparison of conventional stacked SSMs (A) with PHC (B), highlighting lateral/feedback connectivity and hierarchical neuron partitioning.

Signals propagate within each temporal window by circulating multiple times through the NL-SL pipeline—a Multi-Transmission Loop—allowing recurrent spatial aggregation along the connectome before proceeding to the next input token. This recursion emulates biological microcircuits, where information routes through a fixed physical connectome, in contrast to the deep stacking paradigm. Figure 2

Figure 2: Isomorphism of standard stacked SSMs and the PHC architecture. Layer stacking is replaced by region partitioning, MLPs by connectome sub-blocks with Dale's Law and topology constraints.

This design supports fully parallelized training and inference via log-domain prefix scans for all dynamical components, including adaptive thresholding, membrane dynamics, and STP.

PHCSSM: A Biophysically Constrained Spiking State-Space Model

The instantiation of the PHC framework, PHCSSM, realizes the above architectural principles in a fully spiking SSM that enforces all of the following constraints:

  • ALIF neural dynamics: Intrinsic adaptation is realized via learnable, elementwise leaky integration and adaptive spike thresholds using sequential parallel-scan recurrences.
  • Short-term plasticity (STP): Implemented using the Tsodyks–Markram model, supporting both facilitation and synaptic depression through additional affine recurrences.
  • Dale's Law and E/I partitioning: Synaptic weights are sign-constrained, segregating excitatory and inhibitory populations, and further pruned by a region-wise topology mask.
  • Hierarchical Connectome Topology: Each neuron is assigned to a region, and the sparsity/connectivity is controlled at the block level, with explicit feedforward and feedback projections.
  • Online R-STDP: Genuine binary spikes enable native spike-timing-dependent plasticity modulated by a scalar reward signal derived from downstream loss/accuracy, operating external to gradient descent. Figure 3

    Figure 3: Detailed PHCSSM signal flow in the forward pass, showing encoding, masked sensory injection, iterative ALIF and STP scans, and biologically constrained matrix transmission.

The full NL and SL computation is executed in parallel across the temporal axis, with spatial interaction depth decoupled from sequence length.

Empirical Evaluation

PHCSSM is evaluated on the six-dataset UEA Multivariate Time Series Classification Archive under standard protocols. Key results include:

  • SCP2: PHCSSM achieves 59.3% mean test accuracy, surpassing the previous best SSM, LinOSS-IMEX (58.9%), and exceeding all baseline models, including Mamba and S5.
  • MotorImagery: Achieves 53.7%, outperforming Mamba (+6.0pp), S6, and Transformer.
  • EigenWorms: Attains 83.9% accuracy with only 2,701 parameters, surpassing five strong SSM baselines.

Across all datasets, PHCSSM matches or exceeds several unconstrained models while requiring one to two orders of magnitude fewer parameters (1,748–9,485 vs. tens/hundreds of thousands), confirming the parameter efficiency conferred by weight sharing and biological regularization.

Ablation experiments substantiate the non-redundant contribution of each biological constraint. Disabling ALIF, STP, Dale's Law, or R-STDP each degrades accuracy (e.g., –12.3pp on SCP2 for ALIF removal), and removal of Dale's Law/online learning increases training variance, demonstrating strong regularization.

Theoretical and Practical Implications

The PHC paradigm demonstrates that biological constraints are not a performance bottleneck for parallel SSMs and, in fact, act as stabilizing inductive biases, enhancing both optimization regularity and parameter efficiency. By decoupling spatial recurrence from temporal propagation, PHCSSM introduces a new architectural axis—transmission depth M—enabling deep logical processing over fixed parameter budgets without layer stacking.

Pragmatically, these design principles permit:

  • Integration of realistic microcircuit phenomena (STP, Dale's Law) in efficient GPU-compatible sequence models.
  • Seamless transition from parallel training to sequential (RSNN) simulation without post-hoc conversion.
  • Incorporation of empirical connectome data (mouse/human brain atlases) for functional digital twins.
  • Natively supporting reward-modulated plasticity unavailable to continuous models, which may facilitate lifelong/online learning.

This framework directly supports applications in physiological signal decoding, neuroscience simulation, and energy-efficient AI via neuromorphic deployment.

Limitations and Future Directions

Current evaluation is restricted to sequence classification and region sizes up to R=2 (D=16-64). Future work should address:

  • Scaling PHC to larger connectomes and more complex input-output tasks (e.g., language, speech, multivariate regression).
  • Automatic connectome/region structure discovery from biological or task-specific priors.
  • Quantitative analysis of parallel-to-sequential mapping error to support faithful neuromorphic implementation.

Incorporating additional biological mechanisms such as chemotactic modulation, dendritic nonlinearity, or synaptic homeostasis is also viable within this framework.

Conclusion

The PHC framework and its PHCSSM instantiation establish that biologically principled architecture, explicit spatiotemporal recurrence, and full temporal parallelizability are simultaneously achievable in SSMs (2604.01295). The resultant models offer state-of-the-art accuracy on benchmarks while maintaining extreme parameter parsimony and full support for structured, interpretable connectivity and online plasticity. This work uncovers a systematic route for leveraging biological priors as architectural design principles in mainstream deep learning.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 2 likes about this paper.