Spatial Continuous Feature Extractor (SCFE)
- SCFE is a neural module that creates context-rich, spatially coherent features for volumetric and surface data, ensuring continuity across geometric structures.
- It converts 3D feature tensors into 1D sequences via resolution-aware flattening, enabling efficient long-range dependency modeling in segmentation tasks.
- SCFE’s adaptable design—including state-space blocks and field convolution—improves segmentation metrics and intrinsic surface learning across diverse datasets.
A Spatial Continuous Feature Extractor (SCFE) is a neural module designed to produce context-rich, spatially-coherent feature embeddings for high-dimensional geometric or volumetric data. SCFE architectures emphasize continuity: every location in the input domain (voxel grid or surface manifold) receives a descriptor encoding both local and global structure, while preserving geometric relationships. SCFE instances appear in volumetric segmentation frameworks, such as NeuroMamba’s global context module, and in intrinsic surface learning via field convolution. This article details the principles, architectural design, empirical results, and implementation specifics of representative SCFE mechanisms in both 3D volume and surface settings.
1. SCFE in Volumetric Segmentation Architecture
In the context of neuron segmentation using 3D electron microscopy (EM) data, SCFE is implemented within the NeuroMamba framework (Jiang et al., 22 Jan 2026). Its main function is to enable long-range voxel dependency modeling—essential for reconstructing neuronal arbors—without decomposing the volume into disconnected patches, a process that often disrupts continuity. Within the Multi-Perspective Feature Interaction (MPFI) block, the SCFE is paired with the Boundary Discriminative Feature Extractor (BDFE) for a dual approach: BDFE sharpens membrane boundaries and local details, while SCFE ensures global structural coherence.
2. Architectural and Computational Design
Within NeuroMamba, SCFE receives a 3D feature tensor and performs resolution-aware flattening to convert the block into four distinct 1D sequences:
- Transverse-First Scans: flattened in and order
- Axial-First Scans: flattened in and order
Each sequence is fed into a Visual-Mamba block, a state-space-model-based layer with linear time complexity. The outputs are linearly combined using data-driven weights (transverse scans) and (axial scans):
Weights are set as , with , and the anisotropy degree (ratio of axial to transverse voxel spacing). After summation, the 1D embedding is reshaped back to . This workflow preserves voxel-to-voxel relationships across the entire volume.
3. Resolution-Aware Scanning and Adaptivity
SCFE’s scanning strategy directly addresses the anisotropy in EM imaging (e.g., -spacing -spacing). By computing and adjusting , the module emphasizes the lower-resolution axis when necessary. For instance, in highly anisotropic datasets (AC3/AC4: nm, nm), SCFE shifts weighting toward the axial scans. In isotropic datasets (FIB25), weights are evenly split. This mechanism allows SCFE to flexibly adapt to arbitrary voxel resolutions, yielding robust performance across diverse datasets.
4. Integration with State-Space Models and Network Fusion
The 1D sequences produced by SCFE are passed through shared Visual-Mamba layers. These layers employ selective scan, state-space filtering, and a small feed-forward network with residual connections. This design enables patch-free global modeling. In the full MPFI architecture, SCFE’s global features () are fused with BDFE’s local features () within the Cross Feature Interaction (CFI) fusion stage; both feature sets are computed in parallel.
5. SCFE on Surfaces: Intrinsic Feature Extraction via Field Convolution
On curved surfaces, SCFE functionality is realized via field convolution (Mitchel et al., 2021). The operator produces continuous feature descriptors at every point of a 2D Riemannian manifold . Rather than gathering neighbor features into one local frame, every neighbor uses its own tangent-plane chart to “scatter” information towards . The process involves parallel transport, filter weighting using geodesic coordinates, and summation over the local geodesic ball . The field convolution is defined by:
This approach—intrinsic and isometry-invariant—can be discretized on triangle meshes for practical implementation, using band-limited filter parameterizations and complex-valued descriptors.
6. Empirical Results and Hyperparameter Choices
Empirical testing of SCFE in NeuroMamba demonstrates quantitative gains in segmentation quality. Ablation studies indicate:
- Adding SCFE to a baseline reduces Variation of Information (VI) from and Adjusted Rand Index (ARAND) from
- Combined SCFE+BDFE yields VI $0.865$ and ARAND $0.142$
- Including anisotropy priors () reduces ARAND by $0.002 - 0.004$ on several datasets
Recommended hyperparameters are ; typical block shapes are , while Mamba channels are . Computational overhead remains linear in ; full NeuroMamba (including SCFE) reports M parameters, $220.8$G FLOPS, and $0.059$s latency on a V100.
For field convolution, recommended parameters include support radius (segmentation), radial bins –$12$, and band-limit –$2$. Implementation leverages area-normalized weights, complex ReLU, and backpropagation through sum-based convolutions. Downsampling and vector heat-based geodesic computations improve scalability on large meshes.
7. Practical Considerations and Reproducibility
SCFE modules avoid patch-based tokenization, relying instead on global flatten+reshape per block. Memory usage scales predictably with and may be controlled via block size. Tiling with overlap, Adam optimizers (), and accurate axis calibration () are recommended for reproducibility (Jiang et al., 22 Jan 2026). For surfaces, efficient neighborhood search and tangent vector lifting are critical (Mitchel et al., 2021). The patch-free design and intrinsic convolutional structure ensure robust continuity under various geometric and resolution scenarios.
Table: SCFE Mechanisms in Volume vs. Surface Settings
| Domain | SCFE Mechanism | Continuity Model |
|---|---|---|
| 3D EM Volumes | Resolution-aware Mamba | Voxel global context |
| 2D Surfaces | Field Convolution | Geodesic-intrinsic |
SCFE modules are central to state-of-the-art architectures in both volumetric neuron segmentation and geometric surface learning, advancing continuity and adaptivity in neural feature modeling.