Papers
Topics
Authors
Recent
Search
2000 character limit reached

Dense Random Survival Forests

Updated 7 January 2026
  • The paper introduces a dense ensemble framework that grows thousands of survival trees across diverse hyperparameter settings to capture treatment–covariate interactions.
  • It employs a novel splitting rule optimized for detecting heterogeneity in treatment effects within censored time-to-event data, ensuring high sensitivity and controlled Type I error.
  • Spectral clustering on the aggregated proximity matrix facilitates unsupervised discovery of clinically meaningful patient subgroups in both simulated and clinical trial datasets.

Very Dense Random Survival Forests (DRSF) are a methodological framework for unsupervised identification of patient subgroups with heterogeneous treatment response in time-to-event data, characterized by an exceptionally large ensemble of survival trees (up to 100,000) each grown under diverse hyperparameter settings. DRSF is distinguished by its targeted splitting criterion that directly optimizes for treatment–covariate interaction in the presence of censored data, and by the explicit fusion of proximity information across hyperparameter sweeps, culminating in a robust similarity-driven clustering of patients without requiring pre-defined subgroup labels. The technique achieves high sensitivity and stringent Type I error control for heterogeneity detection, with demonstrated interpretability and calibration on both simulated and clinical trial datasets (Li et al., 4 Jan 2026).

1. Algorithmic Structure and Dense Ensemble Construction

DRSF departs from conventional random survival forests by constructing a dense ensemble: rather than training a single forest, it trains forests across a wide grid of hyperparameter configurations including mtry (number of covariates sampled at each split), node size, tree depth, split-search budget (nsplit), and per-variable weighting (xvar.wt). In simulation experiments, each configuration yields 1,500 trees; in clinical datasets, 500 trees per configuration are typical. Combining 60–200 such settings produces a total ensemble containing approximately 50,000–100,000 trees.

All trees are grown using bootstrap samples in the Breiman paradigm but employ a novel splitting rule focused on uncovering treatment–covariate interactions. Once E trees are constructed, a proximity matrix SS is obtained by aggregating the co-occurrence of patients ii and jj in terminal nodes across all trees:

p(i,j)=1Et=1EI(Xi and Xj fall in the same terminal node of tree t).p(i, j) = \frac{1}{E}\sum_{t=1}^E I\left( X_i \text{ and } X_j \text{ fall in the same terminal node of tree } t \right).

Fusing these proximities for all (i,j)(i, j) produces an n×nn \times n similarity matrix used for downstream clustering.

2. Unsupervised Subgroup Discovery via Spectral Clustering

Although each individual tree models survival outcomes with respect to treatment and covariate interactions, no patient subgroup labels are provided to the learner. The methodology is unsupervised in that the subgroups are discovered post hoc from the geometry of the proximity matrix.

Spectral clustering is employed to partition the nn patients into KK subgroups, where KK ranges from 2 to 7. The proximity matrix SS is treated as a weighted adjacency matrix for spectral graph partitioning. This approach allows for identifiably distinct, interpretable patient clusters associated with distinct treatment effects, without reliance on arbitrary or a priori subgroup definitions.

3. Treatment-Heterogeneity–Targeted Splitting Rule

Unlike traditional random survival forests which use log-rank or prognostic score statistics for splitting, DRSF introduces a criterion that directly targets treatment-by-covariate interaction heterogeneity. At each node, given nsn_s observations (Ti,δi,Xi,Wi)(T_i, \delta_i, X_i, W_i), where Wi{0,1}W_i \in \{0, 1\} encodes treatment, a candidate 1D split ss defines child assignments Vi(s){0,1}V_i(s) \in \{0, 1\}. The following Cox model is fit:

h(tVi,Wi)=h0(t)exp{β1Vi+β2Wi+β3(Vi×Wi)}h(t \mid V_i, W_i) = h_0(t)\exp\{ \beta_1 V_i + \beta_2 W_i + \beta_3 (V_i \times W_i) \}

Split scoring is based on a weighted sum:

G(s)=α1a1(s)0.52+(1α1)a2(s)α2G(s) = \alpha_1 \frac{a_1(s) - 0.5}{2} + (1 - \alpha_1) \frac{a_2(s)}{\alpha_2}

where a1(s)a_1(s) is the C-index of the model, a2(s)a_2(s) is the Z-score for testing H0:β3=0H_0: \beta_3 = 0, α1\alpha_1 weights the trade-off between prognostic purity and treatment interaction, and α2\alpha_2 scales the Z-score. Varying α1\alpha_1 interpolates between standard split rules (α1=1\alpha_1=1) and pure interaction focus (α1=0\alpha_1=0). Standard survival forests cannot discover subgroups based on response heterogeneity as they lack any explicit term for β3\beta_3.

4. Computational and Optimization Strategies

Efficiency and scalability in DRSF are achieved through several mechanisms:

  • At each split, only mtrymtry covariates are randomly sampled for consideration.
  • For each covariate, at most nsplitnsplit uniformly random cut-points are evaluated, drastically reducing search space relative to considering all ns1n_s - 1 possible splits.
  • Boundaries on tree depth and leaf size (minimum samples per terminal node) are imposed.
  • Each parameter configuration is trained independently, supporting massive parallelism both across and within configurations.
  • The cost of growing a single tree is O(nsmtrynsplitD)O(n_s \cdot mtry \cdot nsplit \cdot D), with DD as the maximum depth; cost for the entire ensemble is O(Etotaln(mtrynsplitD))O(E_{total}\cdot n \cdot (mtry \cdot nsplit \cdot D)).

High-performing C++ implementations make ensembles at this scale numerically and computationally feasible.

5. Evaluation Metrics and Statistical Calibration

Detection of meaningful treatment heterogeneity is assessed by fitting two Cox models post-clustering:

  • Baseline: 1:h(tW)=h0(t)exp{γW}\ell_1: h(t\mid W) = h_0(t)\exp\{\gamma W\}
  • Subgroup-specific: 2:h(tW,leaf)=h0(t)exp{γ1W+γ2leaf+γ3(leaf×W)}\ell_2: h(t\mid W, \text{leaf}) = h_0(t)\exp\{\gamma_1 W + \gamma_2 \text{leaf} + \gamma_3 (\text{leaf} \times W)\}

The log-likelihood ratio statistic

Λ=2(L1L2)χ2(K1)\Lambda = -2(L_1 - L_2) \sim \chi^2(K-1)

tests the null hypothesis of no subgroup-by-treatment interaction, with pleafp_{leaf} providing an empirical measure of heterogeneity.

To control Type I error at 1%, the significance threshold pp^* is empirically calibrated as the 1st percentile from pleafp_{leaf} values across 1,000 homogeneous (null) simulated datasets, ensuring that P(pleaf<p)1%P(p_{leaf} < p^*) \approx 1\% when no heterogeneity exists.

Power to detect heterogeneity (pleaf<pp_{leaf} < p^*) exceeded 90% in heterogeneous scenarios, with near-exact Type I error under global or null scenarios.

6. Application to Randomized Clinical Trial Data

DRSF was validated on Phase III randomized trials of Panitumumab (studies 263 and 309), analyzing endpoints including progression-free survival (PFS) and overall survival (OS). Spectral clustering of the DRSF-derived similarity matrix revealed two or three clinically interpretable subgroups, distinguished chiefly by KRAS mutation status and ECOG performance status.

For example, in study 263 (PFS endpoint), the subgroups were:

  • KRAS WT (n=455): HR=0.80, p=0.03p=0.03
  • KRAS mutant (n=599): HR ≈ 1.0, p0.05p \gg 0.05

Further subdivision by ECOG status in OS analysis refined identification of patients with maximal benefit. These results replicated known oncology findings (KRAS wild-type response to EGFR inhibitors; improved outcomes with ECOG=0) and satisfied the calibrated error control.

7. Hyperparameter Sweeping and Stability via Fusion

Key forest hyperparameters included mtry{2,3}mtry \in \{2,3\} (case studies), {6,7}\{6,7\} (simulations), node size {50,70,100}\in \{50,70,100\}, maximum depth {2,3,4}\in \{2,3,4\}, nsplit{0,20,50}nsplit \in \{0,20,50\}, treatment-split weight α1{0,0.1,,0.5}\alpha_1 \in \{0, 0.1, \ldots, 0.5\}, variable weighting schemes xvar.wtxvar.wt, and ntree=500ntree=500 or $1,500$. Instead of selecting a single "best" configuration, all are included and fused at the proximity matrix level, stabilizing the similarity measure and rendering subgroup identification robust to hyperparameter mis-specification. Deeper trees or larger nsplitnsplit enhance heterogeneity detection at the cost of computation, but the fusion approach mitigates oversensitivity to any one setting (Li et al., 4 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Very Dense Random Survival Forests.