Landmarking Supermodels via Gradient Boosted Trees
- The paper’s main contribution is the integration of landmark supermodels with gradient boosted trees for dynamic survival prediction through a unified nonparametric framework.
- This approach relaxes parametric and Markovian assumptions by pooling data from multiple landmark points to estimate a conditional hazard surface via Poisson regression.
- Empirical evaluations show that the boosted-tree supermodel outperforms traditional Cox models in nonlinear and high-dimensional settings, ensuring temporal coherence and scalability.
Landmarking supermodels combined with gradient boosted trees provide a nonparametric methodology for dynamic prediction in event history analysis with high-dimensional, time-dependent covariates. This approach directly estimates the future conditional hazard surface via a unified model, relaxing parametric and Markovian assumptions. The estimator leverages gradient boosting on regression trees to capture nonlinearities and high-order interactions, while the statistical estimation procedure is formalized as a sieve M-estimation problem, and is computationally reducible to Poisson regression. This framework is distinguished by its theoretical consistency, flexibility to high-dimensional covariate structures, and avoidance of temporal incoherence common in traditional landmark Cox model approaches (Sandqvist, 24 Jan 2026).
1. Landmarking Supermodels: Definition and Structure
Given independent subjects observed over the interval , each subject is characterized by:
- : counting process for the event of interest,
- : at-risk indicator,
- : a high-dimensional, time-dependent covariate process.
For a fixed landmark time , the future conditional hazard at time , given covariate history up to , is
where .
Traditional landmark analyses fit separate models at fixed landmark times . The landmarking supermodel instead fits a single surface to approximate , pooling all landmarked data jointly for improved statistical efficiency and coherence.
2. Mathematical Formulation and Sieve M-Estimation
The empirical criterion for a candidate log-hazard surface , at a landmark , for subject is: Random landmarks are introduced, independent and uniformly distributed on . The per-subject empirical functional is
The estimation problem is intractable over the class of all measurable bounded functions, so a sequence of sieves is defined; in practice, these are linear spans of regression trees. The estimator is given by
3. Reduction to Poisson Regression and Integration of Gradient Boosted Trees
On each sieve , is piecewise constant on a partition of , i.e., on cell . For landmark ,
Then,
which is the Poisson log-likelihood of counts with exposure . Consequently, the optimization can be performed using standard Poisson-loss gradient boosting, fitting trees to pseudo-residuals at each boosting step. Model training thus accommodates high-dimensional and nonlinear feature effects naturally.
4. Implementation, Computational Details, and Tuning
The approach is implemented as follows:
- Data construction: For each subject and simulated landmark, compute and for all partition cells, yielding a "long" dataset with offsets .
- Boosting algorithm:
- Initialize .
- For : compute negative gradients, fit regression trees (of depth ) to these residuals, update , and truncate to .
- Selection of learning rate (), number of trees (), and tree depth () is by cross-validation on Poisson loss.
- Scalability: The dataset has size on the order of rows per boosting iteration; boosting iteration complexity is using histogram-based splitting or for exact splits. Merging cells sharing coefficients reduces memory. Tree-based methods maintain efficiency as feature dimensionality grows, due to internal variable selection mechanisms.
- Software compatibility: Methodology admits direct implementation in off-the-shelf gradient boosting libraries (e.g., XGBoost, LightGBM) with Poisson objective and exposure (offset).
5. Theoretical Guarantees and Temporal Coherence
Consistency of the estimator relies on two central properties:
- Boundedness ( in ).
- Approximate maximization (the fitted nearly maximizes over up to error).
Under these conditions and as the sieve partitions become increasingly fine, the estimator asymptotically recovers the true log-hazard surface in -norm: Temporal inconsistencies prevalent in landmark Cox models—where sequences of landmark-specific regression coefficients may not jointly define a proper conditional hazard without further assumptions—are avoided. The supermodel approach ensures internal coherence as the estimator is a unified function of , and tree-based sieves are dense in the relevant function space. No additional Markov, homogeneity, or consistency conditions (e.g., Jewell–Nielsen) are necessary (Sandqvist, 24 Jan 2026).
6. Empirical Evaluation and Applied Results
Simulation studies examine three principal settings:
- Scenario 1 (linear Markovian), 3 covariates: For small , the naive Cox method was optimal; at large (), both Cox and boosted-tree supermodels performed equivalently.
- Scenario 2 (nonlinear, non-Markovian): The boosted-tree supermodel yielded superior performance for moderate or large (); Cox supermodel was biased.
- Scenario 3 (high-dimensional noise, ): The boosted-tree supermodel resisted high variance, screening out 47 noise covariates, while the Cox supermodel displayed instability.
Across all settings, increasing the number of landmarks yielded progressive improvements in RMSE of survival predictions.
The primary biliary cirrhosis (PBC) case study utilized data from 312 patients with irregularly timed visits and uniformly distributed landmarks (). Key predictors—bilirubin and albumin—were identified via Shapley values and partial-dependence plots. Generated individual dynamic survival curves matched established clinical prognostic models. Computational time was approximately one minute on a standard laptop using R and XGBoost.
7. Summary and Significance
The landmarking supermodel with gradient boosted trees extends classical landmark analysis to a fully nonparametric regime, enabling dynamic, high-dimensional survival prediction without restrictive model assumptions. The method:
- Simultaneously leverages data from multiple landmark points via a single (t,s,w)-indexed supermodel,
- Frames estimation as a sieve M-estimator of a Poisson-regression likelihood,
- Employs gradient boosted trees for expressive, interpretable function approximation,
- Achieves weak -consistency of the fitted log-hazard,
- Resolves temporal incoherence that affects multi-landmark Cox fitting by construction,
- Remains computationally tractable and scalable to high-dimensional settings through standard machine learning tools (Sandqvist, 24 Jan 2026).
A plausible implication is increased applicability of dynamic event prediction in biomedical and other longitudinal studies, particularly where covariate processes are high-dimensional, non-Markovian, or where temporal interpretability is essential.