Regression-Based Engine Scheduling Heuristic
- The paper introduces a hybrid regression-based scheduling approach that leverages deep neural network regressors as fast surrogates to approximate total tardiness.
- It integrates Lawler’s decomposition with a single-layer LSTM regressor, enabling a recursive, single-pass scheduling scheme that mitigates the NP-hard exponential search space.
- Empirical results demonstrate near-optimal performance (optimality gap ~0.5% for up to 325 jobs) with practical O(n^3) runtime compared to traditional heuristics.
A regression-based engine-scheduling heuristic is a hybrid algorithmic framework designed to solve the classical NP-hard single-machine total-tardiness problem by leveraging deep neural network regressors as fast, polynomial-time surrogates for exact combinatorial evaluation. The approach integrates domain knowledge from classical decomposition (specifically, Lawler’s theorem) with a learned estimator to guide recursive, single-pass scheduling. This paradigm achieves near-optimal scheduling performance on challenging instances (up to approximately 350 jobs), outperforming traditional heuristics while maintaining practical computational efficiency (Bouška et al., 2020).
1. Problem Formulation and Decomposition
The single-machine total-tardiness problem consists of scheduling independent, non-preemptive jobs , each characterized by an integer processing time and due date . A schedule is a permutation of , with completion time for job in given by
where is the position of in . The tardiness is
and the objective is to minimize total tardiness:
Lawler’s decomposition theorem structurally constrains optimal schedules. Specifically, the job with maximal in can only be placed at positions in EDD (Earliest Due Date) order, where is the rank of in EDD. Symmetrically, in SPT (Shortest Processing Time) order, only certain positions for the job with the smallest are considered. Recursively, for each valid , the problem splits into two subproblems (left/right of ), and the overall optimal value is computed as
This recursion naturally induces an exponential tree if all branches are explored optimally.
2. Regression Model Architecture and Training
To circumvent the exponential cost of recursive evaluation, a deep neural network regressor is introduced as a surrogate for the exact evaluation .
- Input Features and Data Preprocessing: Jobs are first sorted in EDD order. Features for each job consist of the normalized processing time , normalized due date (with ), and a positional feature (its EDD rank divided by ). Thus, each input is the sequence .
- Network Design: A single-layer LSTM with 512 hidden units processes the sequence of job vectors. The final hidden state passes through a fully connected layer with linear activation to yield . De-normalization multiplies by to output the estimated total tardiness .
- Training: Instances are generated using the Potts–Van Wassenhove scheme: parameters include job count , due-date range , tardiness-factor , and . Training labels are exact solutions generated by a state-of-the-art DP/branch-and-reduce solver (TTBR), normalized by . Optimization minimizes mean squared error with Adam (learning rate ), and early stopping (patience=5) is applied.
3. Integration with Single-Pass Scheduling
The learned regressor is embedded within a single-pass greedy recursive scheduler based on Lawler’s decomposition. At each recursion:
- EDD- and SPT-eligible position sets (, ) are constructed, filtered by dominance rules.
- The smaller set is selected for branching.
- For each position in the set, the schedule is split: jobs to the left (), the selected job (), and jobs to the right ().
- The cost is estimated as
where .
For (with ), the exact solver is called. Otherwise, the procedure selects minimizing the estimated cost and recursively schedules left and right subproblems. Neural net predictions are cached to prevent redundant inference.
4. Computational Complexity
The time complexity of the heuristic (labeled “dhs” in the source) arises from the following structural properties:
- Each call to executes a single LSTM pass over jobs ().
- Each recursion evaluates all legal placements, each requiring two regressor calls, so each level is .
- With jobs and at most one removed per recursion, the recursion depth is .
Thus, overall worst-case runtime is . Empirically, with practical implementation and modern hardware, the method is efficient even for with seconds of runtime.
5. Empirical Results and Benchmarking
Extensive experiments validate the effectiveness and scalability of the regression-based heuristic. The evaluation spans varying (up to 500), , , with 200 random instances per setting. Competing methods include:
- NBR: classical greedy/local-exchange [Holsenback & Russell, 1992].
- DHS: uses same decomposition but substitutes NBR for the neural regressor.
- TTBR: exact DP/branch-and-reduce with 10 s time limit.
Key metrics are the optimality gap (percentage over optimal) and CPU runtime.
| NBR gap [%] | DHS gap [%] | DHS gap [%] | NBR time [s] | DHS time [s] | DHS time [s] | |
|---|---|---|---|---|---|---|
| 225 | ||||||
| 275 | ||||||
| 325 | ||||||
| 375 | ||||||
| 425 |
For , the regression-based heuristic (DHS) achieves an optimality gap of approximately , about four times better than NBR. Runtime grows as but remains practical (<20 s for ). When increasing to 5000, the heuristic’s gap remains stable, while TTBR and NBR degrade.
6. Implementation Considerations
Efficient implementation requires:
- Use of modern DL frameworks (TensorFlow, PyTorch) for the LSTM regressor.
- On-the-fly computation of and EDD-sorted lists within the schedule.
- Aggressive caching of regressor predictions () for subproblem reuse.
- Tuning of the exact-solver threshold ( achieves a robust tradeoff).
- Parallelization is feasible, especially for GPU-based inference.
7. Significance and Outlook
This regression-based engine-scheduling heuristic demonstrates that neural regressors—when judiciously embedded within classical decomposition frameworks—yield scalable, high-quality solutions for combinatorial scheduling problems. The approach occupies a hybrid space between exact DP/branch-and-reduce and purely handcrafted heuristics, offering polynomial runtime, strong optimality guarantees on moderate-size instances, and generalization to instance regimes beyond those seen in training (Bouška et al., 2020). A plausible implication is that similar regression-guided heuristics may be extensible to other sequencing and vehicle-routing variants where classical decompositions and learned cost proxies can be integrated systematically.