Papers
Topics
Authors
Recent
Search
2000 character limit reached

Regression-Based Engine Scheduling Heuristic

Updated 14 December 2025
  • The paper introduces a hybrid regression-based scheduling approach that leverages deep neural network regressors as fast surrogates to approximate total tardiness.
  • It integrates Lawler’s decomposition with a single-layer LSTM regressor, enabling a recursive, single-pass scheduling scheme that mitigates the NP-hard exponential search space.
  • Empirical results demonstrate near-optimal performance (optimality gap ~0.5% for up to 325 jobs) with practical O(n^3) runtime compared to traditional heuristics.

A regression-based engine-scheduling heuristic is a hybrid algorithmic framework designed to solve the classical NP-hard single-machine total-tardiness problem by leveraging deep neural network regressors as fast, polynomial-time surrogates for exact combinatorial evaluation. The approach integrates domain knowledge from classical decomposition (specifically, Lawler’s theorem) with a learned estimator to guide recursive, single-pass scheduling. This paradigm achieves near-optimal scheduling performance on challenging instances (up to approximately 350 jobs), outperforming traditional heuristics while maintaining practical computational efficiency (Bouška et al., 2020).

1. Problem Formulation and Decomposition

The single-machine total-tardiness problem consists of scheduling nn independent, non-preemptive jobs J={1,,n}J=\{1,\ldots, n\}, each characterized by an integer processing time pj>0p_j > 0 and due date dj0d_j \geq 0. A schedule π\pi is a permutation of JJ, with completion time for job jj in π\pi given by

Cj(π)=i=1k(j)pπ(i),C_{j}(\pi) = \sum_{i=1}^{k(j)} p_{\pi(i)},

where k(j)k(j) is the position of jj in π\pi. The tardiness Tj(π)T_j(\pi) is

Tj(π)=max{0,Cj(π)dj}T_j(\pi) = \max\{0, C_j(\pi) - d_j\}

and the objective is to minimize total tardiness:

T(π)=j=1nTj(π).T(\pi) = \sum_{j=1}^n T_j(\pi).

Lawler’s decomposition theorem structurally constrains optimal schedules. Specifically, the job jpj^p with maximal pjp_j in JJ can only be placed at positions kpos0k \geq pos^0 in EDD (Earliest Due Date) order, where pos0pos^0 is the rank of jpj^p in EDD. Symmetrically, in SPT (Shortest Processing Time) order, only certain positions for the job jdj^d with the smallest djd_j are considered. Recursively, for each valid kk, the problem splits into two subproblems (left/right of kk), and the overall optimal value is computed as

Z(J)=minkKEDD[Z(Jleft)+max(0,(jJleftpj+pjp)djp)+Z(Jright)].Z(J) = \min_{k \in K_\text{EDD}} \left[ Z(J_\text{left}) + \max(0, (\sum_{j \in J_\text{left}} p_j + p_{j^p}) - d_{j^p}) + Z(J_\text{right}) \right].

This recursion naturally induces an exponential tree if all branches are explored optimally.

2. Regression Model Architecture and Training

To circumvent the exponential cost of recursive evaluation, a deep neural network regressor Y^()\widehat{Y}(\cdot) is introduced as a surrogate for the exact evaluation Z()Z(\cdot).

  • Input Features and Data Preprocessing: Jobs are first sorted in EDD order. Features for each job jj consist of the normalized processing time pj/Sp_j/S, normalized due date dj/Sd_j/S (with S=jJpjS = \sum_{j\in J} p_j), and a positional feature αj\alpha_j (its EDD rank divided by J|J|). Thus, each input is the sequence [pj,dj,αj]jJ[p_j, d_j, \alpha_j]_{j \in J}.
  • Network Design: A single-layer LSTM with 512 hidden units processes the sequence of job vectors. The final hidden state passes through a fully connected layer with linear activation to yield ynormRy_\text{norm}\in \mathbb{R}. De-normalization multiplies ynormy_\text{norm} by SS to output the estimated total tardiness Y^(J)\widehat{Y}(J).
  • Training: Instances are generated using the Potts–Van Wassenhove scheme: parameters include job count nn, due-date range RR, tardiness-factor TT, and pmaxp_\text{max}. Training labels are exact solutions Z(J)Z(J) generated by a state-of-the-art DP/branch-and-reduce solver (TTBR), normalized by SS. Optimization minimizes mean squared error with Adam (learning rate 10410^{-4}), and early stopping (patience=5) is applied.

3. Integration with Single-Pass Scheduling

The learned regressor is embedded within a single-pass greedy recursive scheduler based on Lawler’s decomposition. At each recursion:

  • EDD- and SPT-eligible position sets (KEDDK_\text{EDD}, KSPTK_\text{SPT}) are constructed, filtered by dominance rules.
  • The smaller set is selected for branching.
  • For each position kk in the set, the schedule is split: jobs to the left (JLJ_L), the selected job (jselj_\text{sel}), and jobs to the right (JRJ_R).
  • The cost is estimated as

cost=Y^(JL)+penalty+Y^(JR),\text{cost} = \widehat{Y}(J_L) + \text{penalty} + \widehat{Y}(J_R),

where penalty=max(0,CL+pjseldjsel)\text{penalty} = \max(0, C_L + p_{j_\text{sel}} - d_{j_\text{sel}}).

For Jκ|J| \leq \kappa (with κ=5\kappa=5), the exact solver is called. Otherwise, the procedure selects kk minimizing the estimated cost and recursively schedules left and right subproblems. Neural net predictions are cached to prevent redundant inference.

4. Computational Complexity

The time complexity of the heuristic (labeled “dhs” in the source) arises from the following structural properties:

  • Each call to Y^(J)\widehat{Y}(J') executes a single LSTM pass over J|J'| jobs (O(n)\mathcal{O}(n)).
  • Each recursion evaluates all K=O(n)|K| = \mathcal{O}(n) legal placements, each requiring two regressor calls, so each level is O(n2)\mathcal{O}(n^2).
  • With nn jobs and at most one removed per recursion, the recursion depth is O(n)O(n).

Thus, overall worst-case runtime is O(n3)O(n^3). Empirically, with practical implementation and modern hardware, the method is efficient even for n350n \approx 350 with seconds of runtime.

5. Empirical Results and Benchmarking

Extensive experiments validate the effectiveness and scalability of the regression-based heuristic. The evaluation spans varying nn (up to 500), pmax{100,5000}p_\text{max} \in \{100,5000\}, R,T{0.2,0.4,0.6,0.8,1.0}R,T \in \{0.2,0.4,0.6,0.8,1.0\}, with 200 random instances per setting. Competing methods include:

  • NBR: classical greedy/local-exchange [Holsenback & Russell, 1992].
  • DHSnbr_\text{nbr}: uses same decomposition but substitutes NBR for the neural regressor.
  • TTBR10^{10}: exact DP/branch-and-reduce with 10 s time limit.

Key metrics are the optimality gap (percentage over optimal) and CPU runtime.

nn NBR gap [%] DHSnbr_\text{nbr} gap [%] DHSNN_\text{NN} gap [%] NBR time [s] DHSnbr_\text{nbr} time [s] DHSNN_\text{NN} time [s]
225 1.98±0.581.98\pm0.58 1.17±0.471.17\pm0.47 0.58±0.300.58\pm0.30 0.06±0.010.06\pm0.01 1.19±0.421.19\pm0.42 5.03±8.165.03\pm8.16
275 2.12±0.542.12\pm0.54 1.31±0.441.31\pm0.44 0.57±0.280.57\pm0.28 0.09±0.020.09\pm0.02 1.91±0.621.91\pm0.62 6.89±9.626.89\pm9.62
325 2.20±0.502.20\pm0.50 1.39±0.431.39\pm0.43 0.57±0.370.57\pm0.37 0.12±0.020.12\pm0.02 2.87±0.902.87\pm0.90 9.25±11.299.25\pm11.29
375 2.27±0.492.27\pm0.49 1.46±0.441.46\pm0.44 1.23±0.631.23\pm0.63 0.17±0.030.17\pm0.03 4.15±1.314.15\pm1.31 14.61±13.5214.61\pm13.52
425 2.34±0.462.34\pm0.46 1.55±0.411.55\pm0.41 1.71±0.651.71\pm0.65 0.21±0.040.21\pm0.04 5.52±1.715.52\pm1.71 20.60±17.0020.60\pm17.00

For n325n \leq 325, the regression-based heuristic (DHSNN_\text{NN}) achieves an optimality gap of approximately 0.5%0.5\%, about four times better than NBR. Runtime grows as O(n3)O(n^3) but remains practical (<20 s for n425n\leq425). When increasing pmaxp_\text{max} to 5000, the heuristic’s gap remains stable, while TTBR10^{10} and NBR degrade.

6. Implementation Considerations

Efficient implementation requires:

  • Use of modern DL frameworks (TensorFlow, PyTorch) for the LSTM regressor.
  • On-the-fly computation of SS and EDD-sorted lists within the schedule.
  • Aggressive caching of regressor predictions (Y^(J)\widehat{Y}(J)) for subproblem reuse.
  • Tuning of the exact-solver threshold κ\kappa (κ=5\kappa=5 achieves a robust tradeoff).
  • Parallelization is feasible, especially for GPU-based inference.

7. Significance and Outlook

This regression-based engine-scheduling heuristic demonstrates that neural regressors—when judiciously embedded within classical decomposition frameworks—yield scalable, high-quality solutions for combinatorial scheduling problems. The approach occupies a hybrid space between exact DP/branch-and-reduce and purely handcrafted heuristics, offering polynomial runtime, strong optimality guarantees on moderate-size instances, and generalization to instance regimes beyond those seen in training (Bouška et al., 2020). A plausible implication is that similar regression-guided heuristics may be extensible to other sequencing and vehicle-routing variants where classical decompositions and learned cost proxies can be integrated systematically.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Regression-Based Engine-Scheduling Heuristic.