Papers
Topics
Authors
Recent
Search
2000 character limit reached

Dynamic Vector Bin Packing (DVBP)

Updated 30 December 2025
  • Dynamic Vector Bin Packing is a framework for online scheduling of multi-dimensional, time-varying tasks that ensures concurrent resource constraints on homogeneous servers.
  • Its methodology employs Any Fit algorithms such as First Fit, Next Fit, and Move To Front, with competitive analysis providing insights through tight upper and lower bounds.
  • Practical data reduction techniques like time-compression and (1+ε)-deletion enable significant instance size shrinkage while preserving near-optimal resource allocation.

Dynamic Vector Bin Packing (DVBP) generalizes classical bin packing by capturing the temporal and multi-dimensional nature of resource allocation in cloud computing scenarios, where tasks (jobs) exhibit vector-valued resource requirements and overlapping lifespans. The DVBP problem models online assignment of such jobs to homogeneous servers (bins), striving to minimize the total usage or number of bins while respecting concurrent multi-resource constraints at all times. This framework encompasses both practical scheduling applications and a rich theoretical landscape spanning competitive analysis, non-clairvoyant scheduling, and parameterized complexity.

1. Formal Definition and Problem Variants

DVBP is defined by a set of nn jobs (requests) %%%%1%%%%, each with:

  • Arrival time a(r)Q0a(r)\in\mathbb Q_{\ge0}
  • Departure time e(r)Q0e(r)\in\mathbb Q_{\ge0}, with lifespan I(r)=[a(r),e(r))I(r) = [a(r),e(r))
  • dd-dimensional resource demand s(r)[0,1]ds(r)\in[0,1]^d (or, in the integer variant, a(i)Nda^{(i)}\in\mathbb N^d)

At any time tt, an assignment of jobs to bins B1,,BkB_1,\dots,B_k must guarantee

rBj,a(r)t<e(r)s(r)1\sum_{r\in B_j, a(r)\le t<e(r)} s(r) \leq \mathbf{1}

(coordinate-wise), where 1=(1,,1)\mathbf{1}=(1,\ldots,1) represents the per-bin capacity.

Objective functions in DVBP differ according to the underlying system model:

  • MinUsageTime DVBP: Minimize total server rental time,

cost(A,R)=i=1mspan(Ri)\mathrm{cost}(A,R) = \sum_{i=1}^m \operatorname{span}(R_i)

where span(Ri)\operatorname{span}(R_i) denotes the union of activity intervals of jobs in bin BiB_i. The competitive ratio is

CR(A)=supRcost(A,R)OPT(R)\mathrm{CR}(A) = \sup_R \frac{\mathrm{cost}(A,R)}{\mathrm{OPT}(R)}

with OPT(R)\mathrm{OPT}(R) defined as the optimal (possibly clairvoyant) solution.

  • Classical (static) DVBP: Minimize the total number of bins required over the interval set, subject to the multi-dimensional packing constraints at every time.

Parameterization frequently involves:

  • dd: number of resource dimensions
  • μ=maxr(r)\mu = \max_r \ell(r) ((r)=e(r)a(r)\ell(r) = e(r) - a(r)): duration ratio of the longest to shortest job
  • hh: maximum concurrent jobs (instance "height")
  • φ\varphi: number of distinct job "flavors" (unique demand vectors)

2. Any-Fit Algorithms and Algorithmic Schemes

Any Fit algorithms are a class of online procedures that open new bins only if existing bins cannot accommodate an incoming job rr without exceeding the resource constraint in any dimension. Core representatives include:

  • First Fit (FF): Maintain bins in order of creation; scan from oldest to newest to find the first feasible bin for rr.
  • Next Fit (NF): Only one bin is kept open at any time; if rr does not fit, close the current bin and open a new one.
  • Move To Front (MTF): Bins are ordered by recent use; an assigned bin moves to the front, biasing toward temporal alignment among jobs.
  • Best Fit (BF): Place rr in the bin that would maximize post-assignment bin utilization (largest L(B)+s(r)\|L(B)+s(r)\|_{\infty} without overflow); known to have unbounded worst-case ratio even for d=1d=1.

Upon job departure, bins emptied of all jobs are removed from tracking.

The schematic pseudocode for Any Fit algorithms is:

1
2
3
4
5
6
7
Initialize empty bin list L.
For each arriving job r:
  Scan L in prescribed order for bin B where L(B) + s(r) <= 1 (coordinatewise).
  If found, assign r to B (and update L order as required).
  Else, open new bin B', add to L, assign r to B'.
Upon job departure:
  If a bin becomes empty, remove it from L.

The specific reordering and scanning protocols distinguish algorithm variants and critically impact competitive behavior.

3. Competitive Analysis: Upper and Lower Bounds

The theoretical performance of Any Fit algorithms is captured via worst-case competitive ratio bounds, dependent on μ\mu and dd:

Proven Lower Bounds

  • Any Fit (all dd): CR(μ+1)d\mathrm{CR}\ge (\mu+1)d
  • Next Fit: CR2μd\mathrm{CR}\ge 2\mu d
  • Move To Front (d=1d=1): CR2μ\mathrm{CR}\ge 2\mu

These bounds are established via adversarial instance constructions in which carefully crafted job sequences force inefficiencies in online assignment relative to an optimal clairvoyant scheduler.

Nearly Tight Upper Bounds

The main results for MinUsageTime DVBP yield:

Algorithm Competitive Ratio Upper Bound
Move To Front (2μ+1)d+1(2\mu+1)d+1
First Fit (μ+2)d+1(\mu+2)d+1
Next Fit 2μd+12\mu d+1

The Move To Front analysis partitions each bin's active interval into leading and non-leading segments, showing that non-leading intervals can be bounded in aggregate by (2μ+1)dOPT(2\mu+1)d\cdot \mathrm{OPT}, while leading intervals sum to at most OPT\mathrm{OPT}. The gap between lower and upper bounds for MTF in d2d\geq 2 remains open.

First Fit and Next Fit bounds utilize dimension-aware charge arguments to account for multi-dimensional demand and temporal overlap.

4. Data Reduction and Approximation in Large-Scale DVBP

The complexity of DVBP instances motivates preprocessing approaches to shrink problem size while retaining (near-)optimality.

Kernelization and Lower Bounds

It is proven that DVBP admits no polynomial-size kernel in (h+φ)(h+\varphi), even for (1+ε)(1+\varepsilon)-approximation, unless the polynomial hierarchy collapses. Only an exponential-size kernel of O(h2h)O(h\cdot 2^h) is available.

Practical Polynomial-Time Reductions

Despite theoretical hardness, two data reduction rules achieve dramatic real-world instance shrinkage:

  • Time-compression: Remaps all start/end points onto a contiguous range [1,,T][1,\dotsc,T'] with T2nT'\leq 2n, preserving all job interval overlaps, often reducing the number of types TT by a factor of 6\sim 6.
  • (1+ε)(1+\varepsilon)-deletion rule: Greedily deletes up to εL\lfloor \varepsilon L \rfloor bins' worth of requests from the instance, where LL is a time-indexed lower bound on the number of bins, enabling any α\alpha-approximate algorithm on the reduced instance to guarantee an α(1+ε)\alpha (1+\varepsilon)-approximation overall.

These techniques achieve order-of-magnitude reductions in instance size on industry traces (e.g., Azure, Huawei) while preserving bounded optimality gaps and polynomial running time (Murhekar et al., 2023, Bevern et al., 2022).

5. Experimental Evaluations and Empirical Findings

Extensive experiments on both synthetic and real-world traces illustrate key behaviors:

Online Algorithms:

  • On synthetic data (d{1,2,5}d\in\{1,2,5\}, varying μ\mu), Move To Front exhibits the best average-case ratio for all evaluated parameter regimes.
  • First Fit and Best Fit perform comparably (with low variance waste), while Next Fit's performance deteriorates with increasing μ\mu.
  • Heuristics such as Worst Fit and Random Fit display high packing waste and variability.

Data Reduction:

  • Time-compression alone reduces type counts by 1/3\sim 1/3 without reducing items.
  • The combination of time-compression and (1+ε)(1+\varepsilon)-deletion (ε=0.05\varepsilon=0.05) reduces item counts by 12.6×12.6\times (Huawei) and 5.4×5.4\times (Azure), and type counts by 50×\sim 50\times.
  • Reduced instances retain at least 95%95\% of the theoretical maximum deletions, with overall instance size reduction enabling tractability for stronger solvers.
Dataset Initial nn nn after reduction Initial τ\tau τ\tau after reduction Utilization RR Factor KK
Huawei 111,774 8,849 9 5 0.969 1.576
Azure 3,792,000 704,832 74 49 0.946 1.319

6. Open Questions and Research Directions

Several unresolved topics and potential advancements in DVBP research are identified:

  • Closing competitive ratio gaps: For d2d\geq 2, the gap between lower (μ+1)d(\mu+1)d and upper (2μ+1)d+1(2\mu+1)d+1 bounds for Move To Front persists.
  • Algorithmic improvements for low dimensions: Specialized designs for small dd (d=2d=2 or $3$) relevant to practical cloud settings may yield better theoretical and empirical results.
  • Beyond worst-case: Best Fit exhibits unbounded worst-case but competitive average-case ratios, suggesting the need for distributional or smoothed analyses.
  • Clairvoyant or stochastic enhancements: Leveraging predictions (including machine-learned models) of future arrivals or durations remains an open frontier, especially for dd-dimensional cases.
  • Multiple objectives: Incorporation of real-world criteria such as power usage, migration overheads, or startup latencies suggests multi-objective DVBP formulations.

In summary, DVBP offers a mathematically rigorous abstraction for dynamic, multi-resource online scheduling, with a deep body of work characterizing both worst-case guarantees and practical reduction techniques for large-scale deployments in cloud environments (Murhekar et al., 2023, Bevern et al., 2022).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Dynamic Vector Bin Packing (DVBP).