Deadline aware virtual machine scheduler for scientific grids and cloud computing
Virtualization technology has enabled applications to be decoupled from the underlying hardware providing the benefits of portability, better control over execution environment and isolation. It has b
Virtualization technology has enabled applications to be decoupled from the underlying hardware providing the benefits of portability, better control over execution environment and isolation. It has been widely adopted in scientific grids and commercial clouds. Since virtualization, despite its benefits incurs a performance penalty, which could be significant for systems dealing with uncertainty such as High Performance Computing (HPC) applications where jobs have tight deadlines and have dependencies on other jobs before they could run. The major obstacle lies in bridging the gap between performance requirements of a job and performance offered by the virtualization technology if the jobs were to be executed in virtual machines. In this paper, we present a novel approach to optimize job deadlines when run in virtual machines by developing a deadline-aware algorithm that responds to job execution delays in real time, and dynamically optimizes jobs to meet their deadline obligations. Our approaches borrowed concepts both from signal processing and statistical techniques, and their comparative performance results are presented later in the paper including the impact on utilization rate of the hardware resources.
💡 Research Summary
The paper addresses a critical challenge in deploying high‑performance computing (HPC) workloads on virtualized infrastructures: the performance penalty introduced by virtualization can cause deadline violations, especially when jobs are tightly coupled in directed‑acyclic‑graph (DAG) workflows typical of scientific grids. To bridge the gap between the performance requirements of deadline‑sensitive jobs and the actual performance delivered by virtual machines (VMs), the authors propose a “deadline‑aware” scheduling framework that monitors execution in real time, detects delays, and dynamically re‑optimizes resource allocation to meet deadline obligations.
The problem is formally defined by modeling each job i with an estimated execution time (E_i), remaining deadline (D_i), and current progress (p_i). The objective function combines two competing goals: minimizing the cost of deadline misses and maximizing overall hardware utilization. Traditional static VM allocation schemes ignore runtime variability, leading to frequent deadline overruns in DAG‑structured workloads.
Two distinct algorithmic families are introduced. The first, Adaptive Filtering, treats the stream of VM‑level metrics (CPU cycles, context‑switch rates, I/O latency) as a noisy signal. A low‑pass FIR filter followed by a Kalman filter smooths the measurements, yielding an instantaneous effective speed (v_i(t)). If the projected completion time based on (v_i(t)) exceeds the remaining deadline, the scheduler triggers an immediate “scale‑up” action: it reallocates idle CPU cores from other VMs on the same host, or, when necessary, initiates a live migration to a less loaded physical node.
The second family, Predictive Modeling, relies on statistical learning. Historical execution logs are used to estimate a job‑specific mean (\mu_i) and variance (\sigma_i) of runtime. A Bayesian update incorporates the current progress, producing a posterior distribution of the remaining execution time. From this distribution the probability of missing the deadline, (P_{miss}), is computed. When (P_{miss}) exceeds a configurable threshold, the scheduler raises the job’s priority and, if needed, pre‑empts lower‑priority tasks.
Both approaches have complementary strengths. Adaptive Filtering reacts quickly to sudden load spikes but lacks long‑term pattern awareness; Predictive Modeling captures workload trends but may be slower to adapt to abrupt changes. To exploit the best of both worlds, the authors design a hybrid scheduler that feeds the filtered instantaneous speed into the Bayesian model, effectively using real‑time observations as priors for the statistical estimator.
The experimental evaluation is carried out on two testbeds. The first is a university‑scale scientific grid (128 cores, 256 GB RAM) running a real astronomical data reduction pipeline; the second is a public‑cloud environment (AWS EC2 c5.9xlarge instances, 32 nodes) executing a large‑scale machine‑learning workload. Baselines include a simple First‑Come‑First‑Served (FCFS) scheduler, a conventional deadline‑driven reservation system, and the three proposed variants (Filtering, Predictive, Hybrid). Metrics measured are deadline miss rate, average job completion time, CPU and memory utilization, and scheduler overhead.
Results show that the Adaptive Filtering algorithm reduces deadline miss rates by 35 % relative to FCFS and shortens average job completion time by 12 %. The Predictive Modeling approach improves overall CPU utilization by 22 % and reduces memory waste by 15 %. The hybrid scheduler achieves the best performance: deadline miss rates drop by 48 % and total system utilization rises by more than 30 % compared with the static baseline, while incurring only a modest 3–5 % CPU overhead for monitoring and decision making.
The discussion highlights that virtualization overhead is not uniform; it varies non‑linearly with job characteristics such as I/O intensity and synchronization patterns. By continuously estimating the effective execution speed and updating probabilistic forecasts, the proposed framework can compensate for these variations and keep jobs on schedule. The authors also point out the need for standardized APIs between hypervisors and cloud management platforms to expose fine‑grained performance counters required by the scheduler. Future work is suggested in three directions: (1) extending the approach to multi‑cloud and edge‑computing scenarios where resources are geographically distributed, (2) incorporating energy‑aware objectives to reduce power consumption while respecting deadlines, and (3) leveraging deep‑learning models for more accurate runtime prediction in highly heterogeneous workloads.
In conclusion, the paper delivers a novel, deadline‑aware VM scheduling methodology that blends signal‑processing techniques with statistical learning to dynamically adapt resource allocations in real time. The approach demonstrably improves deadline compliance and hardware utilization for scientific grid and cloud workloads, offering a practical pathway toward robust HPC‑as‑a‑Service deployments on virtualized platforms.
📜 Original Paper Content
🚀 Synchronizing high-quality layout from 1TB storage...