Over the past three decades, supercomputers and their workloads have become increasingly complex. Scheduling systems have evolved from traditional heuristics to Deep Reinforcement Learning (DRL) approaches that adapt policies to specific workloads. Though there are several studies that develop various DRL models, no clear consensus exists on the optimal algorithm. This project trains and evaluates representative algorithms from three prominent DRL families: DQN, PPO, and A2C. Statistical testing (Friedman, Nemenyi, Wilcoxon-based confidence intervals) determines whether significant performance differences exist across five industry-standard metrics.
Papers, presentations, and recordings for each submission milestone
April 2026
Initial submission presenting problem statement, literature review findings, proposed methodology, and current progress.
Expected: May 2026
Progress update with DRL training infrastructure implementation, initial training results, and preliminary statistical analysis.
Expected: June 2026
Complete statistical analysis results, Critical Difference diagrams, and comprehensive evaluation of all DRL algorithms.
Expected: July 2026
Complete Honours paper with full results, discussion, conclusions, and future work recommendations.
Researchers and supervisors contributing to this project