Optimization of Task Scheduling in Volunteer Grid Computing Using Reinforcement Learning Techniques
Keywords:
Volunteer Grid Computing, Reinforcement Learning, Q-Learning, Task Scheduling, Resource Optimization, Distributed SystemsAbstract
Volunteer Grid Computing (VGC) harnesses idle computational resources contributed by volunteers, offering a cost-effective and scalable platform for large-scale computing tasks. However, effective task scheduling remains a fundamental challenge due to the dynamic and unreliable nature of volunteer nodes. This paper explores the use of Reinforcement Learning (RL) to optimize task scheduling in VGC. A Q-learning-based scheduler was designed to adaptively assign tasks based on node reliability, availability, and performance history. Results from simulation demonstrate improved throughput and fault tolerance compared to traditional heuristics. This integration offers a promising pathway to autonomously and intelligently manage volunteer grids at scale.
References
Anderson, David P. “BOINC: A System for Public-Resource Computing and Storage.” Proceedings of the 5th IEEE/ACM International Workshop on Grid Computing, 2004, pp. 4–10.
Sarmenta, Luis F. G. “Volunteer Computing.” Doctoral Dissertation, MIT, 2002.
Choi, Sungyoung, et al. “A Resource Scheduling Scheme in Volunteer Grid Environment.” Future Generation Computer Systems, vol. 22, no. 6, 2006, pp. 742–749.
Buyya, Rajkumar, and M. Murshed. “GridSim: A Toolkit for the Modeling and Simulation of Distributed Resource Management and Scheduling for Grid Computing.” Concurrency and Computation: Practice and Experience, vol. 14, no. 13-15, 2002, pp. 1175–1220.
Mao, Hongzi, et al. “Resource Management with Deep Reinforcement Learning.” Proceedings of the 15th ACM Workshop on Hot Topics in Networks, 2016, pp. 50–56.
Xu, Hong, et al. “Online Job Scheduling with Selfish Users in Cloud Computing.” IEEE Transactions on Parallel and Distributed Systems, vol. 28, no. 4, 2017, pp. 1047–1060.
Wu, Hao, et al. “A Deep Reinforcement Learning Based Workflow Scheduling in Cloud Environment.” Future Generation Computer Systems, vol. 102, 2020, pp. 486–495.
Dastjerdi, Amir Vahid, and Rajkumar Buyya. “Fog Computing: Helping the Internet of Things Realize Its Potential.” Computer, vol. 49, no. 8, 2016, pp. 112–116.
Foster, Ian, et al. “Cloud Computing and Grid Computing 360-Degree Compared.” Grid Computing Environments Workshop, 2008.
Dinda, Peter, et al. “Resource Management in the GridSim Toolkit.” Parallel and Distributed Processing Symposium, 2005.
Ghobaei-Arani, Mehrdad, et al. “Resource Management Approaches in Fog Computing: A Comprehensive Review.” Journal of Grid Computing, vol. 18, no. 1, 2020, pp. 1–42.
Litzkow, Michael J., et al. “Condor—A Hunter of Idle Workstations.” Proceedings of the 8th International Conference on Distributed Computing Systems, 1988.
Zhang, Qi, et al. “Dynamic Resource Allocation for Spot Markets in Clouds via Reinforcement Learning.” IEEE INFOCOM, 2018.
Lin, Weiwei, et al. “Task Scheduling with Multi-Agent Reinforcement Learning in Cloud Computing.” Cluster Computing, vol. 22, no. 1, 2019, pp. 1313–1325.
Chen, Lin, et al. “Trust-Based Scheduling for Volunteer Cloud.” IEEE Transactions on Services Computing, vol. 12, no. 1, 2019, pp. 132–143.