Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sparse Mean Field Load Balancing in Large Localized Queueing Systems (2312.12973v2)

Published 20 Dec 2023 in cs.DC, cs.LG, cs.NI, cs.SY, and eess.SY

Abstract: Scalable load balancing algorithms are of great interest in cloud networks and data centers, necessitating the use of tractable techniques to compute optimal load balancing policies for good performance. However, most existing scalable techniques, especially asymptotically scaling methods based on mean field theory, have not been able to model large queueing networks with strong locality. Meanwhile, general multi-agent reinforcement learning techniques can be hard to scale and usually lack a theoretical foundation. In this work, we address this challenge by leveraging recent advances in sparse mean field theory to learn a near-optimal load balancing policy in sparsely connected queueing networks in a tractable manner, which may be preferable to global approaches in terms of wireless communication overhead. Importantly, we obtain a general load balancing framework for a large class of sparse bounded-degree wireless topologies. By formulating a novel mean field control problem in the context of graphs with bounded degree, we reduce the otherwise difficult multi-agent problem to a single-agent problem. Theoretically, the approach is justified by approximation guarantees. Empirically, the proposed methodology performs well on several realistic and scalable wireless network topologies as compared to a number of well-known load balancing heuristics and existing scalable multi-agent reinforcement learning methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Multi-agent reinforcement learning: A selective overview of theories and algorithms, Handbook of Reinforcement Learning and Control (2021) 321–384.
  2. A primer on partially observable markov decision processes (POMDPs), Methods in Ecology and Evolution 12 (2021) 2058–2072.
  3. Multi-agent reinforcement learning for networked system control, in: 8th International Conference on Learning Representations, OpenReview.net, Addis Ababa, Ethiopia, April 26-30, 2020, 2020, pp. 1–17.
  4. Multi-agent reinforcement learning in stochastic networked systems, Advances in Neural Information Processing Systems 34 (2021) 7825–7837.
  5. Scalable and sample efficient distributed policy gradient algorithms in multi-agent networked systems, arXiv preprint arXiv:2212.06357 (2022) 1–30.
  6. S. Guicheng, W. Yang, Review on Dec-POMDP model for MARL algorithms, in: Smart Communications, Intelligent Algorithms and Interactive Methods, Springer, 2022, pp. 29–35.
  7. C. Amato, F. Oliehoek, Scalable planning and learning for multiagent POMDPs, in: Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, PKP Publishing Services Network, 2015, pp. 1–8.
  8. Asymptotically optimal load balancing in large-scale heterogeneous systems with multiple dispatchers, Performance Evaluation 145 (2021) 102146.
  9. Local weak convergence for sparse networks of interacting processes, The Annals of Applied Probability 33 (2023) 843–888.
  10. A survey on large-population systems and scalable multi-agent reinforcement learning, arXiv preprint arXiv:2209.03859 (2022) 1–21.
  11. J.-M. Lasry, P.-L. Lions, Mean field games, Japanese Journal of Mathematics 2 (2007) 229–260.
  12. M. Mitzenmacher, The power of two choices in randomized load balancing, IEEE Transactions on Parallel and Distributed Systems 12 (2001) 1094–1104.
  13. W. Winston, Optimality of the shortest line discipline, Journal of Applied Probability 14 (1977) 181–189.
  14. Universality of power-of-d load balancing in many-server systems, Stochastic Systems 8 (2018) 265–292.
  15. Balancing queues by mean field interaction, Queueing Systems 49 (2005) 335–361.
  16. Asymptotically optimal load balancing topologies, Proceedings of the ACM on Measurement and Analysis of Computing Systems 2 (2018) 1–29.
  17. Graphon mean-field control for cooperative multi-agent reinforcement learning, arXiv preprint arXiv:2209.04808 (2022) 1–25.
  18. K. Cui, H. Koeppl, Learning graphon mean field games and approximate nash equilibria, in: The 10th International Conference on Learning Representations, OpenReview.net, 2022, pp. 1–31.
  19. P. E. Caines, M. Huang, Graphon mean field games and their equations, SIAM Journal on Control and Optimization 59 (2021) 4373–4399.
  20. Learning sparse graphon mean field games, arXiv preprint arXiv:2209.03880 (2022) 1–32.
  21. N. Gast, The power of two choices on graphs: The pair-approximation is accurate?, ACM SIGMETRICS Performance Evaluation Review 43 (2015) 69–71.
  22. D. Rutten, D. Mukherjee, Mean-field analysis for load balancing on spatial graphs, arXiv preprint arXiv:2301.03493 (2023) 1–27.
  23. Learning mean-field control for delayed information load balancing in large queuing systems, in: Proceedings of the 51st International Conference on Parallel Processing, ACM New York, NY, USA, 2022, pp. 1–11.
  24. D. Lipshutz, Open problem—load balancing using delayed information, Stochastic Systems 9 (2019) 305–306.
  25. W. Fischer, K. Meier-Hellstern, The markov-modulated poisson process (mmpp) cookbook, Performance evaluation 18 (1993) 149–171.
  26. Thinning and the law of small numbers, in: IEEE International Symposium on Information Theory, IEEE, 2007, pp. 1491–1495.
  27. The surprising effectiveness of PPO in cooperative, multi-agent games, Proceedings NeurIPS Datasets and Benchmarks (2022) 1–14.
  28. Shared experience actor-critic for multi-agent reinforcement learning, Advances in Neural Information Processing Systems 33 (2020) 10707–10717.
  29. Benchmarking multi-agent deep reinforcement learning algorithms in cooperative tasks, in: Proceedings NeurIPS Datasets and Benchmarks, The MIT Press, 2021, pp. 15–19.
  30. H. Pham, X. Wei, Bellman equation and viscosity solutions for mean-field stochastic control problem, ESAIM: Control, Optimisation and Calculus of Variations 24 (2018) 437–461.
  31. Proximal policy optimization algorithms, arXiv preprint arXiv:1707.06347 (2017) 1–12.
  32. Recurrent model-free RL can be a strong baseline for many POMDPs, in: K. Chaudhuri, S. Jegelka, L. Song, C. Szepesvari, G. Niu, S. Sabato (Eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162, PMLR, 2022, pp. 16691–16723.
  33. Rllib: Abstractions for distributed reinforcement learning, in: International Conference on Machine Learning, PMLR, 2018, pp. 3053–3062.
  34. E. Koenigsberg, Twenty five years of cyclic queues and closed queue networks: A review, Journal of the Operational Research Society 33 (1982) 605–619.
  35. F. P. Preparata, J. Vuillemin, The cube-connected cycles: a versatile network for parallel computation, Communications of the ACM 24 (1981) 300–309.
  36. H. Habibian, A. Patooghy, Fault-tolerant routing methodology for hypercube and cube-connected cycles interconnection networks, The Journal of Supercomputing 73 (2017) 4560–4579.
  37. Geometric mapping of tasks to processors on parallel computers with mesh or torus networks, IEEE Transactions on Parallel and Distributed Systems 30 (2019) 2018–2032.
  38. Algorithms for mapping parallel processes onto grid and torus architectures, in: 23rd Euromicro International Conference on Parallel, Distributed, and Network-Based Processing, IEEE, 2015, pp. 236–243.
  39. Configuring random graph models with fixed degree sequences, SIAM Review 60 (2018) 315–355.
  40. M. Ostilli, Cayley trees and bethe lattices: A concise analysis for mathematicians and physicists, Physica A: Statistical Mechanics and its Applications 391 (2012) 3417–3423.
  41. G. Szabó, I. Borsos, Evolutionary potential games on lattices, Physics Reports 624 (2016) 1–60.
  42. D. T. Gillespie, Exact stochastic simulation of coupled chemical reactions, The Journal of Physical Chemistry 81 (1977) 2340–2361.
  43. Steady-state analysis of shortest expected delay routing, Queueing Systems 84 (2016) 309–354.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Anam Tahir (10 papers)
  2. Kai Cui (28 papers)
  3. Heinz Koeppl (105 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com