Large Language Models for Networking: Workflow, Advances and Challenges (2404.12901v2)
Abstract: The networking field is characterized by its high complexity and rapid iteration, requiring extensive expertise to accomplish network tasks, ranging from network design, configuration, diagnosis and security. The inherent complexity of these tasks, coupled with the ever-changing landscape of networking technologies and protocols, poses significant hurdles for traditional machine learning-based methods. These methods often struggle to generalize and automate complex tasks in networking, as they require extensive labeled data, domain-specific feature engineering, and frequent retraining to adapt to new scenarios. However, the recent emergence of LLMs has sparked a new wave of possibilities in addressing these challenges. LLMs have demonstrated remarkable capabilities in natural language understanding, generation, and reasoning. These models, trained on extensive data, can benefit the networking domain. Some efforts have already explored the application of LLMs in the networking domain and revealed promising results. By reviewing recent advances, we present an abstract workflow to describe the fundamental process involved in applying LLM for Networking. We introduce the highlights of existing works by category and explain in detail how they operate at different stages of the workflow. Furthermore, we delve into the challenges encountered, discuss potential solutions, and outline future research prospects. We hope that this survey will provide insight for researchers and practitioners, promoting the development of this interdisciplinary research field.
- R. Boutaba, M. A. Salahuddin, N. Limam, S. Ayoubi, N. Shahriar, F. Estrada-Solano, and O. M. Caicedo, “A comprehensive survey on machine learning for networking: evolution, applications and research opportunities,” Journal of Internet Services and Applications, vol. 9, no. 1, pp. 1–99, 2018.
- P. Sharma and V. Yegneswaran, “Prosper: Extracting protocol specifications using large language models,” in Proceedings of the 22nd ACM Workshop on Hot Topics in Networks, 2023, pp. 41–47.
- S. K. Mani, Y. Zhou, K. Hsieh, S. Segarra, T. Eberl, E. Azulai, I. Frizler, R. Chandra, and S. Kandula, “Enhancing network management using code generated by large language models,” in Proceedings of the 22nd ACM Workshop on Hot Topics in Networks, 2023, pp. 196–204.
- R. Mondal, A. Tang, R. Beckett, T. Millstein, and G. Varghese, “What do llms need to synthesize correct router configurations?” in Proceedings of the 22nd ACM Workshop on Hot Topics in Networks, 2023, pp. 189–195.
- X. Lian, Y. Chen, R. Cheng, J. Huang, P. Thakkar, and T. Xu, “Configuration validation with large language models,” arXiv preprint arXiv:2310.09690, 2023.
- Y. Zhou, N. Yu, and Z. Liu, “Towards interactive research agents for internet incident investigation,” in Proceedings of the 22nd ACM Workshop on Hot Topics in Networks, 2023, pp. 33–40.
- M. Kotaru, “Adapting foundation models for operator data analytics,” in Proceedings of the 22nd ACM Workshop on Hot Topics in Networks, 2023, pp. 172–179.
- R. Meng, M. Mirchev, M. Böhme, and A. Roychoudhury, “Large language model guided protocol fuzzing,” in Proceedings of the 31st Annual Network and Distributed System Security Symposium (NDSS), 2024.
- Y. Huang, H. Du, X. Zhang, D. Niyato, J. Kang, Z. Xiong, S. Wang, and T. Huang, “Large language models for networking: Applications, enabling techniques, and challenges,” arXiv preprint arXiv:2311.17474, 2023.
- P. Hamadanian, B. Arzani, S. Fouladi, S. K. R. Kakarla, R. Fonseca, D. Billor, A. Cheema, E. Nkposong, and R. Chandra, “A holistic view of ai-driven network incident management,” in Proceedings of the 22nd ACM Workshop on Hot Topics in Networks, 2023, pp. 180–188.
- D. Wu, X. Wang, Y. Qiao, Z. Wang, J. Jiang, S. Cui, and F. Wang, “Large language model adaptation for networking,” arXiv preprint arXiv:2402.02338, 2024.
- B. Paranjape, S. Lundberg, S. Singh, H. Hajishirzi, L. Zettlemoyer, and M. T. Ribeiro, “Art: Automatic multi-step reasoning and tool-use for large language models,” arXiv preprint arXiv:2303.09014, 2023.
- Y. Ge, W. Hua, K. Mei, J. Tan, S. Xu, Z. Li, Y. Zhang et al., “Openagi: When llm meets domain experts,” Advances in Neural Information Processing Systems, vol. 36, 2024.
- S. Zhang, P. Jin, Z. Lin, Y. Sun, B. Zhang, S. Xia, Z. Li, Z. Zhong, M. Ma, W. Jin et al., “Robust failure diagnosis of microservice system through multimodal data,” IEEE Transactions on Services Computing, 2023.
- Chang Liu (864 papers)
- Xiaohui Xie (84 papers)
- Xinggong Zhang (17 papers)
- Yong Cui (29 papers)