Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 48 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 19 tok/s Pro
GPT-4o 107 tok/s Pro
Kimi K2 205 tok/s Pro
GPT OSS 120B 473 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

Large Language Model based Multi-Agents: A Survey of Progress and Challenges (2402.01680v2)

Published 21 Jan 2024 in cs.CL, cs.AI, and cs.MA

Abstract: LLMs have achieved remarkable success across a wide array of tasks. Due to the impressive planning and reasoning abilities of LLMs, they have been used as autonomous agents to do many tasks automatically. Recently, based on the development of using one LLM as a single planning or decision-making agent, LLM-based multi-agent systems have achieved considerable progress in complex problem-solving and world simulation. To provide the community with an overview of this dynamic field, we present this survey to offer an in-depth discussion on the essential aspects of multi-agent systems based on LLMs, as well as the challenges. Our goal is for readers to gain substantial insights on the following questions: What domains and environments do LLM-based multi-agents simulate? How are these agents profiled and how do they communicate? What mechanisms contribute to the growth of agents' capacities? For those interested in delving into this field of study, we also summarize the commonly used datasets or benchmarks for them to have convenient access. To keep researchers updated on the latest studies, we maintain an open-source GitHub repository, dedicated to outlining the research on LLM-based multi-agent systems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (88)
  1. Evaluating multi-agent coordination abilities in large language models, 2023.
  2. Using large language models to simulate multiple humans and replicate human subject studies, 2023.
  3. Playing repeated games with large language models. arXiv preprint arXiv:2305.16867, 2023.
  4. Anonymous. Rethinking the buyer’s inspection paradox in information markets with language agents. In Submitted to The Twelfth International Conference on Learning Representations, 2023. under review.
  5. Chateval: Towards better llm-based evaluators through multi-agent debate, 2023.
  6. Autoagents: A framework for automatic agent generation. arXiv preprint arXiv:2309.17288, 2023.
  7. Multi-agent consensus seeking via large language models. arXiv preprint arXiv:2310.20151, 2023.
  8. Agentverse: Facilitating multi-agent collaboration and exploring emergent behaviors in agents. arXiv preprint arXiv:2308.10848, 2023.
  9. Scalable multi-robot collaboration with large language models: Centralized or decentralized systems? arXiv preprint arXiv:2309.15943, 2023.
  10. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168, 2021.
  11. Collaborating with language models for embodied reasoning. arXiv preprint arXiv:2302.00763, 2023.
  12. Victor Dibia. Multi-agent llm applications — a review of current research, tools, and challenges. https://newsletter.victordibia.com/p/multi-agent-llm-applications-a-review, 2023.
  13. A survey on in-context learning, 2023.
  14. Self-collaboration code generation via chatgpt, 2023.
  15. Improving factuality and reasoning in language models through multiagent debate, 2023.
  16. Can large language models serve as rational players in game theory? a systematic analysis. arXiv preprint arXiv:2312.05488, 2023.
  17. Agent-Based Modeling in Economics and Finance: Past, Present, and Future. INET Oxford Working Papers 2022-10, Institute for New Economic Thinking at the Oxford Martin School, University of Oxford, June 2022.
  18. S33{}^{3}start_FLOATSUPERSCRIPT 3 end_FLOATSUPERSCRIPT: Social-network simulation system with large language model-empowered agents. arXiv preprint arXiv:2307.14984, 2023.
  19. Retrieval-augmented generation for large language models: A survey. arXiv preprint arXiv:2312.10997, 2023.
  20. Did aristotle use a laptop? a question answering benchmark with implicit reasoning strategies, 2021.
  21. Generative agent-based modeling: Unveiling social system dynamics through coupling mechanistic models with generative artificial intelligence. arXiv preprint arXiv:2309.11456, 2023.
  22. What indeed can gpt models do in chemistry? a comprehensive benchmark on eight tasks. arXiv preprint arXiv:2305.18365, 2023.
  23. Measuring massive multitask language understanding. arXiv preprint arXiv:2009.03300, 2020.
  24. Metagpt: Meta programming for multi-agent collaborative framework. arXiv preprint arXiv:2308.00352, 2023.
  25. John J Horton. Large language models as simulated economic agents: What can we learn from homo silicus? Technical report, National Bureau of Economic Research, 2023.
  26. War and peace (waragent): Large language model-based multi-agent simulation of world wars, 2023.
  27. Agentcoder: Multi-agent-based code generation with iterative testing and optimisation, 2023.
  28. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. arXiv preprint arXiv:2311.05232, 2023.
  29. Lyfe agents: Generative agents for low-cost real-time social interactions. arXiv preprint arXiv:2310.02172, 2023.
  30. Decomposed prompting: A modular approach for solving complex tasks, 2023.
  31. The socialai school: Insights from developmental psychology towards artificial socio-cultural agents. arXiv preprint arXiv:2307.07871, 2023.
  32. Retrieval-augmented generation for knowledge-intensive nlp tasks, 2021.
  33. Psychology-informed recommender systems: A human-centric perspective on recommender systems. In Proceedings of the 2022 Conference on Human Information Interaction and Retrieval, CHIIR ’22, page 367–368, New York, NY, USA, 2022. Association for Computing Machinery.
  34. Quantifying the impact of large language models on collective opinion dynamics. arXiv preprint arXiv:2308.03313, 2023.
  35. Camel: Communicative agents for” mind” exploration of large scale language model society. arXiv preprint arXiv:2303.17760, 2023.
  36. Theory of mind for multi-agent collaboration via large language models, 2023.
  37. Api-bank: A comprehensive benchmark for tool-augmented llms, 2023.
  38. Large language model-empowered agents for simulating macroeconomic activities, 2023.
  39. Are you in a masquerade? exploring the behavior and impact of large language model driven social bots in online social networks. arXiv preprint arXiv:2307.10337, 2023.
  40. Tradinggpt: Multi-agent system with layered memory and distinct characters for enhanced financial trading performance, 2023.
  41. Metaagents: Simulating interactions of human behaviors for llm-based task-oriented coordination via collaborative generative agents. arXiv preprint arXiv:2310.06500, 2023.
  42. Let gpt be a math tutor: Teaching math word problem solvers with customized exercise generation. arXiv preprint arXiv:2305.14386, 2023.
  43. Avalonbench: Evaluating llms playing the game of avalon, 2023.
  44. From text to tactic: Evaluating llms playing the game of avalon. arXiv preprint arXiv:2310.05036, 2023.
  45. Dynamic llm-agent network: An llm-agent collaboration framework with agent team optimization. arXiv preprint arXiv:2310.02170, 2023.
  46. Understanding the benefits and challenges of using large language model-based conversational agents for mental well-being support. arXiv preprint arXiv:2307.15810, 2023.
  47. Roco: Dialectic multi-robot collaboration with large language models. arXiv preprint arXiv:2307.04738, 2023.
  48. Alympics: Language agents meet game theory. arXiv preprint arXiv:2311.03220, 2023.
  49. João Moura. Crewai. https://github.com/joaomdmoura/crewAI, 2023.
  50. Welfare diplomacy: Benchmarking language model cooperation. arXiv preprint arXiv:2310.08901, 2023.
  51. Self-adaptive large language model (llm)-based multiagent systems. In 2023 IEEE International Conference on Autonomic Computing and Self-Organizing Systems Companion (ACSOS-C), pages 104–109. IEEE, 2023.
  52. Social simulacra: Creating populated prototypes for social computing systems. In Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology, pages 1–18, 2022.
  53. Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442, 2023.
  54. Communicative agents for software development, 2023.
  55. Tptu: Large language model-based ai agents for task planning and tool usage, 2023.
  56. Artificial Intelligence: A Modern Approach. Prentice Hall Press, USA, 3rd edition, 2009.
  57. Reflexion: Language agents with verbal reinforcement learning, 2023.
  58. Cognitive architectures for language agents. arXiv preprint arXiv:2309.02427, 2023.
  59. Medagents: Large language models as collaborators for zero-shot medical reasoning, 2023.
  60. Putting humans in the natural language processing loop: A survey, 2021.
  61. Adapting llm agents through communication, 2023.
  62. A survey on large language model based autonomous agents, 2023.
  63. Avalon’s game of thoughts: Battle against deception through recursive contemplation. arXiv preprint arXiv:2310.01320, 2023.
  64. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
  65. Lilian Weng. Llm powered autonomous agents. https://lilianweng.github.io/posts/2023-06-23-agent/, 2023.
  66. Epidemic modeling with generative agents. arXiv preprint arXiv:2307.04986, 2023.
  67. Intelligent agents: theory and practice. The Knowledge Engineering Review, 10:115 – 152, 1995.
  68. Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155, 2023.
  69. Chatarena: Multi-agent language game environments for large language models. GitHub repository, 2023.
  70. The rise and potential of large language model based agents: A survey, 2023.
  71. Simulating public administration crisis: A novel generative agent-based simulation system to lower technology barriers in social science research. arXiv preprint arXiv:2311.06957, 2023.
  72. Openagents: An open platform for language agents in the wild. arXiv preprint arXiv:2310.10634, 2023.
  73. Examining inter-consistency of large language models collaboration: An in-depth analysis via debate, 2023.
  74. Magic: Investigation of large language model powered multi-agent in cognition, adaptability, rationality and collaboration, 2023.
  75. Exploring large language models for communication games: An empirical study on werewolf. arXiv preprint arXiv:2309.04658, 2023.
  76. Language agents with reinforcement learning for strategic play in the werewolf game. arXiv preprint arXiv:2310.18940, 2023.
  77. Tree of thoughts: Deliberate problem solving with large language models, 2023.
  78. Co-navgpt: Multi-robot cooperative visual semantic navigation using large language models, 2023.
  79. On generative agents in recommendation, 2023.
  80. Proagent: Building proactive cooperative ai with large language models. arXiv preprint arXiv:2308.11339, 2023.
  81. Building cooperative embodied agents modularly with large language models. arXiv preprint arXiv:2307.02485, 2023.
  82. Exploring collaboration mechanisms for llm agents: A social psychology view, 2023.
  83. Agentcf: Collaborative learning with autonomous language agents for recommender systems, 2023.
  84. Competeai: Understanding the competition behaviors in large language model-based agents, 2023.
  85. Chatgpt research group for optimizing the crystallinity of mofs and cofs. ACS Central Science, 9(11):2161–2170, 2023.
  86. Agents: An open-source framework for autonomous language agents. arXiv preprint arXiv:2309.07870, 2023.
  87. Sotopia: Interactive evaluation for social intelligence in language agents, 2023.
  88. Can large language models transform computational social science? Computational Linguistics, pages 1–53, 2023.
Citations (129)

Summary

  • The paper presents a comprehensive review of LLM-based multi-agent systems, detailing the evolution from single-agent to collective configurations.
  • The paper details methodologies for agent profiling, communication structures, and feedback-driven capability acquisition across various environments.
  • The paper discusses practical applications in problem-solving and world simulations while addressing challenges like hallucination, scalability, and benchmarking.

LLM based Multi-Agents: A Survey of Progress and Challenges

LLMs have achieved significant advancements in autonomous agent tasks, notably in complex problem-solving and world simulations. This survey paper provides an in-depth analysis of LLM-based multi-agent systems (LLM-MAS), focusing on their development, applications, and challenges. It offers insight into domains these systems simulate, how they communicate, and the mechanisms that enhance agent capabilities. Additionally, this paper discusses commonly used datasets and benchmarks, addressing the challenges faced by researchers in this evolving field.

Introduction

LLMs exhibit human-like reasoning and planning capabilities, positioning them as effective autonomous agents in various contexts. This alignment with human expectations for decision-making and action-taking agents has driven the rapid development of LLM-based agents capable of understanding and generating human-like instructions. Survey papers systematically summarize the state of LLM-based agents, reflecting significant advancements.

Research in LLM-MAS has transitioned from single-agent systems to multi-agent configurations, leveraging collective intelligence and specialized agent profiles. Unlike single-agent systems, multi-agent environments simulate complex real-world settings, allowing for collaborative planning, discussion, and decision-making. These systems capitalize on LLMs' communication abilities and domain-wide knowledge, enabling them to specialize in specific tasks. The recent literature shows promising outcomes in employing LLM-MAS for scenarios like software development, multi-robot systems, societal simulations, and policy simulations. Figure 1

Figure 1: The rising trend in the research field of LLM-based Multi-Agents, illustrating growth across various categories.

This burgeoning field has attuned researchers from diverse backgrounds, broadening its impact. Despite remarkable progress, foundational reviews and comprehensive perspectives on the challenges and future of this discipline remain scarce.

Dissecting LLM-MA Systems

This section covers crucial aspects of multi-agent systems: agent-environment interfaces, profiling methods, communication structures, and capability acquisition strategies.

Agents-Environment Interface

Actors within LLM-MA systems interact through interfaces that define how they perceive their environment. These interactions vary from simulated Sandbox environments to real-world Physical spaces:

  • Sandbox: Facilitates free and experimental interactions within virtual settings. Common in game simulations and software development tasks.
  • Physical: Involves real-world environments where agents enact physical actions.
  • None: Lacks external interaction, emphasizing agent-to-agent communication.

Agents Profiling

Agent profiling defines roles within multi-agent systems, embodying specific characteristics and capabilities:

  • Pre-defined: Explicitly specified by system designers.
  • Model-Generated: Created through LLMs based on contextual requirements.
  • Data-Derived: Constructed from pre-existing datasets.

Agents Communication

Agent communication is pivotal for achieving collective intelligence. It comprises:

  • Paradigms: Cooperative, Debate, Competitive formats.
  • Structure: Layered, Decentralized, Centralized frameworks. Figure 2

    Figure 2: The Architecture of LLM-MA Systems.

  • Content: Typically textual, varies by application.

Agents Capabilities Acquisition

Agent capabilities evolve through feedback and strategic adjustment:

  • Feedback Sources: Environment, Agent interactions, Human, None.
  • Adjustments: Memory modules, Self-Evolution, Dynamic Generation. Figure 3

    Figure 3: The Agent Communication Structure in varies forms across systems.

Applications

LLM-MA systems serve two primary applications: problem-solving and world simulations.

Problem Solving

Utilized for complex issues like software development, robotics, and scientific experiments, multi-agent systems emulate specialized roles:

  • Software Development: Emulates distinct roles for collaboration.
  • Robotics/Embodied Agents: Multi-agent collaboration for physical tasks.
  • Science Experiments/Science Debate: Leverages human oversight for complex tasks.

World Simulation

Simulations span societal dynamics, gaming environments, economic models, and policy decision-making:

  • Society/Gaming/Economy: Explore interactive and predictive scenarios based on human-like interactions.
  • Psychology/Policy/Disease Propagation: Enable intricate analyses of social behaviors, policy impacts, and disease outbreaks.

Implementation Tools and Resources

Several frameworks facilitate building and studying multi-agent systems:

  • MetaGPT: Encodes workflow to mitigate hallucinations.
  • CAMEL: Facilitates agent cooperation via model prompting.
  • AutoGen: Offers customization for diverse applications.

Challenges and Opportunities

Research in LLM-MAS presents challenges and opportunities, needing innovative solutions:

Hallucination and Collective Intelligence

Agents must address hallucination risks and leverage collective intelligence for enhanced cooperation.

Multi-Modal Integration and Scalability

Developing multi-modal systems and scaling complex LLM-MAS frameworks present significant research potential.

Evaluation and Benchmarking

Current benchmarks are limited, necessitating comprehensive evaluation frameworks to assess LLM-MAS systems' capabilities.

Conclusion

The survey identifies key elements and applications of LLM-MAS, providing directions for future research. Enhanced frameworks and interdisciplinary exploration promise transformative impacts across domains. Continual research and development will lead to sophisticated applications for complex problem-solving and predictive simulations beyond current capabilities.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.