Papers
Topics
Authors
Recent
Search
2000 character limit reached

Fleet of Agents: Coordinated Problem Solving with Large Language Models

Published 7 May 2024 in cs.CL, cs.AI, cs.LG, and cs.NE | (2405.06691v3)

Abstract: While numerous frameworks have been developed to enhance the reasoning abilities of LLMs, there is a scarcity of methods that effectively balance the trade-off between cost and quality. In this paper, we introduce Fleet of Agents (FoA), a novel and intuitive yet principled framework utilizing LLMs as agents to navigate through dynamic tree searches, employing a genetic-type particle filtering approach. FoA spawns a multitude of agents, each exploring the search space autonomously, followed by a selection phase where resampling based on a heuristic value function optimizes the balance between exploration and exploitation. This mechanism enables dynamic branching, adapting the exploration strategy based on discovered solutions. We conduct extensive experiments on three benchmark tasks, Game of 24'',Mini-Crosswords'', and WebShop'', utilizing four different LLMs,GPT-3.5'', GPT-4'',LLaMA3.2-11B'', and ``LLaMA3.2-90B''. On average across all tasks and LLMs, FoA obtains a quality improvement of ~5% while requiring only ~40% of the cost of previous SOTA methods. Notably, our analyses reveal that (1) FoA achieves the best cost-quality trade-off among all benchmarked methods and (2) FoA + LLaMA3.2-11B surpasses the Llama3.2-90B model. FoA is publicly available at https://github.com/au-clan/FoA.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Graph of thoughts: Solving elaborate problems with large language models. In Proceedings of the AAAI Conference on Artificial Intelligence (2024), vol. 38, pp. 17682–17690.
  2. Language models are few-shot learners. In Advances in Neural Information Processing Systems (2020), H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin, Eds., vol. 33, Curran Associates, Inc., pp. 1877–1901.
  3. Chase, H. Langchain. https://github.com/hwchase17/langchain, 2022.
  4. Daniel, K. Thinking, fast and slow. 2017.
  5. Evolutionary algorithms in engineering applications. Springer Science & Business Media, 2013.
  6. Holland, J. H. Adaptation in natural and artificial systems: an introductory analysis with applications to biology, control, and artificial intelligence. MIT press, 1992.
  7. Metagpt: Meta programming for multi-agent collaborative framework. CoRR abs/2308.00352 (2023).
  8. Flows: Building blocks of reasoning and collaborating ai. arXiv preprint arXiv:2308.01285 (2023).
  9. Large language models are zero-shot reasoners. In Advances in Neural Information Processing Systems (2022), S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, Eds., vol. 35, Curran Associates, Inc., pp. 22199–22213.
  10. Positioning error compensation of an industrial robot using neural networks and experimental study. Chinese Journal of Aeronautics 35, 2 (2022), 346–360. https://www.sciencedirect.com/science/article/abs/pii/S1000936121001187 ; https://www.sciencedirect.com/science/article/pii/S1000936121001187.
  11. Camel: Communicative agents for" mind" exploration of large scale language model society. arXiv preprint arXiv:2303.17760 (2023).
  12. More agents is all you need, 2024.
  13. Chameleon: Plug-and-play compositional reasoning with large language models. ArXiv abs/2304.09842 (2023).
  14. Self-refine: Iterative refinement with self-feedback. arXiv preprint arXiv:2303.17651 (2023).
  15. A hybrid genetic - particle swarm optimization algorithm for the vehicle routing problem. Expert Systems with Applications 37, 2 (2010), 1446–1455. https://dblp.uni-trier.de/db/journals/eswa/eswa37.html#MarinakisM10 ; https://www.sciencedirect.com/science/article/pii/S0957417409006460 ; https://dl.acm.org/doi/10.1016/j.eswa.2009.06.085.
  16. Explaining machine learning classifiers through diverse counterfactual explanations. In Proceedings of the 2020 conference on fairness, accountability, and transparency (2020), pp. 607–617.
  17. Nakajima, Y. Babyagi. https://github.com/yoheinakajima/babyagi, 2023.
  18. Show your work: Scratchpads for intermediate computation with language models. CoRR abs/2112.00114 (2021).
  19. Refiner: Reasoning feedback on intermediate representations. arXiv preprint arXiv:2304.01904 (2023).
  20. Richards, T. B. Autogpt. https://github.com/Significant-Gravitas/Auto-GPT, 2023.
  21. Dynamic optimization of fuzzy cognitive maps for time series forecasting. Knowledge-Based Systems 105 (2016), 29–37. https://dl.acm.org/doi/10.1016/j.knosys.2016.04.023 ; https://dblp.uni-trier.de/db/journals/kbs/kbs105.html#SalmeronF16 ; https://dx.doi.org/10.1016/j.knosys.2016.04.023 ; http://dx.doi.org/10.1016/j.knosys.2016.04.023 ; https://www.sciencedirect.com/science/article/pii/S0950705116300752.
  22. Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. ArXiv abs/2303.17580 (2023).
  23. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171 (2022).
  24. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.
  25. The multidimensional wisdom of crowds. Advances in neural information processing systems 23 (2010).
  26. Autogen: Enabling next-gen LLM applications via multi-agent conversation framework. CoRR abs/2308.08155 (2023).
  27. Tree of thoughts: Deliberate problem solving with large language models. Advances in Neural Information Processing Systems 36 (2024).
  28. React: Synergizing reasoning and acting in language models. In The Eleventh International Conference on Learning Representations (2023).

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 91 likes about this paper.