AutoAct: Automatic Agent Learning from Scratch for QA via Self-Planning
Abstract: Language agents have achieved considerable performance on various complex question-answering tasks by planning with external tools. Despite the incessant exploration in this field, existing language agent systems still struggle with costly, non-reproducible data reliance and face the challenge of compelling a single model for multiple functions. To this end, we introduce AutoAct, an automatic agent learning framework for QA that does not rely on large-scale annotated data and synthetic planning trajectories from closed-source models (e.g., GPT-4). Given limited data with a tool library, AutoAct first automatically synthesizes planning trajectories without any assistance from humans or strong closed-source models. Then, AutoAct leverages a division-of-labor strategy to automatically differentiate based on the target task information and synthesized trajectories, producing a sub-agent group to complete the task. We conduct comprehensive experiments with different LLMs, which demonstrates that AutoAct yields better or parallel performance compared to various strong baselines. Further analysis demonstrates the effectiveness of the division-of-labor strategy, with the trajectory quality generated by AutoAct generally outperforming that of others. Code will be available at https://github.com/zjunlp/AutoAct.
- Rest meets react: Self-improvement for multi-step reasoning llm agent.
- Fireact: Toward language agent fine-tuning. CoRR, abs/2310.05915.
- Reconcile: Round-table conference improves reasoning via consensus among diverse llms. CoRR, abs/2309.13007.
- Agentverse: Facilitating multi-agent collaboration and exploring emergent behaviors in agents. CoRR, abs/2308.10848.
- Alan Colman. 2008. Human embryonic stem cells and clinical applications. Cell Research, 18(1):S171–S171.
- Specializing smaller language models towards multi-step reasoning. In International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pages 10421–10430. PMLR.
- C. A. E. Goodhart. 1984. Problems of Monetary Management: The UK Experience, pages 91–121. Macmillan Education UK, London.
- Reinforced self-training (rest) for language modeling. CoRR, abs/2308.08998.
- Lora: Low-rank adaptation of large language models. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net.
- Large language models can self-improve. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023, pages 1051–1068. Association for Computational Linguistics.
- Do as I can, not as I say: Grounding language in robotic affordances. In Conference on Robot Learning, CoRL 2022, 14-18 December 2022, Auckland, New Zealand, volume 205 of Proceedings of Machine Learning Research, pages 287–318. PMLR.
- Self-alignment with instruction backtranslation. CoRR, abs/2308.06259.
- Encouraging divergent thinking in large language models through multi-agent debate. CoRR, abs/2305.19118.
- Generated knowledge prompting for commonsense reasoning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, pages 3154–3169. Association for Computational Linguistics.
- BOLAA: benchmarking and orchestrating llm-augmented autonomous agents. CoRR, abs/2308.05960.
- Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net.
- Learn to explain: Multimodal reasoning via thought chains for science question answering. In NeurIPS.
- Chameleon: Plug-and-play compositional reasoning with large language models. CoRR, abs/2304.09842.
- Self-refine: Iterative refinement with self-feedback. CoRR, abs/2303.17651.
- Editing personality for llms. CoRR, abs/2310.02168.
- Michael Mintrom. 2015. 12Herbert A. Simon, Administrative Behavior: A Study of Decision-Making Processes in Administrative Organization. In The Oxford Handbook of Classics in Public Policy and Administration. Oxford University Press.
- Yohei Nakajima. 2023. Babyagi. https://github.com/yoheinakajima/babyagi.
- OpenAI. 2022. Chatgpt: Optimizing language models for dialogue. https://openai.com/blog/chatgpt/.
- OpenAI. 2023. GPT-4 technical report. CoRR, abs/2303.08774.
- Anton Osika. 2023. Gpt-engineer. https://github.com/AntonOsika/gpt-engineer.
- Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, UIST 2023, San Francisco, CA, USA, 29 October 2023- 1 November 2023, pages 2:1–2:22. ACM.
- Gorilla: Large language model connected with massive apis. CoRR, abs/2305.15334.
- Virtualhome: Simulating household activities via programs. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pages 8494–8502. Computer Vision Foundation / IEEE Computer Society.
- Making language models better tool learners with execution feedback. CoRR, abs/2305.13068.
- Reasoning with language model prompting: A survey. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023, pages 5368–5393. Association for Computational Linguistics.
- Toolllm: Facilitating large language models to master 16000+ real-world apis. CoRR, abs/2307.16789.
- Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters. In KDD ’20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, August 23-27, 2020, pages 3505–3506. ACM.
- Hugginggpt: Solving AI tasks with chatgpt and its friends in huggingface. CoRR, abs/2303.17580.
- Reflexion: language agents with verbal reinforcement learning. CoRR, abs/2303.11366.
- Alfworld: Aligning text and embodied environments for interactive learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net.
- Llm-planner: Few-shot grounded planning for embodied agents with large language models. CoRR, abs/2212.04088.
- Medagents: Large language models as collaborators for zero-shot medical reasoning. CoRR, abs/2311.10537.
- Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca.
- Torantulino. 2023. Autogpt: build & use ai agents. https://github.com/Significant-Gravitas.
- Llama 2: Open foundation and fine-tuned chat models. CoRR, abs/2307.09288.
- Voyager: An open-ended embodied agent with large language models. CoRR, abs/2305.16291.
- A survey on large language model based autonomous agents. CoRR, abs/2308.11432.
- Self-instruct: Aligning language models with self-generated instructions. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023, pages 13484–13508. Association for Computational Linguistics.
- Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS.
- The rise and potential of large language model based agents: A survey. CoRR, abs/2309.07864.
- Wizardlm: Empowering large language models to follow complex instructions. CoRR, abs/2304.12244.
- Hotpotqa: A dataset for diverse, explainable multi-hop question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, pages 2369–2380. Association for Computational Linguistics.
- Webshop: Towards scalable real-world web interaction with grounded language agents. In NeurIPS.
- React: Synergizing reasoning and acting in language models. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net.
- Lumos: Learning agents with unified data, modular design, and open-source llms. CoRR, abs/2311.05657.
- Large language model as attributed training data generator: A tale of diversity and bias. CoRR, abs/2306.15895.
- Star: Bootstrapping reasoning with reasoning. In NeurIPS.
- Agenttuning: Enabling generalized agent abilities for llms. CoRR, abs/2310.12823.
- Exploring collaboration mechanisms for LLM agents: A social psychology view. CoRR, abs/2310.02124.
- Igniting language intelligence: The hitchhiker’s guide from chain-of-thought reasoning to language agents. CoRR, abs/2311.11797.
- Judging llm-as-a-judge with mt-bench and chatbot arena.
- Least-to-most prompting enables complex reasoning in large language models. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net.
- Webarena: A realistic web environment for building autonomous agents. CoRR, abs/2307.13854.
- Agents: An open-source framework for autonomous language agents. CoRR, abs/2309.07870.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.