Papers
Topics
Authors
Recent
Search
2000 character limit reached

Introducing MAPO: Momentum-Aided Gradient Descent Prompt Optimization

Published 25 Oct 2024 in cs.CL | (2410.19499v3)

Abstract: Momentum-Aided Prompt Optimization (MAPO) enhances the efficiency and efficacy of prompt optimization for LLMs. Building on ProTeGi, MAPO uses positive natural language "gradients" and a momentum-based extension to refine prompts effectively. By tracking gradient history, MAPO avoids local minima and oscillations. It also utilizes beam search and an Upper Confidence Bound (UCB) algorithm for balanced candidate expansion and selection. Benchmark testing shows that MAPO achieves faster convergence time with fewer API calls and higher F1 scores than ProTeGi, proving it as a robust and scalable solution for automated prompt engineering in LLMs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. RLPrompt: Optimizing Discrete Text Prompts with Reinforcement Learning. arXiv.org.
  2. PromptBreeder: Self-Referential Self-Improvement via Prompt Evolution. arXiv.org.
  3. Prompt Optimization with Human Feedback. arXiv.org.
  4. Eureka: Human-Level Reward Design via Coding Large Language Models. arXiv.org.
  5. ETHOS: A Multi-Label Hate Speech Detection Dataset. Complex & Intelligent Systems, 8(6):4663–4678.
  6. OpenAI. 2022. Introducing ChatGPT. Accessed: 2024-10-19.
  7. Automatic Prompt Optimization with “Gradient Descent” and Beam Search. arXiv.org.
  8. A Systematic Survey of Prompt Engineering in Large Language Models: Techniques and Applications. arXiv.org.
  9. The Prompt Report: A Systematic Survey of Prompting Techniques. arXiv.org.
  10. AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts. arXiv.org.
  11. Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled Data. arXiv.org.
  12. William Yang Wang. 2017. “Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection. arXiv.org.
  13. Large Language Models as Optimizers. arXiv.org.
  14. Self-Taught Optimizer (STOP): Recursively Self-Improving Code Generation. arXiv.org.
  15. TEMPERA: Test-Time Prompting via Reinforcement Learning. arXiv.org.
  16. Large Language Models are Human-Level Prompt Engineers. arXiv.org.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.