Instruction Fusion: Advancing Prompt Evolution through Hybridization
Abstract: The fine-tuning of LLMs specialized in code generation has seen notable advancements through the use of open-domain coding queries. Despite the successes, existing methodologies like Evol-Instruct encounter performance limitations, impeding further enhancements in code generation tasks. This paper examines the constraints of existing prompt evolution techniques and introduces a novel approach, Instruction Fusion (IF). IF innovatively combines two distinct prompts through a hybridization process, thereby enhancing the evolution of training prompts for code LLMs. Our experimental results reveal that the proposed novel method effectively addresses the shortcomings of prior methods, significantly improving the performance of Code LLMs across five code generation benchmarks, namely HumanEval, HumanEval+, MBPP, MBPP+ and MultiPL-E, which underscore the effectiveness of Instruction Fusion in advancing the capabilities of LLMs in code generation.
- Program synthesis with large language models. arXiv preprint arXiv:2108.07732.
- A framework for the evaluation of code generation models. https://github.com/bigcode-project/bigcode-evaluation-harness.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
- Multipl-e: A scalable and extensible approach to benchmarking neural code generation.
- Sahil Chaudhary. 2023. Code alpaca: An instruction-following llama model for code generation. https://github.com/sahil280114/codealpaca.
- Evaluating large language models trained on code.
- DeepSeek. 2023. Deepseek coder: Let the code write itself. https://github.com/deepseek-ai/DeepSeek-Coder.
- Code generation using machine learning: A systematic review. Ieee Access.
- Large language models for software engineering: A systematic literature review. arXiv preprint arXiv:2308.10620.
- Mistral 7b.
- Active instruction tuning: Improving cross-task generalization by training on prompt sensitive tasks. arXiv preprint arXiv:2311.00288.
- Starcoder: may the source be with you! arXiv preprint arXiv:2305.06161.
- Is your code generated by chatgpt really correct? rigorous evaluation of large language models for code generation. arXiv preprint arXiv:2305.01210.
- Summary of chatgpt-related research and perspective towards the future of large language models. Meta-Radiology, 1(2):100017.
- Wizardcoder: Empowering code large language models with evol-instruct. arXiv preprint arXiv:2306.08568.
- A conversational paradigm for program synthesis. arXiv preprint.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- Code llama: Open foundation models for code. arXiv preprint arXiv:2308.12950.
- Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
- Richard J Waldinger and Richard CT Lee. 1969. Prow: A step toward automatic program writing. In Proceedings of the 1st international joint conference on Artificial intelligence, pages 241–252.
- Self-instruct: Aligning language model with self generated instructions. arXiv preprint arXiv:2212.10560.
- Codet5: Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. arXiv preprint arXiv:2109.00859.
- Finetuned language models are zero-shot learners.
- Magicoder: Source code is all you need.
- Wizardlm: Empowering large language models to follow complex instructions.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.