Papers
Topics
Authors
Recent
Search
2000 character limit reached

YuLan: An Open-source Large Language Model

Published 28 Jun 2024 in cs.CL and cs.AI | (2406.19853v1)

Abstract: LLMs have become the foundation of many applications, leveraging their extensive capabilities in processing and understanding natural language. While many open-source LLMs have been released with technical reports, the lack of training details hinders further research and development. This paper presents the development of YuLan, a series of open-source LLMs with $12$ billion parameters. The base model of YuLan is pre-trained on approximately $1.7$T tokens derived from a diverse corpus, including massive English, Chinese, and multilingual texts. We design a three-stage pre-training method to enhance YuLan's overall capabilities. Subsequent phases of training incorporate instruction-tuning and human alignment, employing a substantial volume of high-quality synthesized data. To facilitate the learning of complex and long-tail knowledge, we devise a curriculum-learning framework throughout across these stages, which helps LLMs learn knowledge in an easy-to-hard manner. YuLan's training is finished on Jan, 2024 and has achieved performance on par with state-of-the-art LLMs across various English and Chinese benchmarks. This paper outlines a comprehensive technical roadmap for developing LLMs from scratch. Our model and codes are available at https://github.com/RUC-GSAI/YuLan-Chat.

Summary

  • The paper presents a novel three-stage pre-training strategy that enhances multilingual reasoning and factual knowledge retrieval.
  • The paper details the optimization of a Transformer-based architecture featuring rotary position embeddings, RMSNorm, and an optimized tokenizer for stability.
  • The paper demonstrates YuLan’s competitive performance on 22 benchmarks, highlighting its effective human alignment and instruction-tuning methods.

An Overview of the YuLan Open-source LLM

The paper "YuLan: An Open-source LLM" presents the development and evaluation of YuLan, a suite of open-source LLMs consisting of 12 billion parameters. This research was conducted by a team at Renmin University of China, focusing on enhancing the model's capabilities through a comprehensive pre-training and fine-tuning strategy.

Key Contributions

1. Pre-training on Multilingual Data

YuLan's foundational model is pre-trained on a vast dataset comprising approximately 1.7 trillion tokens. These tokens are derived from a diverse corpus that includes English, Chinese, and other multilingual texts. The data sources are extensive, including web pages, code repositories, encyclopedias, academic papers, and various domain-specific documents.

2. Three-stage Pre-training Method

The pre-training process is structured into three distinct stages:

  • Standard Pre-training: Utilizes a next-token prediction approach with a diverse dataset mix.
  • Capability-Enhanced Pre-training: Incorporates educational assessments to boost performance on complex benchmarks like MMLU and C-Eval.
  • Long-tail Knowledge-Aware Pre-training: Focuses on identifying and filling knowledge gaps by synthesizing targeted question-answer pairs.

3. Comprehensive Model Architecture and Optimization

YuLan implements a Transformer-based architecture similar to LLaMA, featuring enhancements like an optimized tokenizer, rotary position embeddings, and RMSNorm for stability. The training employs sophisticated strategies including data and tensor parallelism, enabling efficient use of 96 NVIDIA A800 GPUs.

4. Instruction-Tuning and Human Alignment

YuLan undergoes supervised fine-tuning (instruction-tuning) and further human alignment to adapt to human-like tasks:

  • Curriculum Instruction-Tuning: Gradually transitions from simple to complex tasks using a synthesized dataset of over 41 million instructions.
  • Human Alignment: Implements a difficulty-based curriculum leveraging a reward function to ensure better alignment with human preferences.

Evaluation and Results

YuLan is benchmarked across a range of tasks in areas such as commonsense reasoning, factual knowledge retrieval, reading comprehension, and mathematical reasoning. It demonstrates performance comparable to state-of-the-art LLMs on 22 public benchmarks.

  • Reasoning Tasks: Achieves significant accuracy on datasets like BoolQ and CommonsenseQA.
  • Language Understanding: Performs competitively on challenging datasets like MMLU and C-Eval, with robust capabilities in handling both English and Chinese.
  • Alignment: Excels in alignment benchmarks, showcasing effective human-aligned responses in both English and Chinese.

Implications and Future Directions

YuLan's development offers a valuable roadmap for building and improving large-scale LLMs. Its open-source nature and comprehensive training details provide a foundation for further research, potentially aiding advancements in AI capabilities. Future research could explore refinements in long-tail knowledge integration and cross-lingual transfer learning to enhance performance further.

The release of YuLan and its accessible technical report contributes to the ongoing discourse around transparency and reproducibility in AI, encouraging collaborative progress in the field of natural language processing.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 0 likes about this paper.