āMastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Modelsā, 2024-03-13 ()ā :
Underlying data distributions of natural language, programming code, and mathematical symbols vary vastly, presenting a complex challenge for large language models (LLMs) that strive to achieve high performance across all 3 domains simultaneously. Achieving a very high level of proficiency for an LLM within a specific domain often requires extensive training with relevant corpora, which is typically accompanied by a sacrifice in performance in other domains.
In this paper, we propose to fuse models that are already highly-specialized directly. The proposed fusing framework, UltraFuser, consists of 3 distinct specialists that are already sufficiently trained on language, coding, and mathematics. A token-level gating mechanism is introduced to blend the specialistsā outputs. A two-stage training strategy accompanied by balanced sampling is designed to ensure stability.
To effectively train the fused model, we further construct a high-quality supervised instruction tuning dataset, UltraChat 2, which includes text, code, and mathematical content. This dataset comprises ~300,000 instructions and covers a wide range of topics in each domain.
Experiments show that our model could simultaneously achieve mastery of the 3 crucial domains.