WizardLM

358 posts

WizardLM

WizardLM

@WizardLM_AI

WizardLM, WizardCoder, WizardMath. Evol-Instruct, Arena Learning, RLEIF.

USA Katılım Ekim 2021
680 Takip Edilen12.3K Takipçiler
Sabitlenmiş Tweet
WizardLM
WizardLM@WizardLM_AI·
🎉Today we are announcing Evol-Instruct V2 !!! 🔥 Auto Evol-Instruct is one of the most important technologies for WizardLM-2. Paper link: arxiv.org/pdf/2406.00770 We build a fully automated Evol-Instruct pipeline, allowing WizardLM-2 to be extended from three evolved domains (chat, code and math) of WizardLM-1 to dozens of evolved domains 🚀With Auto Evol-Instruct, You can create high-quality, highly complex instruction tuning data for any task without the need for human efforts! ⚖️We hope that this universal technology can promote fairness and efficiency for all the AI researchers in training and evaluation their own large language models. 👉For more details, please refer to Can Xu's channel:
Can Xu@CanXu20

🔥 Excited to share the other key Technology of WizardLM-2! 📙AutoEvol: Automatic Instruction Evolving for Large Language Models 🚀We build a fully automated Evol-Instruct pipeline to create high-quality, highly complex instruction tuning data: -------- 🧵 -------- 👉Motivation First: Over the past six months, we have dedicated ourselves to exploring methods to scale up synthetic training for LLMs. Although Evol-Instruct has demonstrated excellent performance in creating powerful post-training data, it relies too heavily on the efforts of human experts to design specific evolutionary methods for specific tasks. Once Evol-Instruct is applied to an entirely new complex task, the methods for executing evolution need to be redesigned. This limitation of Evol-Instruct makes scaling up extremely challenging, prompting us to develop a new method, 💻Auto Evol-Insturct💻, that can evolve instruction data automatically. Auto Evol allows the training of WizardLM2 to be conducted with nearly an unlimited number and variety of synthetic data. Let's see: 🧐 1. Limitations of Evol-Instruct: Evol-Instruct takes the high-quality data as a starting point, and further iteratively refines it using LLMs, improving its complexity and diversity. It has demonstrated superior performance across a broad range of public benchmarks that evaluate diverse capabilities, including instruction following (WizardLM), code generation (WizardCoder), and mathematical reasoning (WizardMath). While Evol-Instruct exhibits outstanding performance, its heavy reliance on heuristic efforts presents notable challenges. Whenever it is used for a completely new task, the methods for execution evolution need to be redesigned. Such a process requires a high level of expertise and considerable costs, hindering its adaptation to a wider spectrum of capabilities. 2. We want to build a fully automated Evol-Instruct pipeline Auto Evol-Instruct automatically designs evolving methods that make given instruction data more complex, enabling almost cost-free adaptation to different tasks by only changing the input data of the framework. From below figure, we can see the iterative process of optimizing the initial evolving method e0 into the optimal evolving method e∗, which specifically outlines the transition from et−1 to et. We refer to the model used for evolution as the evol LLM, and the model used for optimization as the optimizer LLM. This optimization process involves two critical stages: (1) Evol Trajectory Analysis: The optimizer LLM carefully analyzes the potential issues and failures exposed in instruction evolution performed by evol LLM, generating feedback for subsequent optimization. (2) Evolving Method Optimization: The optimizer LLM optimizes the evolving method by addressing these identified issues in feedback. These stages alternate and repeat to progressively develop an effective evolving method using only a subset of the instruction data. Once the optimal evolving method is identified, it directs the evol LLM to convert the entire instruction dataset into more diverse and complex forms, thus facilitating improved instruction tuning. 3. Fully AI-driven Evol-Instruct can outperform the Evol-Instruct used by human experts. Our experiments show that the evolving methods designed by Auto Evol-Instruct outperform the Evol-Instruct methods designed by human experts in instruction tuning across various capabilities, including instruction following, mathematical reasoning, and code generation. As shown in the below table, on the instruction following task, Auto Evol-Instruct can achieve a improvement of 10.44% over the Evol method used by WizardLM-1 on MT-bench; on the code task HumanEval, it can achieve a 12% improvement over the method used by WizardCoder; on the math task GSM8k, it can achieve a 6.9% improvement over the method used by WizardMath. 4. Scaling Evol-Instruct to various domains and tasks With the new technology of Auto Evol-Instruct, the evolutionary synthesis data of WizardLM-2 has scaled up from the three domains of chat, code, and math in WizardLM-1 to dozens of domains, covering tasks in all aspects of large language models. This allows Arena Learning to train and learn from an almost infinite pool of high-difficulty instruction data, fully unlocking all the potential of Arena Learning. For more details, please refer to: Paper: arxiv.org/pdf/2406.00770 Project: github.com/nlpxucan/Wizar… We are working with our legal team to publicly release the code of Auto Evol-Instruct.

English
11
44
226
40K
Eric Hartford
Eric Hartford@QuixiAI·
My first popular model was WizardLM-Uncensored, which evolved into Dolphin. @WizardLM_AI is the goat🐐
Newport, SC 🇺🇸 English
1
0
23
1.3K
Chris 🇨🇦
Chris 🇨🇦@llm_wizard·
@WizardLM_AI As my name is "LLM Wizard", you know this is my favourite series of models. Extremely happy to see them coming back!
English
1
0
1
93
Maziyar PANAHI
Maziyar PANAHI@MaziyarPanahi·
@WizardLM_AI Wow! I am not even gonna pretend I am sad! 😂 You guys are gonna SHIP so much! 🔥
GIF
English
1
0
3
169
Pranav :-
Pranav :-@Pranav2278·
@WizardLM_AI Congratulations!!! Can't wait for you all to be so back!
English
1
0
1
140
Cody
Cody@HoldMyData·
@WizardLM_AI Congrats!! Thank you for the memories and good luck with Tencent!
English
1
0
2
117
Harrison Kinsley
Harrison Kinsley@Sentdex·
@WizardLM_AI Whoa, looking forward to this team being able to release stuff again. Congrats on the move team!
English
1
0
13
493
Rishabh Srivastava
Rishabh Srivastava@rishdotblog·
@WizardLM_AI End of an era. Thanks for publishing so much of your work — evol instruct was a revelation back in the day!
English
1
0
2
287
xlr8harder
xlr8harder@xlr8harder·
@WizardLM_AI Congrats, and nice job staying classy. Looking forward to seeing future releases.
English
1
0
5
354
Alex Volkov
Alex Volkov@altryne·
@WizardLM_AI Whoah! Congrats on finding a new home that will let you ship! 👏
English
1
0
7
355
WizardLM retweetledi
Mengkang Hu
Mengkang Hu@aaron_mkhu·
🎉 Thrilled to share our paper accepted by #KDD2025! 🌟AgentGen🌟: An automated environment and task generator that enhances LLM-based agents' planning abilities through diverse, difficulty-controlled synthetic trajectory data. 👇🏻agent-gen.github.io
English
2
2
14
2.7K
WizardLM
WizardLM@WizardLM_AI·
🚀New approach from WaveCoder Team for optimizing code LLMs. The novel feature tree based framework, inspired by AST and Evol-Instruct to modeling semantic relationships, generates more diverse data. The EpiCoder hits SOTA in both challenge file and function benchmarks.
Wavecoder@TeamCodeLLM_AI

🚀 Introducing EpiCoder: a hierarchical feature tree-based framework for diverse and intricate code generation. 🔍 Outperforming benchmarks, it handles everything from simple functions to multi-file projects deftly. 📢 Open source release soon! 🔗 arxiv.org/abs/2501.04694

English
5
5
69
7.9K
WizardLM retweetledi
kaeru
kaeru@mryo39·
GENIAC phase2にて、日本語のローカルLLMを使ってEvol-Instructによるデータセット構築に取り組んだ際の記事を公開しました。(3件目/全4件) zenn.dev/matsuolab/arti…
日本語
0
13
45
7K
WizardLM retweetledi
elvis
elvis@omarsar0·
Agentic Information Retrieval This paper provides a good introduction to agentic information retrieval, which is shaped by the capabilities of LLM agents. I've been developing with this paradigm recently and it does offer lots of interesting ways to optimize retrieval systems.
elvis tweet media
English
11
112
700
63.4K