

Noah Lee
41 posts

@nlee288
MS Student @kaist_ai Interested in LLM, Human Alignment



👀 How to find a better adapted model? ✨ Let the models find it for you! 👉🏻 Introducing Model Swarms, multiple LLM experts collaboratively search for new adapted models in the weight space and discover their new capabilities. 📄 Paper: arxiv.org/abs/2410.11163

🤔How can we systematically assess an LM's proficiency in a specific capability without using summary measures like helpfulness or simple proxy tasks like multiple-choice QA? Introducing the ✨BiGGen Bench, a benchmark that directly evaluates nine core capabilities of LMs.





🤔How can we systematically assess an LM's proficiency in a specific capability without using summary measures like helpfulness or simple proxy tasks like multiple-choice QA? Introducing the ✨BiGGen Bench, a benchmark that directly evaluates nine core capabilities of LMs.

At the upcoming @emnlpmeeting, I will be presenting the ORPO paper with @nlee288 at Miami! 🗓️Nov 14th, 10:30 - 12:00 📍Session F, Riverfront Hall 🔥Excited to meet and chat about RLHF & alignment in LLMs, reward modeling, and diverse applications of alignment methods!🧵



Align LLMs with the preference dataset ONLY with 💡ORPO💡 We introduce ORPO, alignment without reference model & SFT! With awesome dataset from @argilla_io + Mistral(7B) + ORPO, we present 🌟Mistral-ORPO-β🌟 🧵 👉 AlpacaEval 2.0: 12.20% 👉 IFEval: 66.19% 👉 MT-Bench: 7.32





📢New model, Mistral-ORPO-Capybara-7k in ORPO collection!🧵 With 💡ORPO💡 + 7k Capybara preference pair by @argilla_io🔥 + Mistral (7B), you can get the human-aligned chat model within 2.5 hours of fine-tuning👀 👉AlpacaEval 2.0 (LC): 15.9% 👉MT-Bench: 7.44 👉IFEval: 61.27%




Welcome Zephyr 141B to Hugging Chat🔥 🎉A Mixtral-8x22B fine-tune ⚡️Super fast generation with TGI 🤗Fully open source (from the data to the UI) huggingface.co/chat/models/Hu…

The new Mixtral-8x22B base model is a total beast for fine-tuning and has produced some of the highest scores I've ever seen on IFEval and BBH 🤯 We teamed up with @argilla_io and @kaist_ai to cook up a brand new recipe for Zephyr models 🪁 huggingface.co/HuggingFaceH4/… 🧑🍳 Align the base model with Odds Ratio Preference Optimisation (ORPO). This novel algorithm does not require an SFT step to achieve high performance and is thus much more computationally efficient than methods like DPO and PPO 🦫 Use a brand new dataset of 7k high-quality, multi-turn preferences that has been developed by our friends at @argilla_io: huggingface.co/datasets/argil… As usual, we are open sourcing the training code in the Alignment Handbook for the community to build on: github.com/huggingface/al… This has been a epic speed run with @jiwoohong98 @nlee288 @alvarobartt - now I can finally sleep 😂


