Yuning Mao

25 posts

Yuning Mao

Yuning Mao

@yuning_pro

TBD @AIatMeta, 🦙Post-training since Llama 2 https://t.co/OUjOWWA8kL

Moon 가입일 Haziran 2021
199 팔로잉307 팔로워
Yuning Mao 리트윗함
Yiqing Xie
Yiqing Xie@YiqingXieNLP·
Training on issue-solving only does NOT guarantee transfer to other tasks. 🎨Introducing Hybrid-Gym - synthetic training tasks for generalization (hybrid-gym.github.io) +25.4% on SWE-Bench / +7.9% on SWT-Bench / +5.1% on Commit-0 with NO issue-solving / test-gen/... training
Yiqing Xie tweet media
English
1
23
102
15.7K
Yuning Mao 리트윗함
Xianjun Yang
Xianjun Yang@xianjun_agi·
I was laid off by Meta today. As a Research Scientist, my work was just cited by the legendary @johnschulman2 and Nicholas Carlini yesterday. I’m actively looking for new opportunities — please reach out if you have any openings!
Xianjun Yang tweet media
Susan Zhang@suchenzang

👀

English
269
348
4.5K
1.8M
Yuning Mao 리트윗함
Tim Franzmeyer
Tim Franzmeyer@frtimlive·
What if LLMs knew when to stop? 🚧 HALT finetuning teaches LLMs to only generate content they’re confident is correct. 🔍 Insight: Post-training must be adjusted to the model’s capabilities. ⚖️ Tunable trade-off: Higher correctness 🔒 vs. More completeness 📝 with @AIatMeta 🧵
English
1
13
65
9.5K
Yuning Mao 리트윗함
Yiqing Xie
Yiqing Xie@YiqingXieNLP·
How to construct repo-level coding environments in a scalable way? Checkout RepoST: an automated framework to construct repo-level environments using Sandbox Testing (repost-code-gen.github.io) Models trained with RepoST data can generalize well to other datasets (e.g., RepoEval)
Yiqing Xie tweet media
English
3
23
88
10.6K
Yuning Mao 리트윗함
Xianjun Yang
Xianjun Yang@xianjun_agi·
📢My New Paper: Diversity-driven Data Selection for Language Model Tuning through Sparse Autoencoder TLDR: We proposed to use features from SAEs as a measure for data diversity&complexity and proved it's effectiveness on data selection for LLM tuning. arxiv.org/pdf/2502.14050
Xianjun Yang tweet media
English
7
37
180
18.6K
Yuning Mao 리트윗함
Thomas Wolf
Thomas Wolf@Thom_Wolf·
Among the most impressive aspect of the Llama 3.1 release is the accompanying research paper! Close to 100 pages of deep knowledge-sharing on LLMs like we havn't seen very often recently What a treat! It covers everything, pretrainining data, filtering, annealing, synthetic data, scaling laws, infrastructures, parallelism, training recipees, post-training adaptation, tool-use, benchmarking, inference strategies, quantization, vision, speech, videos... Mind-blown! Maybe the single paper you can read today to join the field of LLM from zero right to the frontier Read it here and feel the open-science ai.meta.com/research/publi…
Thomas Wolf tweet media
English
15
250
1.1K
76.1K
Yuning Mao 리트윗함
AI at Meta
AI at Meta@AIatMeta·
Introducing Meta Llama 3: the most capable openly available LLM to date. Today we’re releasing 8B & 70B models that deliver on new capabilities such as improved reasoning and set a new state-of-the-art for models of their sizes. Today's release includes the first two Llama 3 models — in the coming months we expect to introduce new capabilities, longer context windows, additional model sizes and enhanced performance + the Llama 3 research paper for the community to learn from our work. More details ➡️ go.fb.me/i2y41n Download Llama 3 ➡️ go.fb.me/ct2xko
English
340
1.4K
5.7K
1.1M
Yuning Mao 리트윗함
Mikayel Samvelyan
Mikayel Samvelyan@_samvelyan·
Introducing 🌈 Rainbow Teaming, a new method for generating diverse adversarial prompts for LLMs via LLMs It's a versatile tool 🛠️ for diagnosing model vulnerabilities across domains and creating data to enhance robustness & safety 🦺 Co-lead w/ @sharathraparthy & @_andreilupu
English
5
44
179
56.4K
Yuning Mao 리트윗함
AI at Meta
AI at Meta@AIatMeta·
Announcing Purple Llama — A new project to help level the playing field for building safe & responsible generative AI experiences. Purple Llama includes permissively licensed tools, evals & models to enable both research & commercial use. More details ➡️ bit.ly/3ReRNHI
AI at Meta tweet media
English
17
182
843
351K
Yuning Mao
Yuning Mao@yuning_pro·
Honored to be a core contributor of Llama 2 release. Give it a try and I hope you'll be pleasantly surprised. Feel free to DM me for any feedback (especially on safety).
AI at Meta@AIatMeta

We believe an open approach is the right one for the development of today's Al models. Today, we’re releasing Llama 2, the next generation of Meta’s open source Large Language Model, available for free for research & commercial use. Details ➡️ bit.ly/3Dh9hNp

English
1
0
36
5.3K
Yuning Mao
Yuning Mao@yuning_pro·
UniPELT often surpasses the upper bound when taking the best performance of all its submodules used individually on each task, indicating that a mixture of multiple PELT methods may be inherently more effective than single methods. 🧵[4/4]
Yuning Mao tweet media
English
0
0
3
0
Yuning Mao
Yuning Mao@yuning_pro·
On the GLUE benchmark, UniPELT consistently achieves 1~3pt gains compared to the best individual PELT method that it incorporates and even outperforms fine-tuning under different setups, exhibiting superior model effectiveness and robustness. 🧵[3/4]
Yuning Mao tweet media
English
1
0
3
0