Ke Wang

47 posts

Ke Wang banner
Ke Wang

Ke Wang

@wangkeml

PhD student in Machine Learning @ EPFL

Lausanne, Switzerland Присоединился Eylül 2022
122 Подписки86 Подписчики
Ke Wang ретвитнул
Maksym Andriushchenko
Maksym Andriushchenko@maksym_andr·
🚨 Incredibly excited to share that I'm starting my research group focusing on AI safety and alignment at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems in September 2025! 🚨 Hiring. I'm looking for multiple PhD students: both those able to start in Fall 2025 (i.e., as soon as possible) and through centralized programs like CLS, IMPRS, and ELLIS (the deadlines are in November) to start in Spring–Fall 2026. I'm also searching for postdocs, master's thesis students, and research interns. Fill the Google form below if you're interested! Research group. We will focus on developing algorithmic solutions to reduce harms from advanced general-purpose AI models. We're particularly interested in alignment of autonomous LLM agents, which are becoming increasingly capable and pose a variety of emerging risks. We're also interested in rigorous AI evaluations and informing the public about the risks and capabilities of frontier AI models. Additionally, we aim to advance our understanding of how AI models generalize, which is crucial for ensuring their steerability and reducing associated risks. For more information about research topics relevant to our group, please check the following documents: - International AI Safety Report, - An Approach to Technical AGI Safety and Security by DeepMind, - Open Philanthropy’s 2025 RFP for Technical AI Safety Research. Research style. We are not necessarily interested in getting X papers accepted at NeurIPS/ICML/ICLR. We are interested in making an impact: this can be papers (and NeurIPS/ICML/ICLR are great venues), but also open-source repositories, benchmarks, blog posts, even social media posts—literally anything that can be genuinely useful for other researchers and the general public. Broader vision. Current machine learning methods are fundamentally different from what they used to be pre-2022. The Bitter Lesson summarized and predicted this shift very well back in 2019: "general methods that leverage computation are ultimately the most effective". Taking this into account, we are only interested in studying methods that are general and scale with intelligence and compute. Everything that helps to advance their safety and alignment with societal values is relevant to us. We believe getting this—some may call it "AGI"—right is one of the most important challenges of our time. Join us on this journey!
Maksym Andriushchenko tweet media
English
76
90
841
105.8K
Ke Wang ретвитнул
Skander Moalla
Skander Moalla@SkanderMoalla·
🚀 Big time! We can finally do LLM RL fine-tuning with rewards and leverage offline/off-policy data! ❌ You want rewards, but GRPO only works online? ❌ You want offline, but DPO is limited to preferences? ✅ QRPO can do both! 🧵Here's how we do it:
Skander Moalla tweet media
English
3
37
144
25.2K
Ke Wang ретвитнул
Manuel Madeira
Manuel Madeira@manuelmlmadeira·
Excited to present #DeFoG with @qinym710 at #ICML2025 ! Catch our oral today at 3:30 PM (West Exhibition Hall C) and join us at the poster after (4:30–7:00 PM, East Exhibition Hall A-B #E-3004). Come chat graphs & generative models!
GIF
Yiming Qin@qinym710

🚀 Presenting #DeFoG: our discrete flow‑matching framework for graph generation! Catch our #ICML2025 oral presentation today (3:30 – 3:45 PM, in West Exhibition Hall C) and drop by the poster right after (4:30 –7:00). Come chat graphs & generative models! @manuelmlmadeira

English
1
4
10
677
Ke Wang ретвитнул
Yiming Qin
Yiming Qin@qinym710·
🚀 Presenting #DeFoG: our discrete flow‑matching framework for graph generation! Catch our #ICML2025 oral presentation today (3:30 – 3:45 PM, in West Exhibition Hall C) and drop by the poster right after (4:30 –7:00). Come chat graphs & generative models! @manuelmlmadeira
English
1
3
11
1.3K
Ke Wang ретвитнул
Vincent Jung
Vincent Jung@jungvinc·
🧬 New roadmap out in Nature Reviews Molecular Cell Biology! 🤖 We show how RNA-LMs + GNNs can come together to model the RNA interactome & uncover new roles for non-coding RNA. 💊 Clinical links to RNA therapies for cancer & neuro diseases. 📄 Read it: bit.ly/4kNblk6
Nature Reviews Molecular Cell Biology@NatRevMCB

New Online! Decoding the interactions and functions of non-coding RNA with artificial intelligence bit.ly/4kNblk6

English
0
5
12
1.1K
Ke Wang
Ke Wang@wangkeml·
@_simonsmith In contrast, our method updates the model weights directly to inject knowledge, enabling it to generalize across semantically related prompts, as demonstrated by our SOTA performance on the generalization metric.
English
0
0
0
41
Simon Smith
Simon Smith@_simonsmith·
This looks interesting, but how does it compare to simply having facts stored in a database for retrieval? Like, why store updated facts in the model at all?
Yiming Qin@qinym710

How can we inject new knowledge into LLMs without full retraining, forgetting, or breaking past edits? We introduce MEMOIR 📖— a scalable framework for lifelong model editing that reliably rewrites thousands of facts sequentially using a residual memory module. 🔥 🧵1/7

English
5
0
0
235
Ke Wang
Ke Wang@wangkeml·
@_simonsmith For instance, it can fail to recognize rephrased edits, such as from "Where is the Eiffel Tower?" to "Which city is the Eiffel Tower located in?"
English
0
0
0
36
Ke Wang
Ke Wang@wangkeml·
@_simonsmith While GRACE achieves near-perfect accuracy, its reliance on explicit storage limits its ability to generalize to rephrased prompts regarding the same edited knowledge.
English
0
0
1
42
Ke Wang
Ke Wang@wangkeml·
@_simonsmith Thanks for liking and sharing our work! One of the baselines in our comparisons, GRACE, closely resembles what you described. It stores edited facts by directly saving modified output patterns in an external dictionary.
English
0
0
0
37
Ke Wang
Ke Wang@wangkeml·
@kaelrickwyn Thank you for liking and sharing our work!
English
0
0
0
575
Yiming Qin
Yiming Qin@qinym710·
This lets MEMOIR generalize to rephrased prompts. It rewrites facts like "The Eiffel Tower is in Berlin" and still answers correctly when asked "What is the location of the Eiffel Tower?" 🗼✅ 🧵 4/7
English
1
1
34
3.3K
Ke Wang
Ke Wang@wangkeml·
Very happy to share our new work: MEMOIR 📖! MEMOIR is a lifelong model editing method to edit factual knowledge in LLMs with minimal forgetting even with thousands of edits!
Yiming Qin@qinym710

How can we inject new knowledge into LLMs without full retraining, forgetting, or breaking past edits? We introduce MEMOIR 📖— a scalable framework for lifelong model editing that reliably rewrites thousands of facts sequentially using a residual memory module. 🔥 🧵1/7

English
0
4
17
1K
Ke Wang ретвитнул
Yiming Qin
Yiming Qin@qinym710·
How can we inject new knowledge into LLMs without full retraining, forgetting, or breaking past edits? We introduce MEMOIR 📖— a scalable framework for lifelong model editing that reliably rewrites thousands of facts sequentially using a residual memory module. 🔥 🧵1/7
Yiming Qin tweet media
English
26
169
1.2K
110.1K
Ke Wang ретвитнул
Yiming Qin
Yiming Qin@qinym710·
Happy to share that DeFoG: Discrete Flow Matching for Graph Generation will be showcased as a Spotlight Poster at #ICML2025 ! -> Explore the paper: arxiv.org/abs/2410.04263 -> Open-source code: github.com/manuelmlmadeir… Looking forward to your feedback on our repository!
Manuel Madeira@manuelmlmadeira

Are you interested in graph generation, from molecular discovery 🧪 to social networks 🌐? You’ll love DeFoG 🌬️😶‍🌫️, our new framework that delivers state-of-the-art performance in diverse graph generation tasks with unmatched efficiency! 🤩 📄: arxiv.org/abs/2410.04263 🧵1/9

English
0
8
34
6.8K
Ke Wang ретвитнул
Yiming Qin
Yiming Qin@qinym710·
Happy to release the code for DeFoG (github.com/manuelmlmadeir…), accepted as an ICML 2025 Spotlight Poster, with @manuelmlmadeira! → Efficient training & sampling → Wide dataset support & extensible dataloader → Hydra-powered CLI & WandB → Docker/Conda setup
English
0
1
10
614