Guy Van den Broeck

51 posts

Guy Van den Broeck banner
Guy Van den Broeck

Guy Van den Broeck

@guyvdb

Professor of Computer Science at UCLA @UCLAComSci; working on Artificial Intelligence

Los Angeles, CA เข้าร่วม Nisan 2008
1.6K กำลังติดตาม3.9K ผู้ติดตาม
Guy Van den Broeck
Guy Van den Broeck@guyvdb·
@trunghlt @IanLi1118 new tokens depend on already unmasked tokens, but they do not depend on each other when unmasking multiple tokens in a single step
English
1
0
6
83
Trung H
Trung H@trunghlt·
@IanLi1118 Before yielding tokens, the last layer has access to previous hidden outputs. Why do we assume it generates tokens independently?
Vietnam 🇻🇳 English
1
0
1
552
Guy Van den Broeck รีทวีตแล้ว
Ian Li
Ian Li@IanLi1118·
One of the biggest promises of Diffusion LLMs is parallel generation: predicting multiple tokens at once to bypass the sequential bottleneck of autoregressive models. However, parallel generation comes with a price. For example: Should the sentence “He is from [MASK] [MASK]” be filled with [New] [York] or [San] [Diego]? If a diffusion model predicts both at the exact same time, it assumes independence and may produce... [San] [York]. 🤦‍♂️ We argue this arises from a structural misspecification: models are restricted to fully factorized outputs because parameterizing the full joint distribution would require a prohibitively massive output head. This is the Factorization Barrier crippling parallel generation. Here is how we broke it with CoDD.
Ian Li tweet media
English
8
31
309
21.9K
Guy Van den Broeck รีทวีตแล้ว
Guy Van den Broeck รีทวีตแล้ว
Guy Van den Broeck รีทวีตแล้ว
Guy Van den Broeck รีทวีตแล้ว
Discrete Diffusion Reading Group
Discrete Diffusion Reading Group@diffusion_llms·
📢Feb 2 (Mon): Planned Diffusion 🙅Diffusion language models are capable of parallelizing text generation but can struggle with coherence in low time-step regimes. 💡Planned Diffusion unlocks a new axis of parallelism: Token-level parallelism ➡️ semantic parallelism ✍️Planned diffusion first generates a structured plan, then diffuses semantically independent spans of text in parallel according to the plan. This Monday, Daniel Israel (UCLA) (@danielmisrael) and Tian Jin (MIT) (@jintian) will discuss their exciting Planned Diffusion paper as joint first authors. Collaborators: Ellie Cheng (elliecheng.com), Guy Van den Broeck (@guyvdb), Aditya Grover (@adityagrover_), Suvinay Subramanian (@suvinay), Michael Carbin (@mcarbin) Paper link: arxiv.org/abs/2510.18087
Discrete Diffusion Reading Group tweet media
English
1
9
48
10K
Guy Van den Broeck รีทวีตแล้ว
Luis Lamb
Luis Lamb@luislamb·
@RealAAAI⁩ AAAI2021Conference - Neuro-Symbolic AI Panel, during the COVID-19 crisis. With ⁦@kerstingAIML⁩ ⁦@guyvdb⁩ ⁦@mattbotvinick⁩ Marta Kwiatkowska, Leslie Pack Kaelbling. Just a picture 🙂
Luis Lamb tweet media
Română
0
3
8
567
Yuanqi Du
Yuanqi Du@YuanqiD·
Kudos to the SPIGM workshop best paper and best poster winners!
Yuanqi Du tweet mediaYuanqi Du tweet media
English
1
4
13
3.9K
Guy Van den Broeck รีทวีตแล้ว
NeSy 2026
NeSy 2026@nesyconf·
Recordings of the NeSy 2025 keynotes are now available! 🎥 Check out insightful talks from @guyvdb , @tkipf and @dlmcguinness on our new Youtube channel. Topics include using symbolic reasoning for LLM, and object-centric representations @NeSyconference" target="_blank" rel="nofollow noopener">youtube.com/@NeSyconference
English
0
6
14
2.9K
Guy Van den Broeck รีทวีตแล้ว
Alex Chen
Alex Chen@itisalex3·
What happens when we compress the KV cache of prompts with multiple instructions? 🤔 Existing compression methods can lead to some instructions being ignored. 🙀 We propose simple changes to KV cache eviction that fix this problem alongside other pitfalls to be aware of. 💯
Alex Chen tweet media
English
2
2
17
2.5K
Guy Van den Broeck รีทวีตแล้ว
Daniel Israel
Daniel Israel@danielmisrael·
"An hour of planning can save you 10 hours of doing." ✨📝 Planned Diffusion 📝 ✨ makes a plan before parallel dLLM generation. Planned Diffusion runs 1.2-1.8× faster than autoregressive and an order of magnitude faster than diffusion, while staying within 0.9–5% AR quality.
English
7
48
320
38.4K
Aayush Karan
Aayush Karan@aakaran31·
We found a new way to get language models to reason. 🤯 No RL, no training, no verifiers, no prompting. ❌ With better sampling, base models can achieve single-shot reasoning on par with (or better than!) GRPO while avoiding its characteristic loss in generation diversity.
English
73
249
1.7K
266.9K
Guy Van den Broeck รีทวีตแล้ว
Daniel Israel
Daniel Israel@danielmisrael·
🔦Adaptive Parallel Decoding (APD) has been accepted as a spotlight paper at @NeurIPSConf ! I thank my collaborators, reviewers, and program organizers for this honor. A thread for those interested 🧵 (1/n)
Daniel Israel tweet media
English
12
25
175
17K
Guy Van den Broeck รีทวีตแล้ว
NeSy 2026
NeSy 2026@nesyconf·
@e_giunchiglia @guyvdb How can reverend Bayes help us to incorporate constraints? With NeSy of course 👀 With applications in non-toxic LLM generation and safe AI driving! @guyvdb
NeSy 2026 tweet mediaNeSy 2026 tweet media
English
1
1
6
476
Guy Van den Broeck รีทวีตแล้ว
NeSy 2026
NeSy 2026@nesyconf·
@e_giunchiglia Now, @guyvdb is giving the opening keynote arguing why symbolic AI is still relevant in the age of LLMs... With the help of Shrek!
NeSy 2026 tweet media
English
1
5
12
674
Guy Van den Broeck รีทวีตแล้ว
NeSy 2026
NeSy 2026@nesyconf·
@e_giunchiglia @guyvdb Behind all of these very nice methods are one central trick... Circuits! ➕✖️ These are tractable generative neural networks 😍
NeSy 2026 tweet mediaNeSy 2026 tweet media
English
1
2
6
486