Luke Rowe

105 posts

Luke Rowe banner
Luke Rowe

Luke Rowe

@Luke22R

PhD student at @Mila_Quebec, focusing on autonomous driving. Previously @Waymo and @torc_robotics.

Montréal, Québec เข้าร่วม Şubat 2022
175 กำลังติดตาม157 ผู้ติดตาม
ทวีตที่ปักหมุด
Luke Rowe
Luke Rowe@Luke22R·
🚀 Our method, Poutine, was the best-performing entry in the 2025 Waymo Vision-based End-to-End Driving Challenge at #CVPR2025! Our 3 B-parameter VLM Poutine scored 7.99 RFS on the official test set—comfortably ahead of every other entry (see figure).
Luke Rowe tweet media
English
4
11
22
3.2K
Luke Rowe รีทวีตแล้ว
Felix Heide
Felix Heide@_FelixHeide_·
Are we done with object detection? What about tiny objects beyond 200 meters? 🔎 Telescope 🔭 addresses long-range perception by explicitly tackling extreme scale imbalance ⚖️ in images. It hinges on a learnable hyperbolic foveation transform from a low-resolution image, magnifying distant regions 🔍 while compressing nearby ones - effectively normalizing object scales with minimal computational overhead. Objects are detected in the transformed (Riemannian) space using a novel bounding box parameterization and are then mapped back to the original image. Project: light.princeton.edu/telescope/
English
22
121
1.1K
187.9K
Luke Rowe รีทวีตแล้ว
Daphne Cornelisse
Daphne Cornelisse@daphne_cor·
New preprint led by @saeedrmd! “Artificial Intelligence for Modeling and Simulation of Mixed Automated and Human Traffic." Drawing on 200+ papers, it offers a snapshot of where the field stands and outlines promising directions ahead.
Daphne Cornelisse tweet media
English
1
2
16
1.4K
Luke Rowe รีทวีตแล้ว
Felix Heide
Felix Heide@_FelixHeide_·
ScenarioControl 🚗🛣️ - Scenario Generation from a single Dashcam Image 📸 or Text Prompt 💬!! Excited to introduce a new vision-language control mechanism for learned driving scenario generation. Given a single dashcam image or a scene prompt or an image, we generate a full scene layout 🧩, temporally consistent rollouts, including map 🗺️, agents 🚗, and ego video🛣️ ScenarioControl enables direct, fine-grained control over layout and traffic while preserving realism. It operates in a vectorized latent space with a new cross-global control mechanism to fuse vision-language inputs with scene structure while preserving realism. Interfaces seamlessly with generative video models! Project: light.princeton.edu/ScenarioControl Super fun project by Lili Gao, @Yanbo_Xu_ , William Koch, Samuele Ruffino, @Luke22R , Behdad Chalaki, Dmitriy Rivkin, Julian Ost, @rogg1111, Mario Bijelic.
English
0
11
39
22.1K
Luke Rowe รีทวีตแล้ว
Zhijian Liu
Zhijian Liu@zhijianliu_·
Reasoning VLAs can think. They just can't think fast. Until now. Introducing FlashDrive⚡ 🚀 716 ms → 159 ms on RTX PRO 6000 (up to 5.7×) ✅ Zero accuracy loss FlashDrive = streaming inference + DFlash speculative reasoning + ParoQuant W4A8 Real-time reasoning for autonomous driving is here! z-lab.ai/projects/flash…
English
32
162
1.3K
162.9K
Luke Rowe รีทวีตแล้ว
Felix Heide
Felix Heide@_FelixHeide_·
Chop the gradients ✂️! We found that truncating decoder gradients in latent video diffusion to a fixed window allows us to finetune on videos with pixel-wise perceptual losses without running out of memory. Pixel losses have been essential for image generation and reconstruction, but until now, they haven't scaled to long-duration, high-resolution video diffusion due to recursive activation accumulation in causal decoders, leading to OOM during training 💥📉. Project: light.princeton.edu/chopgrad/ Video diffusion models can do a lot more 🚀 when you can backprop the decoder! Post-process neural rendered scenes, super-resolve videos, harmonize lighting in controlled synthetic driving scenes, and inpaint videos — all in a single step ⚡ with a quick finetune from a standard diffusion model.
English
4
39
349
28.1K
Luke Rowe รีทวีตแล้ว
Felix Heide
Felix Heide@_FelixHeide_·
WorldFlow3D: Unbounded 3D World Generation 🌍 by Flow Through Hierarchical Distributions, without VAEs ! We reformulate 3D generation as flowing through sequentially finer 3D distributions, cutting training time by more than half ⏱️ compared to existing approaches! Vectorized map layouts provide full scene controllability 🗺️, and a novel flow-field alignment process enables causally coherent, spatially unbounded generation 🌍. This generative method generalizes across both real and synthetic data distributions! Project: light.princeton.edu/worldflow3d Project led by @amogh7joshi and Julian Ost — will be super fun to build on this! 🔥
English
2
27
239
19.4K
Luke Rowe รีทวีตแล้ว
Joseph Suarez 🐡
Joseph Suarez 🐡@jsuarez·
Releasing PufferLib 4.0: Train agents in seconds
English
39
94
1.1K
178.3K
Luke Rowe รีทวีตแล้ว
Patrick Yin
Patrick Yin@patrickhyin·
We’re releasing OmniReset, a framework for training robot policies using large-scale RL and diverse resets for contact-rich, dexterous manipulation. OmniReset pushes the frontier of robustness and dexterity, without any reward engineering or demonstrations. Try the policies yourself in our interactive simulator! weirdlabuw.github.io/omnireset/ (1/N 🧵)
English
21
95
468
107.6K
Luke Rowe รีทวีตแล้ว
Tenny Yin
Tenny Yin@tennyyin·
🎮 Can we learn interactive world models from letting robots “play”? ➡️ Introducing ✨PlayWorld: a framework for training high-fidelity video world models from large-scale autonomous play experience that enables: → Accurate dynamics prediction → Reliable policy evaluation → RL fine-tuning entirely inside the world model 🌐robot-playworld.github.io
English
2
31
188
34.3K
Luke Rowe รีทวีตแล้ว
Helen Tianyue Zhang
Helen Tianyue Zhang@HelennnnnnZhang·
🚨 New Paper 🚨 TL;DR we derive scaling laws of lr, momentum, and batch size for modern first-order optimizers through the lens of recent convergence bounds for LMO, a framework that includes normalized SGD, signSGD (approximating Adam), and Muon x.com/orvieto_antoni…
Antonio Orvieto@orvieto_antonio

Optimization theory for adaptive methods actually predicts most of what we know about hyperparameter scaling in LLM pretraining, and suggests new strategies as well. We did a deep dive here.

English
1
15
88
8.7K
Luke Rowe รีทวีตแล้ว
Lucas Maes
Lucas Maes@lucasmaes_·
JEPA are finally easy to train end-to-end without any tricks! Excited to introduce LeWorldModel: a stable, end-to-end JEPA that learns world models directly from pixels, no heuristics. 15M params, 1 GPU, and full planning <1 second. 📑: le-wm.github.io
English
107
558
3.9K
935.9K
Luke Rowe รีทวีตแล้ว
Joan Rodriguez
Joan Rodriguez@joanrod_ai·
Introducing @QuiverAI, a new AI lab and product company focused on frontier vector design. We’ve raised an $8.3M seed round led by @a16z, with support from amazing angels and investors. Our first model, Arrow-1.0, generates SVGs from images and text. It’s available now in public beta at app.quiver.ai
English
307
292
4.8K
1.3M
Luke Rowe รีทวีตแล้ว
Junyoung Seo
Junyoung Seo@jyseo_cv·
What if a world model could render not an imagined place, but the actual city? We introduce Seoul World Model, the first world simulation model grounded in a real-world metropolis. TL;DR: We made a world model RAG over millions of street-views. proj: seoul-world-model.github.io
English
42
207
1.5K
176.5K
Luke Rowe รีทวีตแล้ว
Andrew Davison
Andrew Davison@AjdDavison·
I'm not sure about the details but I'm convinced that how to publish and create impact is due to change very significantly in the near future. The value of writing and reading 8 page PDFs is rapidly dropping. What is the right way to publish the nugget of a research contribution?
Jon Barron@jon_barron

If I was a grad student today, I would: 1) Not write papers, 2) push my (agent-written) code to a public repo ~weekly, 3) maintain (via agents) a writeup.tex (manually verified) and a skill.md in the repo, and 4) work towards establishing skill usage as the new "citation" format.

English
12
15
123
21.6K
Luke Rowe รีทวีตแล้ว
Siva Reddy
Siva Reddy@sivareddyg·
Montreal deep tech scene is getting hot!! Many recent hires of Cohere, Mistral, Periodic Labs, Poolside are all based in Montreal. And now, AMI will have an office here 🔥 It's a no-brainer, though. @Mila_Quebec has the highest concentration of deep learning expertise with interdisciplinary connections. Thanks to recent US regulation changes on immigration, no more brain drain! Let's build more in Canada!
Yann LeCun@ylecun

Unveiling our new startup Advanced Machine Intelligence (AMI Labs). We just completed our seed round: $1.03B / 890M€, one the largest seeds ever, probably the largest for a European company. We're hiring! [the background image is the Veil Nebula - a picture I took from my backyard, most appropriate for an unveiling] More details here: techcrunch.com/2026/03/09/yan…

English
19
49
756
72.5K
Luke Rowe รีทวีตแล้ว
Benjamin Thérien @ ICLR 2026
Benjamin Thérien @ ICLR 2026@benjamintherien·
Are frontier LLMs trained across datacenters? One thing is certain: if the pre-training optimizer’s critical batch size is too small, they are NOT! Excited to announce MuLoCo, a pre-training optimizer that can efficiently pre-train across datacenters while having large enough batch sizes to warrant doing so. 🧵1/N
Benjamin Thérien @ ICLR 2026 tweet media
English
3
33
94
17.6K
Luke Rowe รีทวีตแล้ว
Daphne Cornelisse
Daphne Cornelisse@daphne_cor·
Benchmarks are a key driver of progress. But how should we evaluate human-like driving? Does the Waymo Open Sim Agent Challenge (WOSAC) really capture what matters? Looking forward to any feedback!
Daphne Cornelisse tweet media
English
3
5
41
6.9K
Luke Rowe รีทวีตแล้ว
Emiliano Penaloza
Emiliano Penaloza@emilianopp_·
Remember all the self-distillation papers that came out last week. Well, we also propose it 😅, but… But alongside something better 😎 π-Distill We show that with this method, you can distill closed-source frontier models even tho their traces are hidden 🔒. Both our methods can reach and even surpass the performance of the industry-standard SFT + RL with access to reasoning traces 🤯. 🔬And we spent ~100,000 hours GPU hours on a comprehensive analysis, not because the method is finicky, but because we wanted to understand why it works so well. 🧵 1/10
English
11
78
432
50.9K
Luke Rowe รีทวีตแล้ว
Yipeng Zhang
Yipeng Zhang@yipengzz·
How can we predict multiple plausible targets from a single context in joint-embedding self-supervised learning (SSL)? Check out our paper titled “Self-Supervised Learning from Structural Invariance” accepted at #ICLR2026! Previously Best Paper Award at @unireps 2025. arxiv.org/abs/2602.02381 We introduce AdaSSL, which models the target uncertainty and relaxes the standard assumption that the positive pair share the same semantic features. Derived from first principles, we realize @ylecun’s JEPA with a learned latent variable for jointly learning better representations and world models, extending SSL’s utility to a broader range of data types. 1/🧵
English
2
23
82
10.6K