Lars Ankile

149 posts

Lars Ankile

Lars Ankile

@larsankile

ML for robotics.

Palo Alto, CA Katılım Aralık 2012
593 Takip Edilen587 Takipçiler
Sabitlenmiş Tweet
Lars Ankile
Lars Ankile@larsankile·
How can we enable finetuning of humanoid manipulation policies, directly in the real world? In our new paper, Residual Off-Policy RL for Finetuning BC Policies, we demonstrate real-world RL on a bimanual humanoid with 5-fingered hands (29 DoF) and improve pre-trained policies with ~15-75 minutes of robot interaction. By learning residual corrections on frozen BC policies using sample-efficient off-policy RL, we achieve significant improvements in sample efficiency, enabling policy finetuning directly on the hardware — to our knowledge, one of the first examples of this on a humanoid with bimanual dexterous hands. (If you know of other examples, let me know!)
English
11
57
260
51.8K
Lars Ankile retweetledi
Marcel Torné
Marcel Torné@marceltornev·
Just told my parents I’m hosting a “AI coding tool workshop” for them later this week. They’re so excited 😄 Has anyone here tried teaching their parents an AI coding tool before? How did it go?!
English
2
2
12
467
Lars Ankile
Lars Ankile@larsankile·
@Garmin @GarminFitness why are you so opposed to me getting access to _my own data_ from my Garmin watch that _I bought_? It's maybe a lot to ask, but maybe you can try to not throttle/block my one API request I do per day to look at my data?
English
1
0
5
252
Lars Ankile
Lars Ankile@larsankile·
@Garmin, why do you not allow personal use of the Connect API?
English
0
1
4
223
Lars Ankile retweetledi
Younghyo Park
Younghyo Park@younghyo_park·
What's different between these two BC policies? It's the same architecture, training budget, and data collection setup — the only difference is the controller gains! Controller gains are an understudied design parameter in robot learning. In our new work (w/ @BronarsToni*, @pulkitology), we show how they act as an inductive bias across BC, RL, and Sim2Real transfer, with real consequences on performance. Here's what we found 🧵 * Equal Contribution 📄arxiv: arxiv.org/abs/2604.02523 🔗website: younghyopark.me/tune-to-learn/
English
7
75
465
141.1K
Lars Ankile
Lars Ankile@larsankile·
@1x_tech any updates on when my Neo will ship? 🙏
English
0
0
7
228
Lars Ankile
Lars Ankile@larsankile·
@seungwookh This is really cool! I gotta admit, I had my doubts that this could even work when you first told me about it so big kudos for taking it all the way 🙏
English
1
0
2
493
Lars Ankile retweetledi
Seungwook Han
Seungwook Han@seungwookh·
Can language models learn useful priors without ever seeing language? We pre-pre-train transformers on neural cellular automata — fully synthetic, zero language. This improves language modeling by up to 6%, speeds up convergence by 40%, and strengthens downstream reasoning. Surprisingly, it even beats pre-pre-training on natural text! Blog: hanseungwook.github.io/blog/nca-pre-p… (1/n)
Seungwook Han tweet media
English
47
262
1.7K
247.8K
Lars Ankile retweetledi
Tanishq Kumar
Tanishq Kumar@tanishqkumar07·
I've been working on a new LLM inference algorithm. It's called Speculative Speculative Decoding (SSD) and it's up to 2x faster than the strongest inference engines in the world. Collab w/ @tri_dao @avnermay. Details in thread.
English
134
455
4.1K
606.6K
Lars Ankile retweetledi
Kushal
Kushal@kushalk_·
🤖 Can a single robot policy manipulate diverse tools without ever seeing them before? Introducing SimToolReal 🔨 : a generalist dexterous manipulation policy that transfers zero-shot sim→real to unseen tools + unseen tasks All videos are 1x speed (60 Hz control) 🧵👇
English
21
77
334
60.5K
Lars Ankile retweetledi
Zhanpeng He
Zhanpeng He@zhanpeng_he·
Excited to organize the ICRA 2026 Workshop on Reinforcement Learning in the Era of Imitation Learning (June 1 · Vienna 🇦🇹)! Imitation learning scales robot policies — but robustness & real-world adaptation remain open challenges. How can RL improve real-world robot performance? Speakers: @svlevine, @chelseabfinn, @pulkitology, @RobotPlatt, @davheld, @JasonMa2020, @GeorgiaChal More information: rl4il-icra.github.io Organizing w/ @stephentian_, @XiaomengXu11, @ric_and_robots, @albertyu101 #ICRA2026 #Robotics #ReinforcementLearning #ImitationLearning
Zhanpeng He tweet media
English
0
12
82
5.6K
Lars Ankile retweetledi
Kevin Zakka
Kevin Zakka@kevin_zakka·
Some exciting Friday news 🙂 We just open-sourced our system identification toolbox in MuJoCo 3.5. Get started today: "pip install mujoco[sysid]" mjlab v1.1 is also out featuring a brand new RGB-D renderer and now fully available on PyPI. Install with: "pip install mjlab"
English
12
15
152
9.2K
Lars Ankile
Lars Ankile@larsankile·
@ZechenZhang5 How do you typically store and structure results in your repo?
English
0
0
0
11
Zechen Zhang
Zechen Zhang@ZechenZhang5·
How I use it: 1. Point Claude at my research repo 2. It explores code, results, docs 3. Drafts the paper using all these principles 4. Fetches BibTeX via APIs (never from memory) 5. I iterate with it in an IDE with pdf compiled every often like LaTeX workshop extension
English
2
0
14
3.6K
Lars Ankile retweetledi
Qiyang (Colin) Li
Qiyang (Colin) Li@qiyang_li·
Flow policies are getting popular in robotics as they capture multimodal prior data well, and synergize nicely with action chunking. But it is unclear how to best train them with RL effectively. We found something that works pretty well! (spoiler: use Adjoint Matching) 🧵1/N
Qiyang (Colin) Li tweet media
English
8
58
452
86K