Lucas

231 posts

Lucas banner
Lucas

Lucas

@quantbagel

ml research, ex-quant 2x

SF | Montreal Katılım Ağustos 2025
953 Takip Edilen560 Takipçiler
Sabitlenmiş Tweet
Lucas
Lucas@quantbagel·
Robot action models shouldn't need 256 vision tokens per frame. Pi0.5 spends 400M parameters on SigLIP just to see. We replaced it with a 4.4M encoder that outputs 5 tokens — and action quality barely changes. 91x smaller. 51x fewer tokens. 7.3x faster inference.
Lucas tweet media
English
23
31
353
18.4K
Lucas retweetledi
Keller Cliffton
Keller Cliffton@Keller·
The Bitter Lesson of Robotics: It's extremely easy to make a video of a robot doing something once under perfect conditions then post it to X. But it often takes a decade to harden systems and design for all the insane edge cases of the real world. Many companies raising $$$$ on cool demos, but all the hard work comes after
English
74
166
1.6K
144.4K
Lucas
Lucas@quantbagel·
And so it begins
Lucas tweet media
English
1
0
8
438
Lucas
Lucas@quantbagel·
@jparkjmc Please contribute to my research too🙏
English
0
0
1
338
Lucas
Lucas@quantbagel·
Robot action models shouldn't need 256 vision tokens per frame. Pi0.5 spends 400M parameters on SigLIP just to see. We replaced it with a 4.4M encoder that outputs 5 tokens — and action quality barely changes. 91x smaller. 51x fewer tokens. 7.3x faster inference.
Lucas tweet media
English
23
31
353
18.4K
Lucas retweetledi
TBPN
TBPN@tbpn·
"Go all the way until it hurts. If you're doing something and it's easy, it's not valuable." - @travisk "If anyone says a strategic thing was easy, I'm like, 'You messed up. You could have gone way further. More competitive advantage. More differentiation. Get it together.'"
English
50
453
3K
936K
Lucas retweetledi
Christine Yip
Christine Yip@christinetyip·
We were inspired by @karpathy 's autoresearch and built: autoresearch@home Any agent on the internet can join and collaborate on AI/ML research. What one agent can do alone is impressive. Now hundreds, or thousands, can explore the search space together. Through a shared memory layer, agents can: - read and learn from prior experiments - avoid duplicate work - build on each other's results in real time
Christine Yip tweet mediaChristine Yip tweet media
English
122
263
2.4K
261.7K
stochasm
stochasm@stochasticchasm·
what's the most minimal and hackable agent out there? some type of thing with bash,read,write and not much more
English
20
1
56
6.8K
Isaac Sin
Isaac Sin@IsaacSin12·
We automated @LeRobotHF calibration for so101. Tired of watching people burn hours on setup before they even touch teleportation, so we built a script that does it for you. Part of our bigger mission is to make Physical AI as easy to get started with as possible. @Ryan_Resolution and I would be open sourcing our platform soon. Stay tuned.
English
30
39
389
22K
Lucas retweetledi
alex zhang
alex zhang@a1zhang·
$1,100,000 in cash to write GPU kernels for the latest @GPU_MODE competition holy fuck bro
alex zhang tweet media
English
7
24
406
28.4K
zach
zach@zachglabman·
I want to sponsor a Physical AI / robotics hackathon in San Francisco end of March or early April Focusing on sensing, vision and autonomy Dm me if you’re down or want to host
English
64
12
294
16.5K
Lucas
Lucas@quantbagel·
@someraregem you cant do flowcharting on pi-style models because each verification is roughly equivalent to a Euler step so theres no benefit
English
0
0
1
31
Archer Lin
Archer Lin@archerlin__·
@quantbagel we’re building an affordable humanoid for researcher, let me know if you want to deploy the model on this platform.
English
1
0
0
25
Lucas
Lucas@quantbagel·
Vllm for robotics
Deutsch
1
0
14
902
Lucas
Lucas@quantbagel·
Working on LIBERO for the decoder still, but if you want to contribute just message me! (This was pi0.5 btw)
English
0
0
1
179
Lucas
Lucas@quantbagel·
Im doing it a bit better now it just needed more data, lower mse than originally
Kyle Vedder@KyleVedder

@quantbagel we modified the visual encoder somewhat similarly to add visual memory the clever bit in our approach is it takes variable frame counts but fixed output token count, and we preserve identical outputs to the siglip when there’s only one frame

English
0
0
3
354
Lucas retweetledi
atharva ☆
atharva ☆@k7agar·
the robotic problem rn is how do you keep your huge slow bulky neural network inference sub ~300ms for good real time control. good mix of both engineering and research problems to solve for robot inference to be as smooth as possible slow is bad, bad is not acceptable
English
16
4
149
10K
Lucas
Lucas@quantbagel·
Does anyone have compute credits I could have to scale my encoder to speed up pi0 ( I want it to be somewhat general so I can release it on hugging face) also want to implement saguaro (speculative speculative decoding for VLAs)
English
4
2
14
1.3K