Ben Eisner

282 posts

Ben Eisner banner
Ben Eisner

Ben Eisner

@BenAEisner

ML/Robotics Researcher. Ph.D. Student in Robotics at CMU. Learning for Manipulation.

Pittsburgh, PA Katılım Aralık 2011
190 Takip Edilen395 Takipçiler
Ben Eisner retweetledi
Kallol Saha
Kallol Saha@_ksaha·
We will be presenting at CoRL 2025; come see our Oral Session talk (2-3PM) & Poster (4:30-6PM) on 29th September! Huge thanks to my co-authors @ambermli, Angela, @yu_lifan5260, @BenAEisner, and our amazing advisors Prof. @davheld & Prof. Maxim Likhachev. See you in Seoul! 🇰🇷
English
0
1
3
303
Ben Eisner retweetledi
Yufei Wang
Yufei Wang@YufeiWang25·
Introducing ArticuBot🤖at #RSS2025, in which we learn a single policy for manipulating diverse articulated objects across 3 robot embodiments in different labs, kitchens & lounges, achieved via large-scale simulation and hierarchical imitation learning. articubot.github.io 🧵
English
3
31
89
6.9K
Ben Eisner retweetledi
Vlad Feinberg
Vlad Feinberg@FeinbergVlad·
Definitely lots of blood, sweat, and tears went into this one! Congrats to the whole team on getting 2.0 Flash shipped, I'm very grateful to have had the chance to work with everyone on this!
Sundar Pichai@sundarpichai

We’re kicking off the start of our Gemini 2.0 era with Gemini 2.0 Flash, which outperforms 1.5 Pro on key benchmarks at 2X speed (see chart below). I’m especially excited to see the fast progress on coding, with more to come.  Developers can try an experimental version in AI Studio and Vertex AI today. It is also available to try in @GeminiApp on the web today, mobile coming soon.

English
3
14
214
40.9K
Ben Eisner retweetledi
Eric Cai
Eric Cai@eywcai·
Introducing TAX3D, in which we extend relative-placement methods to generalizable deformable manipulation! #CoRL2024 (1/🧵) Our approach generalizes to: - Diverse unseen objects - Diverse unseen configurations - Multimodal placements
GIF
English
2
15
49
7.4K
Ben Eisner retweetledi
Tesla Optimus
Tesla Optimus@Tesla_Optimus·
Navigating by myself
English
2.7K
6.8K
49.4K
7.2M
Ben Eisner retweetledi
Mohammad Nomaan Qureshi
Mohammad Nomaan Qureshi@qunomaan·
1/ 🎉 How can we develop methods to generate synthetic, photorealistic data for training #AI models in robotics? We present SplatSim, a step in this direction. SplatSim is a scalable framework that generates photorealistic data for manipulation tasks using existing simulators as a physics backbone — enabling zero-shot Sim2Real policy transfer for RGB policies! 🌍 Check out our project page: splatsim.github.io and paper on arXiv arxiv.org/abs/2409.10161 With @_sparshgarg_ Francisco Yandun, @davheld, George Kantor, @abhi_silwal @CMU_Robotics
English
6
48
230
33.4K
Ben Eisner
Ben Eisner@BenAEisner·
@__tinygrad__ @realGeorgeHotz Wow, that's wild. The FP16 + FP32 accumulate details are publicly posted for the 4090 (ada whitepaper), but the L40S documentation doesn't break it down into accumulation type. Is that published? Or did you just have to benchmark to figure out empirically what was going on?
English
0
0
1
82
the tiny corp
the tiny corp@__tinygrad__·
@BenAEisner @realGeorgeHotz Because NVIDIA is scamming you. FP16 with FP32 accumulate is nerfed to half speed on 4090. FP16 with FP16 accumulate is very hard to use. Still way cheaper per FLOP even with nerf.
English
1
0
7
314
Ben Eisner
Ben Eisner@BenAEisner·
@realGeorgeHotz 2. Any insight into why the 8x L40s machines in this benchmark tinybox posted (public.tableau.com/app/profile/da…) train resnet 2.5x-ish faster than tinybox, even though the total fp16 flops on an L40s node should be like only 50% more? Is it just GPU memory + cpu configuration?
English
0
0
1
100
Ben Eisner
Ben Eisner@BenAEisner·
@realGeorgeHotz Super excited about this - considering buying one for our lab's AI+robotics research. Having trouble doing flops/$ calculations, though. 2 (maybe naive?) Qs: 1. Why is advertised NVIDIA tinybox flops 991 tflops, when each 4090 theoretically has 330 tflops of pure fp16 ops?
English
2
0
2
373
Ben Eisner
Ben Eisner@BenAEisner·
@sweetgreen seems that there’s something buggy w/ the sweetgreen app this week. 2x I’ve tried to place outpost orders this week, which charged my card via Apple Pay but then immediately refunded saying something went wrong. But my daily Sweetpass credit was still used up. 😢
English
1
0
0
89
John Hewitt
John Hewitt@johnhewtt·
I’m joining the Columbia Computer Science faculty as an assistant professor in fall 2025, and hiring my first students this upcoming cycle!! There’s so much to understand and improve in neural systems that learn from language — come tackle this with me!
John Hewitt tweet media
English
122
52
887
100.4K
Ben Eisner
Ben Eisner@BenAEisner·
My friends made a really, really awesome photo-to-Etch-a-Sketch robot. youtube.com/watch?v=iQhhut…. Draws human sketches at lightning speed - rarely see such high-quality 1x-speed robots. Amazing work!
YouTube video
YouTube
English
0
3
20
5.1K
Ben Eisner
Ben Eisner@BenAEisner·
@asoare159 Yeah I've thought about using diffusion/generative modeling for either the distance directly or diffusing some part of the latent representation, but still need to try it out. Since the points themselves are highly highly correlated, I'd bet that latents would work better.
English
0
0
1
28
Alexander Soare
Alexander Soare@asoare159·
@BenAEisner Luckily since ur only predicting distance, u don't need to deal with distributions in non-linear spaces. Although I imagine the distribution is quite funky for many pt pairs. I wonder if u could use generative modelling as is the trend in e2e robot learning (ACT, diffusion policy
English
1
0
0
46
Ben Eisner
Ben Eisner@BenAEisner·
(1/N) How can we get robots to make precise placement predictions when solving rearrangement tasks, just by watching demonstrations? In our ICLR 2024 paper, “Deep SE(3)-Equivariant Geometric Reasoning for Precise Placement Tasks”, we do just that! Paper: arxiv.org/abs/2404.13478
GIF
English
3
15
70
7K
Ben Eisner retweetledi
Murtaza Dalal
Murtaza Dalal@mihdalal·
LLMs are capable of high-level planning, but they require pre-trained skills! Our #ICLR2024 paper instead uses LLM guidance to train RL agents from scratch to solve 25+ long-horizon robotics tasks across four benchmarks w/ >85% success rates Paper & code: mihdalal.github.io/planseqlearn
English
5
35
129
36.3K
Ben Eisner
Ben Eisner@BenAEisner·
@asoare159 Yes it could be used to predict intermediate poses! We've used this to do say, pre-grasp - grasp sequences, similar to your suggestion. The only tricky bit (which we're working on) is handling multimodality natural in most free-space motion seen in demos.
English
1
0
1
46
Alexander Soare
Alexander Soare@asoare159·
@BenAEisner Awesome! Skimming the paper I got the impression that you predict just grasp and target pose. I suppose that the method could also be used to predict intermediate poses? Say you have to slot an object into a receptacle and your approach trajectory matters.
English
1
0
0
92