Abrar Anwar

239 posts

Abrar Anwar banner
Abrar Anwar

Abrar Anwar

@_abraranwar

CS PhD student at @USCViterbi + intern @nvidia | prev intern @Cornell @SandiaLabs | undergrad @UTCompSci

Austin, TX Katılım Ağustos 2013
750 Takip Edilen676 Takipçiler
Sabitlenmiş Tweet
Abrar Anwar
Abrar Anwar@_abraranwar·
Robots are deployed for long periods of time, but how can they answer questions and generate goals based on their long-horizon history? During my internship at #NVIDIA, we built ReMEmbR, a retrieval-augmented memory for embodied robots. 1/8 🧵 nvidia-ai-iot.github.io/remembr/
English
5
31
263
34.6K
Abrar Anwar retweetledi
Hung-Chieh Fang
Hung-Chieh Fang@hungchiehfang·
Manipulation demands dexterous in-hand tool use, rich contact handling, and long-horizon stability. Introducing DexDrummer, our sim2real framework that unifies these skills in a drumming testbed. w/ @amberxie_, @jenngrannen, Kenneth Llontop, @DorsaSadigh (Videos at 1x speed)
English
9
11
40
6.1K
Abrar Anwar retweetledi
Jesse Thomason
Jesse Thomason@_jessethomason_·
This fall semester, I will be moving my GLAMOR lab (glamor.rocks) to the School of Interactive Computing (IC) at the Georgia Institute of Technology!
English
11
10
145
17.7K
Abrar Anwar retweetledi
Jiaheng Hu
Jiaheng Hu@JiahengHu1·
VLA models are capable generalists. But can they continually self-improve? Such Continual Reinforcement Learning (CRL) problems are traditionally considered very challenging. Surprisingly, we found that with the right setup, the simplest CRL recipe can work really well! arxiv.org/abs/2603.11653
English
7
50
268
43.4K
Abrar Anwar retweetledi
Amber Xie
Amber Xie@amberxie_·
Introducing HandelBot 🎹🤖, a real-world piano playing robot! Piano is extremely hard (even for humans!). We take a small but exciting step to replicate this beautiful skill w HandelBot. Our insight is combining sim priors w real world refinement & RL. w/ @haozhiq @DorsaSadigh
English
11
31
166
36.8K
Abrar Anwar retweetledi
Anthony Liang
Anthony Liang@aliangdw·
Just released RBM-1M-OOD! We collected and annotated 1k+ robot trajectories of varying expertise across 4 universities for evaluating reward models. 📂Dataset: huggingface.co/datasets/robom… Also give Robometer a try on your own robot trajectories here: huggingface.co/spaces/robomet…
Anthony Liang@aliangdw

Super excited to share Robometer, a reward model that works zero-shot across robots, tasks, and scenes! Try fine-tuning Robometer on your own dataset! 🌐Project website: robometer.github.io 💻Code: github.com/robometer/robo…

English
0
14
48
5.8K
Abrar Anwar retweetledi
Ville🤖
Ville🤖@VilleKuosmanen·
Tested robometer on an oven temperature selection task (ground truth is 200). Results based on temperature referenced in prompt: - 160: highest when rotating dial - 180: highest at start - 200: highest at end really impressive instruction following!
Jesse Zhang@Jesse_Y_Zhang

A reward model that works, zero-shot, across robots, tasks, and scenes? Introducing Robometer: Scaling general-purpose robotic reward models with 1M+ trajectories. Enables zero-shot: online/offline/model-based RL, data retrieval + IL, automatic failure detection, and more! 🧵 (1/12)

English
4
8
32
3.8K
Abrar Anwar retweetledi
pfung
pfung@philfung·
added the excellent Robometer algorithm!! Easily compare your fav robot reward functions (Robometer, TOPReward, GVL, etc) in one website: philfung.github.io/rewardscope Thanks @aliangdw, @yigitkkorkmaz, @Jesse_Y_Zhang, @JiahuiZhang__32 for amazing Robometer paper!
pfung@philfung

Inspired by the TopReward paper, I made a lil web tool to test these robot manipulation rewards on your own videos. Try: philfung.github.io/rewardscope Record yourself folding a towel, upload it, and compare: 1. TopReward (this paper) 2. GVL (Deepmind) 3. Brute Force (i.e. at each frame, ask LLM to reply with a probability) TopReward (Qwen3VL-8B) holds its own surprisingly well against the others, even if those use ChatGPT! Great work @DJiafei, UW, AllenAI, thanks for pushing @VilleKuosmanen.

San Francisco, CA 🇺🇸 English
1
8
55
4.3K
Abrar Anwar retweetledi
Abrar Anwar retweetledi
Yiğit Korkmaz
Yiğit Korkmaz@yigitkkorkmaz·
Oh, btw, it also works with humanoids, notice how the progress plateaus when the robot gets slower. It understands the difference between "opening" and "closing" a dishwasher, too. @adcock_brett just letting you know 🙃
GIF
GIF
Jesse Zhang@Jesse_Y_Zhang

A reward model that works, zero-shot, across robots, tasks, and scenes? Introducing Robometer: Scaling general-purpose robotic reward models with 1M+ trajectories. Enables zero-shot: online/offline/model-based RL, data retrieval + IL, automatic failure detection, and more! 🧵 (1/12)

English
0
3
19
2.2K
Abrar Anwar
Abrar Anwar@_abraranwar·
Such a good analysis from @memmelma on Robometer and TOPReward This is hopefully the start of a new line of work on robot reward models, and comparisons like this help figure out what we need to do next!
Marius Memmel@memmelma

There’s a discussion going on rn about two recent robotic reward models: TOPReward⛰️ and Robometer🌡️ Which one is better? It depends entirely on your objective! Here is a deep dive into the conceptual differences, strengths, and weaknesses of both. 🧵👇

English
0
0
11
610
Abrar Anwar retweetledi
Marius Memmel
Marius Memmel@memmelma·
There’s a discussion going on rn about two recent robotic reward models: TOPReward⛰️ and Robometer🌡️ Which one is better? It depends entirely on your objective! Here is a deep dive into the conceptual differences, strengths, and weaknesses of both. 🧵👇
Marius Memmel tweet media
English
3
18
55
14.1K
Abrar Anwar retweetledi
Jiafei Duan
Jiafei Duan@DJiafei·
With both TOPReward and RoboMeter now released, I’m sure people will compare them , and honestly, they’re both strong, just optimized for different goals. For TOPReward, we’re pitching it as a general-purpose reward model that works across robotics and non-robotics. We go from an off-the-shelf VLM like Qwen-VL 3 (which, notably, wasn’t trained on robotics data) to TOPReward with no reward training, no fine-tuning, and no in-context prompting. And have fun building excited to see what people make! But if you’re curious about a simple side-by-side between RoboMeter and TOPReward (implemented by @VilleKuosmanen), check it out 👇
Marius Memmel@memmelma

There’s a discussion going on rn about two recent robotic reward models: TOPReward⛰️ and Robometer🌡️ Which one is better? It depends entirely on your objective! Here is a deep dive into the conceptual differences, strengths, and weaknesses of both. 🧵👇

English
5
6
79
8.6K
Abrar Anwar retweetledi
Jesse Zhang
Jesse Zhang@Jesse_Y_Zhang·
This is some great analysis by @memmelma on comparing TOPReward and Robometer. Another point to stress: while Robometer is the best *robotics* reward model right now, TOPReward still performs well and likely performs better on non-robotics tasks as it doesnt finetune anything. Overall, TOPReward performs well, outperforms GVL significantly on our benchmark, and is even better than some finetuned models (VLAC, RoboDopamine). Nice work on TOPReward, @DJiafei! We have now added TOPReward results to our website table at robometer.github.io Kind of crazy to see how many reward models came out in the last 2 months, really made us work hard on our baselines table.
Marius Memmel@memmelma

There’s a discussion going on rn about two recent robotic reward models: TOPReward⛰️ and Robometer🌡️ Which one is better? It depends entirely on your objective! Here is a deep dive into the conceptual differences, strengths, and weaknesses of both. 🧵👇

English
0
3
33
2.7K