
Bones Studio
23 posts

Bones Studio
@TheBonesStudio
The Source of ground-truth multimodal human motion & behavior data for Physical AI — from world models to humanoid control | 5+ years, trusted by leading labs


Need high-quality motion for humanoid robots or digital humans? Meet Kimodo: our new diffusion model trained on 700 hours of optical mocap data for easy, controllable, and high-fidelity motion generation. @NVIDIAAI research.nvidia.com/labs/sil/proje…

288 hours of high-quality, text-annotated human motion data are now available! 140k motion sequences! Do you know that a large part of SONIC's training data is now open-sourced? Check out the dataset here 👇🏻 from our friends at Bones Studio! Full human + G1 retargeted motion! Stie🌐:bones.studio/datasets/seed Data💿:huggingface.co/datasets/bones… SONIC training code coming VERY VERY soon!




BONES-SEED is more than just a larger dataset. This is professional studio mocap. Way higher quality than what was previously available out there. Less artifacts -> transfers directly to higher quality motion, both for animation and robotics! 🤖








With SONIC, we CAN do many things! Now you can too.

We trained a humanoid with 22-DoF dexterous hands to assemble model cars, operate syringes, sort poker cards, fold/roll shirts, all learned primarily from 20,000+ hours of egocentric human video with no robot in the loop. Humans are the most scalable embodiment on the planet. We discovered a near-perfect log-linear scaling law (R² = 0.998) between human video volume and action prediction loss, and this loss directly predicts real-robot success rate. Humanoid robots will be the end game, because they are the practical form factor with minimal embodiment gap from humans. Call it the Bitter Lesson of robot hardware: the kinematic similarity lets us simply retarget human finger motion onto dexterous robot hand joints. No learned embeddings, no fancy transfer algorithms needed. Relative wrist motion + retargeted 22-DoF finger actions serve as a unified action space that carries through from pre-training to robot execution. Our recipe is called "EgoScale": - Pre-train GR00T N1.5 on 20K hours of human video, mid-train with only 4 hours (!) of robot play data with Sharpa hands. 54% gains over training from scratch across 5 highly dexterous tasks. - Most surprising result: a *single* teleop demo is sufficient to learn a never-before-seen task. Our recipe enables extreme data efficiency. - Although we pre-train in 22-DoF hand joint space, the policy transfers to a Unitree G1 with 7-DoF tri-finger hands. 30%+ gains over training on G1 data alone. The scalable path to robot dexterity was never more robots. It was always us. Deep dives in thread:

