Not enough data to create a plot.
Try a different view from the menu above.
Pfaff, Nicholas
Empirical Analysis of Sim-and-Real Cotraining Of Diffusion Policies For Planar Pushing from Pixels
Wei, Adam, Agarwal, Abhinav, Chen, Boyuan, Bosworth, Rohan, Pfaff, Nicholas, Tedrake, Russ
-- In imitation learning for robotics, cotraining with demonstration data generated both in simulation and on real hardware has emerged as a powerful recipe to overcome the "sim2real gap". This work seeks to elucidate basic principles of this sim-and-real cotraining to help inform simulation design, sim-and-real dataset creation, and policy training. Focusing narrowly on the canonical task of planar pushing from camera inputs enabled us to be thorough in our study. These experiments confirm that cotraining with simulated data can dramatically improve performance in real, especially when real data is limited. The results also suggest that reducing the domain gap in physics may be more important than visual fidelity for nonprehensile manipulation tasks. Perhaps surprisingly, having some visual domain gap actually helps the cotrained policy - binary probes reveal that high-performing policies learn to distinguish simulated domains from real. We conclude by investigating this nuance and mechanisms that facilitate positive transfer between sim-and-real. In total, our experiments span over 40 real-world policies (evaluated on 800+ trials) and 200 simulated policies (evaluated on 40,000+ trials). Foundation models trained on large datasets have transformed natural language processing [2]][3] and computer vision [4]. However, this data-driven recipe has been challenging to replicate in robotics since real-world data for imitation learning can be expensive and time-consuming to collect [5]. Fortunately, alternative data sources, such as simulation and video, contain useful information for robotics. In particular, simulation is promising since it can automate robot-specific data collection.
Scalable Real2Sim: Physics-Aware Asset Generation Via Robotic Pick-and-Place Setups
Pfaff, Nicholas, Fu, Evelyn, Binagia, Jeremy, Isola, Phillip, Tedrake, Russ
Simulating object dynamics from real-world perception shows great promise for digital twins and robotic manipulation but often demands labor-intensive measurements and expertise. We present a fully automated Real2Sim pipeline that generates simulation-ready assets for real-world objects through robotic interaction. Using only a robot's joint torque sensors and an external camera, the pipeline identifies visual geometry, collision geometry, and physical properties such as inertial parameters. Our approach introduces a general method for extracting high-quality, object-centric meshes from photometric reconstruction techniques (e.g., NeRF, Gaussian Splatting) by employing alpha-transparent training while explicitly distinguishing foreground occlusions from background subtraction. We validate the full pipeline through extensive experiments, demonstrating its effectiveness across diverse objects. By eliminating the need for manual intervention or environment modifications, our pipeline can be integrated directly into existing pick-and-place setups, enabling scalable and efficient dataset creation.