Decoding Animal Behavior to Train Robots with EgoPet with Amir Bar
EPISODE 692
|
JULY
9,
2024
Watch
Follow
Share
About this Episode
Today, we're joined by Amir Bar, a PhD candidate at Tel Aviv University and UC Berkeley to discuss his research on visual-based learning, including his recent paper, “EgoPet: Egomotion and Interaction Data from an Animal’s Perspective.” Amir shares his research projects focused on self-supervised object detection and analogy reasoning for general computer vision tasks. We also discuss the current limitations of caption-based datasets in model training, the ‘learning problem’ in robotics, and the gap between the capabilities of animals and AI systems. Amir introduces ‘EgoPet,’ a dataset and benchmark tasks which allow motion and interaction data from an animal's perspective to be incorporated into machine learning models for robotic planning and proprioception. We explore the dataset collection process, comparisons with existing datasets and benchmark tasks, the findings on the model performance trained on EgoPet, and the potential of directly training robot policies that mimic animal behavior.
About the Guest
Amir Bar
Tel Aviv University and UC Berkeley
Resources
- EgoPet: Egomotion and Interaction Data from an Animal's Perspective
- DETReg: Unsupervised Pretraining with Region Priors for Object Detection
- Visual Prompting via Image Inpainting
- Sequential Modeling Enables Scalable Learning for Large Vision Models
- Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation
- Egocentric 4D Perception (EGO4D)

