Ego-Exo4D
E
Ego Exo4D
Overview :
Ego-Exo4D is a multimodal, multi-view video dataset and benchmark challenge focused on capturing first-person and external perspectives of skill-based human activities. It supports multi-modal machine perception research for daily activities. The dataset was collected by 839 volunteers wearing cameras in 13 cities worldwide, capturing 1422 hours of skill-based human activity videos. The dataset provides three types of paired video-aligned natural language datasets: expert annotations, participant-provided tutorial-style narratives, and one-sentence atomic action descriptions. Ego-Exo4D also captures multi-view and multi-sensory modalities, including multiple cameras, seven microphone arrays, two IMUs, a barometer, and a magnetometer. The dataset was recorded strictly adhering to privacy and ethical policies with informed consent from participants. For more information, please visit the official website.
Target Users :
Supports multi-modal machine perception research for video analysis and understanding of daily activities
Total Visits: 3.0K
Top Region: US(94.96%)
Website Views : 59.6K
Features
Multimodal Multi-view Video Dataset
Synchronized first-person and external viewpoints
Multi-sensory modalities including microphones, IMUs, barometers, etc.
Three types of natural language datasets
Supports research on multi-modal machine perception of daily activities
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase