The Ego-Exo4D consortium, consisting of FAIR and university partners, has conducted a comprehensive study involving over 800 skilled participants across various countries to capture perspectives on human activities. The consortium is set to open source the collected data and annotations for use in novel benchmark tasks, with a public benchmark challenge planned for the next […]
- The consortium has gathered diverse perspectives on skilled human activities using over 1,400 hours of video data from skilled participants in various countries.
- The released data and annotations are intended to facilitate research in AI understanding of human skills in video, with potential applications in augmented reality systems, robot learning, and social networks.
- Existing datasets and learning paradigms are limited in their ability to capture and understand the fluid interaction between the first- and third-person perspectives of human activities, highlighting the need for more comprehensive and diverse datasets in this area.