-
EgoPAT3D
An Egocentric Action Target Prediction Dataset
News
Sept 17, 2021
Dataset github access is available
here.
Sept 28, 2021
Baseline model becomes
available.
Mar 24, 2022
Our paper is available on
Arxiv.
Jun 19, 2022
Paper is accepted at CVPR 2022.
About EgoPAT3D Dataset
Data Modalities
- RGB (.png), depth, IR, IMU, temperature frames compressed in multiple
channels in Matroska format (.mkv)
- RGB color frames (.png) extracted from
the Matroska file
- RGB videos (.mp4) of the color frames
from each recording
- Point clouds (.ply) of each scene, produced using AzureKinfu
- Labeled hand/action frames (.txt)
- Hand pose inference results using Google MediaPipe Hands (.txt)
Specifications
- 15 household scenes
- 15 point cloud files (one for each scene)
- 150 total recordings(10 recordings in each scene, with different object
configurations in each recording)
- 15000 hand-object actions (100 per recording)
- ~600 min of RGBD video (~4 min per video)
- ~1,080,000 RGB frames at 30 fps
- ~900,000 hand action frames (assuming ~2 seconds per hand-object action)
Live Visitor Statistics (via RevolverMaps)