TNT15 Dataset

Multimodal Motion Capture Dataset (TNT15)

Video-based human motion capture has been a very active research area for decades now. The articulated structure of the human body, occlusions, partial observations and image ambiguities makes it very hard to accurately track the high number of degrees of freedom of the human pose. Recent approaches have shown, that adding sparse orientation cues from Inertial Measurement Units (IMUs) helps to disambiguate and improves full-body human motion capture. As a complementary data source, inertial sensors allow for accurate estimation of limb orientations even under fast motions.
In the research landscape of marker-less motion capture, publicly available benchmarks for video-based trackers (e.g. HumanEva, Human3.6M) generally lack inertial data. One exception is the MPI08 dataset, which provides inertial data of 5 IMUs along with video data.
This new dataset, called TNT15, consists of synchronized data streams from 8 RGB-cameras and 10 IMUs. In contrast to MPI08 it has been recorded in a normal office room environment and the high number of 10 IMUs can be used for new tracking approaches or improved evaluation purposes.

3D scans

8 RGB-cameras

10 IMUs

The data recordings comprise four actors performing five activities:

walking,

running on the spot,

rotating arms,

jumping and skiing exercises,

dynamic punching.

In total, the dataset contains more than 4:30 minutes of video data, which amounts to almost 13 thousand frames at a frame rate of 50 Hz. Note that the videos are compressed with high quality, however if you would like to obtain the sequences as lossless png images please contact marcard@tnt.uni-hannover.de.

The TNT15 dataset used in the TPAMI article is freely available for your own tests and experiments. However it is restricted to research purposes only. If you use this data, please acknowledge the effort that went into data collection by citing the corresponding article Human Pose Estimation from Video and IMUs BibTeX.