Human grasping database for activities of daily living with depth, color and kinematic data streams

Sci Data. 2018 May 29;5:180101. doi: 10.1038/sdata.2018.101.


This paper presents a grasping database collected from multiple human subjects for activities of daily living in unstructured environments. The main strength of this database is the use of three different sensing modalities: color images from a head-mounted action camera, distance data from a depth sensor on the dominant arm and upper body kinematic data acquired from an inertial motion capture suit. 3826 grasps were identified in the data collected during 9-hours of experiments. The grasps were grouped according to a hierarchical taxonomy into 35 different grasp types. The database contains information related to each grasp and associated sensor data acquired from the three sensor modalities. We also provide our data annotation software written in Matlab as an open-source tool. The size of the database is 172 GB. We believe this database can be used as a stepping stone to develop big data and machine learning techniques for grasping and manipulation with potential applications in rehabilitation robotics and intelligent automation.

Publication types

  • Dataset

MeSH terms

  • Activities of Daily Living*
  • Biomechanical Phenomena
  • Databases, Factual
  • Humans