Chalearn LAP IsoGD Database
Owing to the limited amount of training samples on the released gesture datasets, it is hard to apply them on real applications. Therefore, we have built a large-scale gesture dataset: Chalearn LAP RGB-D Isolated Gesture Dataset (IsoGD). The focus of the challenges is "large-scale" learning and "user independent", which means vgestures per each class are more than 200 RGB and depth videos, and training samples from the same person do not appear in the validation and testing sets. The Chalearn LAP IsoGD dataset are derived from the Chalearn Gesture Dataset (CGD)  that is used on "one-shot-learning". Because the CGD dataset has totally more than 54,000 gestures which are split into subtasks. To reuse the CGD dataset, we finally obtained 249 gesture labels and manually labeled temporal segmentation to obtain the start and end frames for each gesture in continuous videos from the CGD dataset.
Database Infomation and Format
This database includes 47933 RGB-D gesture videos (about 9G). Each RGB-D video represents one gesture only, and there are 249 gestures labels performed by 21 different individuals.
The database has been divided to three sub-datasets for the convenience of using, and these three subsets are mutually exclusive.
|Sets||# of Labels||# of Gestures||# of RGB Vidoes||# of Depth Vidoes||# of Performers||Label Provided|
The test protocol are shown in the above table, which includes three subsets: train.txt for training set, valid.txt for validation set and test.txt for testing set.
train.txt ==> Training Set . Each row format: RGB_video_name depth_video_name GestureLabel
valid.txt ==> Validation Set. Each row format: RGB_video_name depth_video_name
test.txt ==> Testing Set. Each row fromat: RGB_videoname depth_video_name
The validation and testing sets are only provided the samples, no labels. Gesure labels are ranged from 1 to 249, if labels of videos were provided for that sub-dataset. Please note that the index of the labels starts from 1.
1) Isolate gesture recognition using RGB and depth videos
2) Large-scale Learning
3) User Independent: the uses in training set will not disappear in testing and validation set.
Publication and Result
To use both datasets please cite:
Jun Wan, Yibing Zhao, Shuai Zhou, Isabelle Guyon, Sergio Escalera and Stan Z. Li, "ChaLearn Looking at People RGB-D Isolated and Continuous Datasets for Gesture Recognition", CVPR workshop, 2016. [PDF]
The above reference should be cited in all documents and papers that report experimental results based on the Chalearn LAP IsoGD Database.
To obtain the database, please follow the steps below:
- Download and print the document Agreement for using Chalearn LAP IsoGD
- Sign the agreement
- Send the agreement to firstname.lastname@example.org
- Check your email to find a login account and a password of our website after one day, if your application has been approved.
- Download the Chalearn LAP IsoGD database from our website with the authorized account within 48 hours.
Copyright Note and Contacts
The database is released for research and educational purposes. We hold no liability for any undesirable consequences of using the database. All rights of the Chalearn LAP IsoGD are reserved.
 Guyon, I., Athitsos, V., Jangyodsuk, P., Escalante, H. & Hamner, B. (2013). Results and analysis of the chalearn gesture challenge 2012.
Room 1411, Intelligent Building
95 Zhongguancun Donglu,
Beijing 100190, China.Email:
jun.wan at ia.ac.cn
joewan10 at gmail.com