Gesture Phase Segmentation Dataset: An Extension for Development of Gesture Analysis Models

Raúl A. Sánchez-Ancajima, Sarajane Marques Peres, Javier A. López-Céspedes, José L. Saly-Rosas-solano, Ronald M. Hernández, Miguel A. Saavedra-López

Research output: Contribution to journalOriginal Articlepeer-review


In recent years, experts in theory of gesture have been showing some interest in automating the discovery of gesture information. Such an automation can help them in reducing the inherent subjectivity of gesture studies. Usually, to produce information for linguistic and psycholinguistic studies, the researchers analyze a video of people speaking and gesturing. This annotation task is costly and it is the goal of automation. Such videos compose the datasets that allow the development of automated models capable to carry out part of the analysis of gestures. In this paper, we present a detailed documentation about the Gesture Phase Segmentation Dataset, publicized in UCI Machine Learning Repository, and an extension of such dataset. Such dataset is especially prepared to be used in the development of models capable to carry out the segmentation of gestures in their phases. The extended dataset is composed by nine videos of three people gesturing and telling stories. The data was captured with Microsoft Kinect Sensor and they are represented by spatial coordinates and temporal information (velocity and acceleration). The data are labeled following four phase of gesture (preparation, stroke, hold and retraction) and rest positions.

Original languageAmerican English
Pages (from-to)139-155
Number of pages17
JournalJournal of Internet Services and Information Security
Issue number4
StateIndexed - Nov 2022

Bibliographical note

Publisher Copyright:
© 2022, Innovative Information Science and Technology Research Group. All rights reserved.


  • Dataset
  • Gesture Segmentation
  • Gesture Studies
  • Phases of Gesture


Dive into the research topics of 'Gesture Phase Segmentation Dataset: An Extension for Development of Gesture Analysis Models'. Together they form a unique fingerprint.

Cite this