2023
Conference article  Open Access

SegmentCodeList: unsupervised representation learning for human skeleton data retrieval

Sedmidubsky J., Carrara F., Amato G.

3D skeleton sequence  Segment similarity  Unsupervised feature learning  Variational AutoEncoder  Segment code list  Action retrieval 

Recent progress in pose-estimation methods enables the extraction of sufficiently-precise 3D human skeleton data from ordinary videos, which offers great opportunities for a wide range of applications. However, such spatio-temporal data are typically extracted in the form of a continuous skeleton sequence without any information about semantic segmentation or annotation. To make the extracted data reusable for further processing, there is a need to access them based on their content. In this paper, we introduce a universal retrieval approach that compares any two skeleton sequences based on temporal order and similarities of their underlying segments. The similarity of segments is determined by their content-preserving low-dimensional code representation that is learned using the Variational AutoEncoder principle in an unsupervised way. The quality of the proposed representation is validated in retrieval and classification scenarios; our proposal outperforms the state-of-the-art approaches in effectiveness and reaches speed-ups up to 64x on common skeleton sequence datasets.

Source: ECIR 2023 - 45th European Conference on Information Retrieval, pp. 110–124, Dublin, Ireland, 2-6/4/2023


Metrics



Back to previous page
BibTeX entry
@inproceedings{oai:it.cnr:prodotti:479562,
	title = {SegmentCodeList: unsupervised representation learning for human skeleton data retrieval},
	author = {Sedmidubsky J. and Carrara F. and Amato G.},
	doi = {10.1007/978-3-031-28238-6_8},
	booktitle = {ECIR 2023 - 45th European Conference on Information Retrieval, pp. 110–124, Dublin, Ireland, 2-6/4/2023},
	year = {2023}
}

AI4Media
A European Excellence Centre for Media, Society and Democracy


OpenAIRE