2023
Conference article  Open Access

SegmentCodeList: unsupervised representation learning for human skeleton data retrieval

Sedmidubsky J, Carrara F, Amato G

3D skeleton sequence  Segment similarity  Unsupervised feature learning  Variational AutoEncoder  Segment code list  Action retrieval 

Recent progress in pose-estimation methods enables the extraction of sufficiently-precise 3D human skeleton data from ordinary videos, which offers great opportunities for a wide range of applications. However, such spatio-temporal data are typically extracted in the form of a continuous skeleton sequence without any information about semantic segmentation or annotation. To make the extracted data reusable for further processing, there is a need to access them based on their content. In this paper, we introduce a universal retrieval approach that compares any two skeleton sequences based on temporal order and similarities of their underlying segments. The similarity of segments is determined by their content-preserving low-dimensional code representation that is learned using the Variational AutoEncoder principle in an unsupervised way. The quality of the proposed representation is validated in retrieval and classification scenarios; our proposal outperforms the state-of-the-art approaches in effectiveness and reaches speed-ups up to 64x on common skeleton sequence datasets.


Metrics



Back to previous page
BibTeX entry
@inproceedings{oai:it.cnr:prodotti:479562,
	title = {SegmentCodeList: unsupervised representation learning for human skeleton data retrieval},
	author = {Sedmidubsky J and Carrara F and Amato G},
	doi = {10.1007/978-3-031-28238-6_8},
	year = {2023}
}

AI4Media
A European Excellence Centre for Media, Society and Democracy


OpenAIRE