D 2023

Text-to-Motion Retrieval: Towards Joint Understanding of Human Motion Data and Natural Language

MESSINA, Nicola, Jan SEDMIDUBSKÝ, Falchi FABRIZIO and Tomáš REBOK

Basic information

Original name

Text-to-Motion Retrieval: Towards Joint Understanding of Human Motion Data and Natural Language

Authors

MESSINA, Nicola (380 Italy), Jan SEDMIDUBSKÝ (203 Czech Republic, guarantor, belonging to the institution), Falchi FABRIZIO (380 Italy) and Tomáš REBOK (203 Czech Republic, belonging to the institution)

Edition

New York, NY, USA, 46th International Conference on Research and Development in Information Retrieval (SIGIR), p. 2420-2425, 6 pp. 2023

Publisher

Association for Computing Machinery

Other information

Language

English

Type of outcome

Stať ve sborníku

Field of Study

10200 1.2 Computer and information sciences

Country of publisher

Czech Republic

Confidentiality degree

není předmětem státního či obchodního tajemství

Publication form

electronic version available online

References:

RIV identification code

RIV/00216224:14330/23:00130552

Organization unit

Faculty of Informatics

ISBN

978-1-4503-9408-6

UT WoS

001118084002091

Keywords in English

human motion data;skeleton sequences;CLIP;BERT;deep language models;ViViT;motion retrieval;cross-modal retrieval

Tags

International impact, Reviewed
Změněno: 14/3/2024 13:10, doc. RNDr. Jan Sedmidubský, Ph.D.

Abstract

V originále

Due to recent advances in pose-estimation methods, human motion can be extracted from a common video in the form of 3D skeleton sequences. Despite wonderful application opportunities, effective and efficient content-based access to large volumes of such spatio-temporal skeleton data still remains a challenging problem. In this paper, we propose a novel content-based text-to-motion retrieval task, which aims at retrieving relevant motions based on a specified natural-language textual description. To define baselines for this uncharted task, we employ the BERT and CLIP language representations to encode the text modality and successful spatio-temporal models to encode the motion modality. We additionally introduce our transformer-based approach, called Motion Transformer (MoT), which employs divided space-time attention to effectively aggregate the different skeleton joints in space and time. Inspired by the recent progress in text-to-image/video matching, we experiment with two widely-adopted metric-learning loss functions. Finally, we set up a common evaluation protocol by defining qualitative metrics for assessing the quality of the retrieved motions, targeting the two recently-introduced KIT Motion-Language and HumanML3D datasets. The code for reproducing our results is available here: https://github.com/mesnico/text-to-motion-retrieval.

Links

EF16_019/0000822, research and development project
Name: Centrum excelence pro kyberkriminalitu, kyberbezpečnost a ochranu kritických informačních infrastruktur