TY - JOUR
T1 - DGU-HAU
T2 - A Dataset for 3D Human Action Analysis on Utterances
AU - Park, Jiho
AU - Park, Kwangryeol
AU - Kim, Dongho
N1 - Publisher Copyright:
© 2023 by the authors.
PY - 2023/12
Y1 - 2023/12
N2 - Constructing diverse and complex multi-modal datasets is crucial for advancing human action analysis research, providing ground truth annotations for training deep learning networks, and enabling the development of robust models across real-world scenarios. Generating natural and contextually appropriate nonverbal gestures is essential for enhancing immersive and effective human–computer interactions in various applications. These applications include video games, embodied virtual assistants, and conversations within a metaverse. However, existing speech-related human datasets are focused on style transfer, so they have limitations that make them unsuitable for 3D human action analysis studies, such as human action recognition and generation. Therefore, we introduce a novel multi-modal dataset, DGU-HAU, a dataset for 3D human action on utterances that commonly occurs during daily life. We validate the dataset using a human action generation model, Action2Motion (A2M), a state-of-the-art 3D human action generation model.
AB - Constructing diverse and complex multi-modal datasets is crucial for advancing human action analysis research, providing ground truth annotations for training deep learning networks, and enabling the development of robust models across real-world scenarios. Generating natural and contextually appropriate nonverbal gestures is essential for enhancing immersive and effective human–computer interactions in various applications. These applications include video games, embodied virtual assistants, and conversations within a metaverse. However, existing speech-related human datasets are focused on style transfer, so they have limitations that make them unsuitable for 3D human action analysis studies, such as human action recognition and generation. Therefore, we introduce a novel multi-modal dataset, DGU-HAU, a dataset for 3D human action on utterances that commonly occurs during daily life. We validate the dataset using a human action generation model, Action2Motion (A2M), a state-of-the-art 3D human action generation model.
KW - 3D human action analysis
KW - human activity understanding
KW - motion capture
KW - multi-modal dataset
KW - utterance dataset
UR - http://www.scopus.com/inward/record.url?scp=85179327882&partnerID=8YFLogxK
U2 - 10.3390/electronics12234793
DO - 10.3390/electronics12234793
M3 - Article
AN - SCOPUS:85179327882
SN - 2079-9292
VL - 12
JO - Electronics (Switzerland)
JF - Electronics (Switzerland)
IS - 23
M1 - 4793
ER -