Please use this identifier to cite or link to this item: doi:10.22028/D291-40347
Volltext verfügbar? / Dokumentlieferung
Title: Dilated Temporal Fully-Convolutional Network for Semantic Segmentation of Motion Capture Data
Author(s): Cheema, Noshaba
Hosseini, Somayeh
Sprenger, Janis
Herrmann, Erik
Du, Han
Fischer, Klaus
Slusallek, Philipp
Language: English
Publisher/Platform: arXiv
Year of Publication: 2018
DDC notations: 004 Computer science, internet
400 Language, linguistics
Publikation type: Other
Abstract: Semantic segmentation of motion capture sequences plays a key part in many data-driven motion synthesis frameworks. It is a preprocessing step in which long recordings of motion capture sequences are partitioned into smaller segments. Afterwards, additional methods like statistical modeling can be applied to each group of structurally-similar segments to learn an abstract motion manifold. The segmentation task however often remains a manual task, which increases the effort and cost of generating large-scale motion databases. We therefore propose an automatic framework for semantic segmentation of motion capture data using a dilated temporal fully-convolutional network. Our model outperforms a state-of-the-art model in action segmentation, as well as three networks for sequence modeling. We further show our model is robust against high noisy training labels.
DOI of the first publication: 10.48550/arXiv.1806.09174
URL of the first publication: https://arxiv.org/abs/1806.09174
Link to this record: urn:nbn:de:bsz:291--ds-403477
hdl:20.500.11880/36294
http://dx.doi.org/10.22028/D291-40347
Date of registration: 21-Aug-2023
Notes: Poster
Faculty: MI - Fakultät für Mathematik und Informatik
Department: MI - Informatik
Professorship: MI - Prof. Dr. Philipp Slusallek
Collections:SciDok - Der Wissenschaftsserver der Universität des Saarlandes

Files for this record:
There are no files associated with this item.


Items in SciDok are protected by copyright, with all rights reserved, unless otherwise indicated.