Gait Recognition from Motion Capture Data
Identifikátory výsledku
Kód výsledku v IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F00216224%3A14330%2F18%3A00102051" target="_blank" >RIV/00216224:14330/18:00102051 - isvavai.cz</a>
Výsledek na webu
<a href="https://doi.org/10.1145/3152124" target="_blank" >https://doi.org/10.1145/3152124</a>
DOI - Digital Object Identifier
<a href="http://dx.doi.org/10.1145/3152124" target="_blank" >10.1145/3152124</a>
Alternativní jazyky
Jazyk výsledku
angličtina
Název v původním jazyce
Gait Recognition from Motion Capture Data
Popis výsledku v původním jazyce
Gait recognition from motion capture data, as a pattern classification discipline, can be improved by the use of machine learning. This paper contributes to the state-of-the-art with a statistical approach for extracting robust gait features directly from raw data by a modification of Linear Discriminant Analysis with Maximum Margin Criterion. Experiments on the CMU MoCap database show that the suggested method outperforms thirteen relevant methods based on geometric features and a method to learn the features by a combination of Principal Component Analysis and Linear Discriminant Analysis. The methods are evaluated in terms of the distribution of biometric templates in respective feature spaces expressed in a number of class separability coefficients and classification metrics. Results also indicate a high portability of learned features, that means, we can learn what aspects of walk people generally differ in and extract those as general gait features. Recognizing people without needing group-specific features is convenient as particular people might not always provide annotated learning data. As a contribution to reproducible research, our evaluation framework and database have been made publicly available. This research makes motion capture technology directly applicable for human recognition.
Název v anglickém jazyce
Gait Recognition from Motion Capture Data
Popis výsledku anglicky
Gait recognition from motion capture data, as a pattern classification discipline, can be improved by the use of machine learning. This paper contributes to the state-of-the-art with a statistical approach for extracting robust gait features directly from raw data by a modification of Linear Discriminant Analysis with Maximum Margin Criterion. Experiments on the CMU MoCap database show that the suggested method outperforms thirteen relevant methods based on geometric features and a method to learn the features by a combination of Principal Component Analysis and Linear Discriminant Analysis. The methods are evaluated in terms of the distribution of biometric templates in respective feature spaces expressed in a number of class separability coefficients and classification metrics. Results also indicate a high portability of learned features, that means, we can learn what aspects of walk people generally differ in and extract those as general gait features. Recognizing people without needing group-specific features is convenient as particular people might not always provide annotated learning data. As a contribution to reproducible research, our evaluation framework and database have been made publicly available. This research makes motion capture technology directly applicable for human recognition.
Klasifikace
Druh
J<sub>imp</sub> - Článek v periodiku v databázi Web of Science
CEP obor
—
OECD FORD obor
10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)
Návaznosti výsledku
Projekt
—
Návaznosti
S - Specificky vyzkum na vysokych skolach
Ostatní
Rok uplatnění
2018
Kód důvěrnosti údajů
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Údaje specifické pro druh výsledku
Název periodika
ACM Transactions on Multimedia Computing, Communications and Applications (TOMM), special issue on Representation, Analysis and Recognition of 3D Humans
ISSN
1551-6857
e-ISSN
—
Svazek periodika
14
Číslo periodika v rámci svazku
1s
Stát vydavatele periodika
US - Spojené státy americké
Počet stran výsledku
18
Strana od-do
1-18
Kód UT WoS článku
000433517100008
EID výsledku v databázi Scopus
2-s2.0-85042907000