Robust gait-based gender classification using depth cameras

Other authors

Universitat de Barcelona

Universitat Autònoma de Barcelona

Universitat Oberta de Catalunya (UOC)

Publication date

2019-03-22T09:56:41Z

2019-03-22T09:56:41Z

2013-01-02



Abstract

This article presents a new approach for gait-based gender recognition using depth cameras, that can run in real time. The main contribution of this study is a new fast feature extraction strategy that uses the 3D point cloud obtained from the frames in a gait cycle. For each frame, these points are aligned according to their centroid and grouped. After that, they are projected into their PCA plane, obtaining a representation of the cycle particularly robust against view changes. Then, final discriminative features are computed by first making a histogram of the projected points and then using linear discriminant analysis. To test the method we have used the DGait database, which is currently the only publicly available database for gait analysis that includes depth information. We have performed experiments on manually labeled cycles and over whole video sequences, and the results show that our method improves the accuracy significantly, compared with state-of-the-art systems which do not use depth information. Furthermore, our approach is insensitive to illumination changes, given that it discards the RGB information. That makes the method especially suitable for real applications, as illustrated in the last part of the experiments section.

Document Type

Article


Published version

Language

English

Publisher

EURASIP Journal on Image and Video Processing

Related items

EURASIP Journal on Image and Video Processing, 2013, 2013(1)

https://jivp-eurasipjournals.springeropen.com/articles/10.1186/1687-5281-2013-1

Recommended citation

Igual, L., Lapedriza, A. & Borràs, R. (2013). Robust gait-based gender classification using depth cameras. EURASIP Journal on Image and Video Processing, 2013(1). doi: 10.1186/1687-5281-2013-1

1687-5176

10.1186/1687-5281-2013-1

This item appears in the following Collection(s)

Articles [361]