2020
DOI: 10.1007/s00521-020-04811-z
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal feature fusion for CNN-based gait recognition: an empirical comparison

Abstract: People identification in video based on the way they walk (i.e. gait) is a relevant task in computer vision using a non-invasive approach. Standard and current approaches typically derive gait signatures from sequences of binary energy maps of subjects extracted from images, but this process introduces a large amount of non-stationary noise, thus, conditioning their efficacy. In contrast, in this paper we focus on the raw pixels, or simple functions derived from them, letting advanced learning techniques to ex… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
39
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
2

Relationship

3
7

Authors

Journals

citations
Cited by 64 publications
(40 citation statements)
references
References 62 publications
1
39
0
Order By: Relevance
“…For example, Handojoseno et al used the wavelet coefficients of electroencephalogram (EEG) signals as the input of the Multilayer Perceptron Neural Network and k-Nearest Neighbor classifier, which can predict the transition from walking to FOG with 87% sensitivity and 73% accuracy [14]. The second group generally uses three-dimensional (3D) motion analysis systems [15][16][17][18], plantar pressure measurement systems [19][20][21] or inertial sensors (accelerometers, gyroscopes or magnetometers) [22][23][24][25] to obtain more intuitive gait kinematics or dynamics signals. Delval et al used multiple cameras to capture the gait kinematics signals of patients with reflective markers attached to their bodies from different angles [17], and Okuno et al used a plantar pressure measurement system (1.92 × 0.88 m) to record the soles of the patients walking [19].…”
Section: Related Workmentioning
confidence: 99%
“…For example, Handojoseno et al used the wavelet coefficients of electroencephalogram (EEG) signals as the input of the Multilayer Perceptron Neural Network and k-Nearest Neighbor classifier, which can predict the transition from walking to FOG with 87% sensitivity and 73% accuracy [14]. The second group generally uses three-dimensional (3D) motion analysis systems [15][16][17][18], plantar pressure measurement systems [19][20][21] or inertial sensors (accelerometers, gyroscopes or magnetometers) [22][23][24][25] to obtain more intuitive gait kinematics or dynamics signals. Delval et al used multiple cameras to capture the gait kinematics signals of patients with reflective markers attached to their bodies from different angles [17], and Okuno et al used a plantar pressure measurement system (1.92 × 0.88 m) to record the soles of the patients walking [19].…”
Section: Related Workmentioning
confidence: 99%
“…In [70], four different CNN architectures are proposed for gait feature extraction, including (i) 2D CNN: linear CNN with four 2D convolutions, (ii) 3D CNN: 3D CNN with four 3D convolutions, (iii) ResNet‐A: residual CNN with a 2D convolution and three residual blocks, and (iv) ResNet‐B: residual CNN with a 2D convolution and four residual blocks. Three different modalities are separately attempted in the CNN model, including grey pixels, optical flow channels, and depth maps.…”
Section: Gait Recognition Approachesmentioning
confidence: 99%
“…Kinect, mobiles) or techniques that produce different kinds of data like depth [7] or optical flow [8]. Thus, some works exploit multimodality and show that it leads to better representations and improved results [2,9,10,11,12]. Nevertheless, their common limitation is their inability to handle missing modalities.…”
Section: Introductionmentioning
confidence: 99%