RT Journal Article T1 Multimodal feature fusion for CNN-based gait recognition: an empirical comparison A1 Castro, Francisco M. A1 Marín-Jiménez, Manuel J. A1 Guil-Mata, Nicolás A1 Pérez de la Blanca, Nicolás K1 Arquitectura de ordenadores AB This paper focuses on identifying people based on their gait using a non-invasive approach. Traditional methods rely on gait signatures derived from binary energy maps, which introduce noise. Instead, the authors explore the use of raw pixel data and compare different Convolutional Neural Network (CNN) architectures across three modalities: gray pixels, optical flow, and depth maps. Tested on the TUM-GAID and CASIA-B datasets, the study finds that (i) raw pixel values are competitive with traditional silhouette-based features, (ii) combining pixel data with optical flow and depth maps yields state-of-the-art results even at lower image resolutions, and (iii) the choice of CNN architecture significantly impacts performance. PB Springer London YR 2020 FD 2020 LK https://hdl.handle.net/10630/32705 UL https://hdl.handle.net/10630/32705 LA eng NO Castro, F.M., Marín-Jiménez, M.J., Guil, N. et al. Multimodal feature fusion for CNN-based gait recognition: an empirical comparison. Neural Comput & Applic 32, 14173–14193 (2020). https://doi.org/10.1007/s00521-020-04811-z DS RIUMA. Repositorio Institucional de la Universidad de Málaga RD 20 ene 2026