Continúan las tareas de actualización de RIUMA, NO SE DEBEN REALIZAR ENVÍOS HASTA NUEVO AVISO.

Foreground detection by probabilistic modeling of the features discovered by stacked denoising autoencoders in noisy video sequences

Research Projects

Organizational Units

Journal Issue

Center

Abstract

A robust foreground detection system is presented, which is resilient to noise in video sequences. The proposed model divides each video frame in patches that are fed to a stacked denoising autoencoder, which is responsible for the extraction of significant features from each image patch. After that, a probabilistic model that is composed of a mixture of Gaussian distributions decides whether the given feature vector describes a patch belonging to the background or the foreground. In order to test the model robustness, several trials with noise of different types and intensities have been carried out. A comparison with other ten state of the art foreground detection algorithms has been drawn. The algorithms have been ranked according to the obtained results, and our proposal appears among the first three positions in most case and its the one that best performs on average.

Description

Licencia CC BY-NC-ND. Versión definitiva disponible en el DOI indicado. García-González, J., Ortiz-de-Lazcano-Lobato, J. M., Luque-Baena, R. M., Molina-Cabello, M. A., & López-Rubio, E. (2019). Foreground detection by probabilistic modeling of the features discovered by stacked denoising autoencoders in noisy video sequences. Pattern Recognition Letters, 125, 481-487.

Bibliographic citation

García-González, J., Ortiz-de-Lazcano-Lobato, J. M., Luque-Baena, R. M., Molina-Cabello, M. A., & López-Rubio, E. (2019). Foreground detection by probabilistic modeling of the features discovered by stacked denoising autoencoders in noisy video sequences. Pattern Recognition Letters, 125, 481–487.

Collections

Endorsement

Review

Supplemented By

Referenced by

Creative Commons license

Except where otherwised noted, this item's license is described as Attribution-NonCommercial-NoDerivatives 4.0 Internacional