For Full-Text PDF, please login, if you are a member of IEICE,|
or go to Pay Per View on menu list, if you are a nonmember of IEICE.
Discrimination between Genuine and Cloned Gait Silhouette Videos via Autoencoder-Based Training Data Generation
Yuki HIROSE Kazuaki NAKAMURA Naoko NITTA Noboru BABAGUCHI
IEICE TRANSACTIONS on Information and Systems
Publication Date: 2019/12/01
Online ISSN: 1745-1361
Type of Manuscript: PAPER
Category: Pattern Recognition
discrimination of cloned data, gait silhouette clone, training data generation, gait recognition, anti-spoofing,
Full Text: PDF(2.4MB)>>
Spoofing attacks are one of the biggest concerns for most biometric recognition systems. This will be also the case with silhouette-based gait recognition in the near future. So far, gait recognition has been fortunately out of the scope of spoofing attacks. However, it is becoming a real threat with the rapid growth and spread of deep neural network-based multimedia generation techniques, which will allow attackers to generate a fake video of gait silhouettes resembling a target person's walking motion. We refer to such computer-generated fake silhouettes as gait silhouette clones (GSCs). To deal with the future threat caused by GSCs, in this paper, we propose a supervised method for discriminating GSCs from genuine gait silhouettes (GGSs) that are observed from actual walking people. For training a good discriminator, it is important to collect training datasets of both GGSs and GSCs which do not differ from each other in any aspect other than genuineness. To this end, we propose to generate a training set of GSCs from GGSs by transforming them using multiple autoencoders. The generated GSCs are used together with their original GGSs for training the discriminator. In our experiments, the proposed method achieved the recognition accuracy of up to 94% for several test datasets, which demonstrates the effectiveness and the generality of the proposed method.