Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/174147
Title: | GaitSpike: event-based gait recognition with spiking neural network | Authors: | Tao, Ying Chang, Chip Hong Sa¨ıgh, Sylvain Gao, Shengyu |
Keywords: | Engineering | Issue Date: | 2024 | Source: | Tao, Y., Chang, C. H., Sa¨ıgh, S. & Gao, S. (2024). GaitSpike: event-based gait recognition with spiking neural network. 2024 IEEE 6th International Conference on Artificial Intelligence Circuits and Systems (AICAS), 357-361. https://dx.doi.org/10.1109/AICAS59952.2024.10595896 | Project: | MOE-T2EP50220-0003 | Conference: | 2024 IEEE 6th International Conference on Artificial Intelligence Circuits and Systems (AICAS) | Abstract: | Existing vision-based gait recognition systems are mostly designed based on video footage acquired with RGB cameras. Appearance-, model- and motion-based techniques commonly used by these systems require silhouette segmentation, skeletal contour detection and optical flow patterns, respectively for features extraction. The extracted features are typically classified by convolutional neural networks to identify the person. These preprocessing steps are computationally intensive due to the high visual data redundancies and their accuracies can be influenced by background variations and non-locomotion related external factors. In this paper, we propose GaitSpike, a new gait recognition system that synergistically combines the advantages of sparsity-driven event-based camera and spiking neural network (SNN) for gait biometric classification. Specifically, a domain-specific locomotion-invariant representation (LIR) is proposed to replace the static Cartesian coordinates of the raw address event representation of the event camera to a floating polar coordinate reference to the motion center. The aim is to extract the relative motion information between the motion center and other human body parts to minimize the intra-class variance to promote the learning of inter-class features by the SNN. Experiments on a real event-based gait dataset DVS128-Gait and a synthetic event-based gait dataset EV-CASIA-B show that GaitSpike achieves comparable accuracy as RGB camera based gait recognition systems with higher computational efficiency, and outperforms the state-of-the-art event camera based gait recognition systems. | URI: | https://hdl.handle.net/10356/174147 | URL: | https://aicas2024.org/ | ISBN: | 979-8-3503-8363-8 | ISSN: | 2834-9857 | DOI: | 10.1109/AICAS59952.2024.10595896 | Schools: | School of Electrical and Electronic Engineering | Organisations: | CNRS@CREATE | Rights: | © 2024 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/AICAS59952.2024.10595896. | Fulltext Permission: | open | Fulltext Availability: | With Fulltext |
Appears in Collections: | EEE Conference Papers |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
2024037707.pdf | 952.95 kB | Adobe PDF | View/Open |
Page view(s)
153
Updated on Jan 13, 2025
Download(s) 50
147
Updated on Jan 13, 2025
Google ScholarTM
Check
Altmetric
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.