Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/174147
Title: GaitSpike: event-based gait recognition with spiking neural network
Authors: Tao, Ying
Chang, Chip Hong
Sa¨ıgh, Sylvain
Gao, Shengyu
Keywords: Engineering
Issue Date: 2024
Source: Tao, Y., Chang, C. H., Sa¨ıgh, S. & Gao, S. (2024). GaitSpike: event-based gait recognition with spiking neural network. 2024 IEEE 6th International Conference on Artificial Intelligence Circuits and Systems (AICAS), 357-361. https://dx.doi.org/10.1109/AICAS59952.2024.10595896
Project: MOE-T2EP50220-0003 
Conference: 2024 IEEE 6th International Conference on Artificial Intelligence Circuits and Systems (AICAS)
Abstract: Existing vision-based gait recognition systems are mostly designed based on video footage acquired with RGB cameras. Appearance-, model- and motion-based techniques commonly used by these systems require silhouette segmentation, skeletal contour detection and optical flow patterns, respectively for features extraction. The extracted features are typically classified by convolutional neural networks to identify the person. These preprocessing steps are computationally intensive due to the high visual data redundancies and their accuracies can be influenced by background variations and non-locomotion related external factors. In this paper, we propose GaitSpike, a new gait recognition system that synergistically combines the advantages of sparsity-driven event-based camera and spiking neural network (SNN) for gait biometric classification. Specifically, a domain-specific locomotion-invariant representation (LIR) is proposed to replace the static Cartesian coordinates of the raw address event representation of the event camera to a floating polar coordinate reference to the motion center. The aim is to extract the relative motion information between the motion center and other human body parts to minimize the intra-class variance to promote the learning of inter-class features by the SNN. Experiments on a real event-based gait dataset DVS128-Gait and a synthetic event-based gait dataset EV-CASIA-B show that GaitSpike achieves comparable accuracy as RGB camera based gait recognition systems with higher computational efficiency, and outperforms the state-of-the-art event camera based gait recognition systems.
URI: https://hdl.handle.net/10356/174147
URL: https://aicas2024.org/
ISBN: 979-8-3503-8363-8
ISSN: 2834-9857
DOI: 10.1109/AICAS59952.2024.10595896
Schools: School of Electrical and Electronic Engineering 
Organisations: CNRS@CREATE 
Rights: © 2024 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/AICAS59952.2024.10595896.
Fulltext Permission: open
Fulltext Availability: With Fulltext
Appears in Collections:EEE Conference Papers

Files in This Item:
File Description SizeFormat 
2024037707.pdf952.95 kBAdobe PDFThumbnail
View/Open

Page view(s)

153
Updated on Jan 13, 2025

Download(s) 50

147
Updated on Jan 13, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.