Spike-event Object Detection for Neuromorphic Vision

Citation Author(s):
Yuan-Kai
Wang
Fu-Jen Catholic University
Submitted by:
Yuan-Kai Wang
Last updated:
Mon, 07/08/2024 - 15:59
DOI:
10.21227/x8x3-mw77
License:
92 Views
Categories:
0
0 ratings - Please login to submit your rating.

Abstract 

Neuromorphic vision is one of the novel research fields that studies neuromorphic cameras and spiking neural networks (SNNs) for computer vision. Instead of computing on frame-based images, spike events are streamed from neuromorphic cameras, and novel object detection algorithms have to deal with spike events to achieve detection tasks. In this paper, we propose a solution of the novel object detection method with spike events. Spike events are first decoded to event images according to the computational methodology of neuromorphic theory. The event images can be realized as change-detected images of moving objects with a high frame rate. A redesigned deep learning framework is proposed for the object detection to deal with the event images. We propose a deep SNN method that can be realized by the conversion of successful convolution neural networks but trained by event images. We also design a methodology to build event-image datasets by object tracking algorithms. The proposed solution therefore includes spike event decoding, a redesigned deep SNN, and an event-image dataset algorithm. Experiments are conducted not only on the MNIST-DVS dataset, which is a benchmark dataset for the study of neuromorphic vision, but also on our event pedestrian detection dataset. The experimental results show that the performance of our automatic labeling algorithm is close to the model trained on manual labeled data. Moreover, the algorithm efficiency can be further improved with the PAFBenchmark dataset. The model comparison result shows that our proposed model has higher accuracy than existing SNN methods, better energy efficiency, and lower energy consumption than existing CNN methods. It demonstrates that our deep SNN method is a feasible solution for the study of neuromorphic vision. The intuition that deep SNN trained with more learning data can achieve better accuracy is also confirmed in this brand new research field. 

Instructions: 

The FJU pedestrian detection (FJUPD) dataset is the main dataset used in our paper "Spike-event Object Detection for Neuromorphic Vision". FJUPD is captured by DAVIS346 and has event images with a size of 346*260.The FJUPD records pedestrians in three scales and is fabricated to evaluate the effects of semi-automatic labeling and the deep SNN model in the published IEEE Access paper.The event images in the dataset are decoded by the surface of active events method (SAE), with time steps ranging from 10,000 to 20,000.The FJUPD has two background situations and three object scales.

The dataset download operation will get the compressed ZIP file of the FJUPD dataset. There are 3 folders after decompressing the zip file: raw, labeled, paper.  In the "raw" folder, according to the different situations, the FJUPD is split into two folds. Each fold contains SAE decoded images, with time steps ranging from 10,000 to 20,000, and the original event data (.aedat4). In the "labeled" folder, there are semi-automatic labels by discriminative correlation filter with channel and spatial reliability (CSR-DCF) and the corresponding SAE decoded images. In addition to the "raw" and "labeled" folders, there is a folder named "paper" in the FJUPD. It contains training, testing, and validation sets for the published IEEE Access paper.