Publication

Egomotion from event-based SNN optical flow

Conference Article

Conference

ACM International Conference on Neuromorphic Systems (ACM-ICONS)

Edition

2023

Pages

8:1-8

Doc link

https://doi.org/10.1145/3589737.3605978

File

Download the digital copy of the doc pdf document

Abstract

We present a method for computing egomotion using event cameras with a pre-trained optical flow spiking neural network (SNN). To address the aperture problem encountered in the sparse and noisy normal flow of the initial SNN layers, our method includes a sliding-window bin-based pooling layer that computes a fused full flow estimate. To add robustness to noisy flow estimates, instead of computing the egomotion from vector averages, our method optimizes the intersection of constraints. The method also includes a RANSAC step to robustly deal with outlier flow estimates in the pooling layer. We validate our approach on both simulated and real scenes and compare our results favorably to the state-of-the-art methods. However, our method may be sensitive to datasets and motion speeds different from those used for training, limiting its generalizability.

Categories

computer vision.

Author keywords

spiking neural network, event camera, optical flow, egomotion

Scientific reference

Y. Tian and J. Andrade-Cetto. Egomotion from event-based SNN optical flow, 2023 ACM International Conference on Neuromorphic Systems, 2023, Santa Fe, NM, USA, pp. 8:1-8.