Abstract
In event-based neuromorphic processing, computer vision finds an efficient alternative capable of optimizing computational and energy resources, inspired by the dynamics of biological neural systems. In the development of real-time processing systems, it is crucial to visually represent the information captured by sensors and to explore its content with precision. Thus, machine learning models are implemented with the capability of being deployed on hardware devices with limited capabilities, depending on the intended purpose, ensuring savings in computational resources. The aim of this work was to evaluate the limits of the implemented neuron model, leaky-integrate and fire (LIF), for fitting convolutional layers of a neural network. To this end, the characteristics of the LIF neuron model used are summarized, as well as the details of its implementation in a hardware design, using configurable parameters. The experimental phase considered two convolution approaches to compare performance, Matlab R2022a software and a spiking convolutional processor for an FPGA, using sample recordings from the MNIST-DVS dataset and Sobel kernels for edge detection. The results reflect that the number of spikes generated by both approaches is very similar and their distribution by frame addresses is directly proportional.