The rapid growth of edge devices demands efficient neural network accelerators for low-latency, energy-efficient AI tasks like object detection and pose estimation. Despite benefits such as improved privacy and reduced cloud dependence, edge AI faces challenges including hardware heterogeneity, resource constraints, and the balance between real-time performance and accuracy. This paper benchmarks STMicroelectronics STM32N6, Luxonis OAK-FCC 4P, and SONY IMX500 on object detection and human pose estimation tasks, evaluating power consumption, inference time, accuracy, and memory usage. Using YOLOv5 and YOLOv8 models in Small and Nano configurations on COCO and COCO-Pose datasets, the study highlights the trade-offs between efficiency and accuracy. Results show IMX500 leads in inference time and power efficiency but is limited in memory compared to N6 and OAK-FCC 4P. This study provides a comprehensive evaluation of the suitability of different hardware architectures, guiding the optimization of lightweight neural network deployment on resource-constrained platforms.

Comparative Analysis of Lightweight Vision Models on Embedded Accelerator Devices

Corti G.;Veronesi C.;Bartoli P.;Giudici A.;Palermo F.;Matteucci M.;Mentasti S.
2026-01-01

Abstract

The rapid growth of edge devices demands efficient neural network accelerators for low-latency, energy-efficient AI tasks like object detection and pose estimation. Despite benefits such as improved privacy and reduced cloud dependence, edge AI faces challenges including hardware heterogeneity, resource constraints, and the balance between real-time performance and accuracy. This paper benchmarks STMicroelectronics STM32N6, Luxonis OAK-FCC 4P, and SONY IMX500 on object detection and human pose estimation tasks, evaluating power consumption, inference time, accuracy, and memory usage. Using YOLOv5 and YOLOv8 models in Small and Nano configurations on COCO and COCO-Pose datasets, the study highlights the trade-offs between efficiency and accuracy. Results show IMX500 leads in inference time and power efficiency but is limited in memory compared to N6 and OAK-FCC 4P. This study provides a comprehensive evaluation of the suitability of different hardware architectures, guiding the optimization of lightweight neural network deployment on resource-constrained platforms.
2026
Lecture Notes in Computer Science
9783032101846
9783032101853
Hardware Benchmarking
Lightweight Neural Network
Low-Power Edge Devices
Neural Network Accelerators
File in questo prodotto:
File Dimensione Formato  
978-3-032-10185-3_18.pdf

Accesso riservato

: Publisher’s version
Dimensione 1.79 MB
Formato Adobe PDF
1.79 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1312355
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact