TY - JOUR
T1 - Benchmarking edge computing devices for grape bunches and trunks detection using accelerated object detection single shot multibox deep learning models
AU - Magalhães, Sandro Costa
AU - dos Santos, Filipe Neves
AU - Machado, Pedro
AU - Moreira, António Paulo
AU - Dias, Jorge
N1 - Funding Information:
Sandro Costa Magalhães was granted by the Portuguese funding agency, Fundação para a Ciência e Tecnologia (FCT), and the European Social Fund (ESF) under scholarship SFRH/BD/147117/2019. This work was supported by the European Union's Horizon 2020 Research and Innovation Program under Grant 101004085.
Funding Information:
Sandro Costa Magalhães was granted by the Portuguese funding agency, Fundação para a Ciência e Tecnologia (FCT) , and the European Social Fund (ESF) under scholarship SFRH/BD/147117/2019 . This work was supported by the European Union’s Horizon 2020 Research and Innovation Program under Grant 101004085 .
Publisher Copyright:
© 2022 The Author(s)
PY - 2023/1
Y1 - 2023/1
N2 - Purpose: Visual perception enables robots to perceive the environment. Visual data is processed using computer vision algorithms that are usually time-expensive and require powerful devices to process the visual data in real-time, which is unfeasible for open-field robots with limited energy. This work benchmarks the performance of different heterogeneous platforms for object detection in real-time. This research benchmarks three architectures: embedded GPU—Graphical Processing Units (such as NVIDIA Jetson Nano 2GB and 4GB, and NVIDIA Jetson TX2), TPU—Tensor Processing Unit (such as Coral Dev Board TPU), and DPU—Deep Learning Processor Unit (such as in AMD/Xilinx ZCU104 Development Board, and AMD/Xilinx Kria KV260 Starter Kit). Methods: The authors used the RetinaNet ResNet-50 fine-tuned using the natural VineSet dataset. After the trained model was converted and compiled for target-specific hardware formats to improve the execution efficiency. Conclusions and Results: The platforms were assessed in terms of performance of the evaluation metrics and efficiency (time of inference). Graphical Processing Units (GPUs) were the slowest devices, running at 3FPS to 5FPS, and Field Programmable Gate Arrays (FPGAs) were the fastest devices, running at 14FPS to 25FPS. The efficiency of the Tensor Processing Unit (TPU) is irrelevant and similar to NVIDIA Jetson TX2. TPU and GPU are the most power-efficient, consuming about 5W. The performance differences, in the evaluation metrics, across devices are irrelevant and have an F1 of about 70% and mean Average Precision (mAP) of about 60%.
AB - Purpose: Visual perception enables robots to perceive the environment. Visual data is processed using computer vision algorithms that are usually time-expensive and require powerful devices to process the visual data in real-time, which is unfeasible for open-field robots with limited energy. This work benchmarks the performance of different heterogeneous platforms for object detection in real-time. This research benchmarks three architectures: embedded GPU—Graphical Processing Units (such as NVIDIA Jetson Nano 2GB and 4GB, and NVIDIA Jetson TX2), TPU—Tensor Processing Unit (such as Coral Dev Board TPU), and DPU—Deep Learning Processor Unit (such as in AMD/Xilinx ZCU104 Development Board, and AMD/Xilinx Kria KV260 Starter Kit). Methods: The authors used the RetinaNet ResNet-50 fine-tuned using the natural VineSet dataset. After the trained model was converted and compiled for target-specific hardware formats to improve the execution efficiency. Conclusions and Results: The platforms were assessed in terms of performance of the evaluation metrics and efficiency (time of inference). Graphical Processing Units (GPUs) were the slowest devices, running at 3FPS to 5FPS, and Field Programmable Gate Arrays (FPGAs) were the fastest devices, running at 14FPS to 25FPS. The efficiency of the Tensor Processing Unit (TPU) is irrelevant and similar to NVIDIA Jetson TX2. TPU and GPU are the most power-efficient, consuming about 5W. The performance differences, in the evaluation metrics, across devices are irrelevant and have an F1 of about 70% and mean Average Precision (mAP) of about 60%.
KW - Embedded systems
KW - Heterogeneous platforms
KW - Object detection
KW - RetinaNet resNet
KW - SSD resNet
UR - http://www.scopus.com/inward/record.url?scp=85149896694&partnerID=8YFLogxK
U2 - 10.1016/j.engappai.2022.105604
DO - 10.1016/j.engappai.2022.105604
M3 - Article
AN - SCOPUS:85149896694
SN - 0952-1976
VL - 117
JO - Engineering Applications of Artificial Intelligence
JF - Engineering Applications of Artificial Intelligence
M1 - 105604
ER -