← Back to Articles
⚠️
IEEE Published Article
This article is published by IEEE and the copyright belongs to IEEE. Please click here to access the full text.

Analysis of Neural Network Inference Response Times on Embedded Platforms

View PDF

Abstract

The response time of Artificial Neural Network (ANN)-inference is of utmost importance in embedded applications, particularly continual stream-processing. Predictive maintenance applications require timely predictions of state changes. This study serves to enable the reader to estimate the response time of a given model based on the underlying platform, and emphasizes the relevance of benchmarking generic ANN applications on edge devices. We analyze the influence of net parameters, activation functions as well as single-and multithreading on execution times. Potential side effects such as tact rate variances or other hardware-related influences are being outlined and accounted for. The results underline the complexity of task-partitioning and scheduling strategies while emphasizing the necessity of precise concertation of the parameters to achieve optimal performance on any platform. This study shows that cutting-edge frameworks don’t necessarily perform the required concertations automatically for all configurations, which may negatively impact performance.

Keywords

ANN inference tensorflow lite embedded systems benchmarking response times

Authors

P. Huber
University of Applied Sciences, Kempten&Technical University of Munich, Kempten, Germany
U. Göhner
University of Applied Sciences Kempten, Kempten, Germany
M. Trapp
Fraunhofer Institute for Cognitive Systems&Technical University of Munich, Munich, Germany
J. Zender
University of Applied Sciences Kempten, Kempten, Germany
R. Lichtenberg
University of Applied Sciences Kempten, Kempten, Germany

Publication Details

Type
proceedings
Publisher
IEEE
Volume
Issue
ISSN
Citations
1
Views
0