[Zurück]


Zeitschriftenartikel:

M. Lechner, A. Jantsch:
"Blackthorn: Latency Estimation Framework for CNNs on Embedded Nvidia Platforms";
IEEE Access, 9 (2021), 21053760; S. 110074 - 110084.



Kurzfassung englisch:
With more powerful yet efficient embedded devices and accelerators being available for Deep Neural Networks (DNN), machine learning is becoming an integral part of edge computing. As the number of such devices increases, finding the best platform for a specific application has become more challenging. A common question for application developers is to find the most cost-effective combination of a DNN and a device while still meeting latency and accuracy requirements. In this work, we propose Blackthorn, a layer-wise latency estimation framework for embedded Nvidia GPUs based on analytical models. We provide accurate predictions for each layer, helping developers to find bottlenecks and optimize the architecture of a DNN to fit target platforms. Our framework can quickly evaluate and compare large amounts of network optimizations without needing to build time-consuming execution engines. Our experimental results on Jetson TX2 and Jetson Nano devices show a per-layer estimation error of 6.104% Root-Mean-Square-Percentage-Error (RMSPE) and 5.888% RMSPE, which significantly outperforms current state-of-the-art methods. At network level, the average latency error is below 3% for the tested DNNs.

Schlagworte:
Hardware;Computational Modeling;Estimation;Benchmark Testing;Computer Architecture;Tools;Analytical Models;Estimation;Neural Network Hardware


"Offizielle" elektronische Version der Publikation (entsprechend ihrem Digital Object Identifier - DOI)
http://dx.doi.org/10.1109/ACCESS.2021.3101936


Erstellt aus der Publikationsdatenbank der Technischen Universität Wien.