Bitte benutzen Sie diese Referenz, um auf diese Ressource zu verweisen:
doi:10.22028/D291-46349
Titel: | A Performance Study of Deep Neural Network Representations of Interpretable ML on Edge Devices with AI Accelerators |
VerfasserIn: | Schauer, Julian Goodarzi, Payman Morsch, Jannis Schütze, Andreas |
Sprache: | Englisch |
Titel: | Sensors |
Bandnummer: | 25 |
Heft: | 18 |
Verlag/Plattform: | MDPI |
Erscheinungsjahr: | 2025 |
Freie Schlagwörter: | edge computing smart sensors interpretable ML AI accelerator latency energy efficiency |
DDC-Sachgruppe: | 500 Naturwissenschaften |
Dokumenttyp: | Journalartikel / Zeitschriftenartikel |
Abstract: | With the rising adoption of machine learning (ML) and deep learning (DL) applications, the demand for deploying these algorithms closer to sensors has grown significantly, particularly in sensor-driven use cases such as predictive maintenance (PM) and condition monitoring (CM). This study investigated a novel application-oriented approach to representing interpretable ML inference as deep neural networks (DNNs) regarding the latency and energy efficiency on the edge, to tackle the problem of inefficient, high-effort, and uninterpretable-implementation ML algorithms. For this purpose, the interpretable deep neural network representation (IDNNRep) was integrated into an open-source interpretable ML toolbox to demonstrate the inference time and energy efficiency improvements. The goal of this work was to enable the utilization of generic artificial intelligence (AI) accelerators for interpretable ML algorithms to achieve efficient inference on edge hardware in smart sensor applications. This novel approach was applied to one regression and one classification task from the field of PM and validated by implementing the inference on the neural processing unit (NPU) of the QXSP-ML81 Single-Board Computer and the tensor processing unit (TPU) of the Google Coral. Different quantization levels of the implementation were tested against common Python and C++ implementations. The novel implementation reduced the inference time by up to 80% and the mean energy consumption by up to 76% at the lowest precision with only a 0.4% loss of accuracy compared to the C++ implementation. With the successful utilization of generic AI accelerators, the performance was further improved with a 94% reduction for both the inference time and the mean energy consumption. |
DOI der Erstveröffentlichung: | 10.3390/s25185681 |
URL der Erstveröffentlichung: | https://doi.org/10.3390/s25185681 |
Link zu diesem Datensatz: | urn:nbn:de:bsz:291--ds-463497 hdl:20.500.11880/40641 http://dx.doi.org/10.22028/D291-46349 |
ISSN: | 1424-8220 |
Datum des Eintrags: | 1-Okt-2025 |
Fakultät: | NT - Naturwissenschaftlich- Technische Fakultät |
Fachrichtung: | NT - Systems Engineering |
Professur: | NT - Prof. Dr. Andreas Schütze |
Sammlung: | SciDok - Der Wissenschaftsserver der Universität des Saarlandes |
Dateien zu diesem Datensatz:
Datei | Beschreibung | Größe | Format | |
---|---|---|---|---|
sensors-25-05681.pdf | 4,24 MB | Adobe PDF | Öffnen/Anzeigen |
Diese Ressource wurde unter folgender Copyright-Bestimmung veröffentlicht: Lizenz von Creative Commons