Abstract
Resource constraints and hardware aging make long-term neural inference on embedded and IoT devices increasingly challenging. We present Neural Caching, a lightweight inference mechanism that exploits the piecewise-linear structure of Artificial Neural Networks (ANNs) with piecewise linear activation functions to accelerate evaluation without retraining or model compression. Our approach, Neural Caching, associates each locally linear region of an ANN with a cached affine mapping and reuses it for subsequent inputs falling within the same or nearby activation region. This enables full predictions to be computed via a single matrix multiplication instead of potential multiple matrix multiplications. Experiments on four human-activity recognition datasets demonstrate up to an order-of-magnitude reduction in inference time while preserving classification performance metrics within ±0.1 % of baseline accuracy. By lowering computational load and power draw, neural caching extends device lifetime and supports sustainable, long-term deployment of machine-learning models in real-world IoT environments.
Keywords: artificial neural networks, inference-time caching, hardware longevity, time-series classification
How to Cite:
Sallinger, C., Stippel, C., Panner, E., Poschenreither, P., Hoch, R. & Schwendinger, B., (2025) “Neural Caching: Improving Longevity of Smart IoT Devices running Artificial Neural Networks”, IoT Workshop Proceedings 1(1), 1-7. doi: https://doi.org/10.34749/3061-1008.2025.1
Rights: Copyright © 2025 The author(s)
Downloads:
Download PDF
95 Views
20 Downloads
