Poster
MEET: Towards Memory-Efficient Temporal Delta-Sigma Deep Neural Networks
Zeqi Zhu · Ibrahim Batuhan Akkaya · Luc Waeijen · Egor Bondarev · Arash Pourtaherian · Orlando Moreira
[
Abstract
]
Abstract:
Deep Neural Networks (DNNs) are accurate but compute-intensive, leading to substantial energy consumption during inference. Exploiting temporal redundancy through - convolution in video processing has proven to greatly enhance computation efficiency. However, temporal - DNNs typically require substantial memory for storing neuron states to compute inter-frame differences, hindering their on-chip deployment. To mitigate this memory cost, directly compressing the states can disrupt the linearity of temporal - convolution, causing accumulated errors in long-term - processing. Thus, we propose , an optimization framework for mory-fficient emporal - DNNs. MEET transfers the state compression challenge to a well-established weight compression problem by trading fewer activations for more weights and introduces a co-design of network architecture and suppression method to optimize for mixed spatial-temporal execution. Evaluations on three vision applications demonstrate a reduction of 5.113.3 in total memory compared to the most computation-efficient temporal DNNs, while preserving the computation efficiency and model accuracy in long-term - processing. MEET facilitates the deployment of temporal - DNNs within on-chip memory of embedded event-driven platforms, empowering low-power edge processing.
Live content is unavailable. Log in and register to view live content