Skip to yearly menu bar Skip to main content


Poster

MEET: Towards Memory-Efficient Temporal Delta-Sigma Deep Neural Networks

Zeqi Zhu · Ibrahim Batuhan Akkaya · Luc Waeijen · Egor Bondarev · Arash Pourtaherian · Orlando Moreira


Abstract: Deep Neural Networks (DNNs) are accurate but compute-intensive, leading to substantial energy consumption during inference. Exploiting temporal redundancy through Δ-Σ convolution in video processing has proven to greatly enhance computation efficiency. However, temporal Δ-Σ DNNs typically require substantial memory for storing neuron states to compute inter-frame differences, hindering their on-chip deployment. To mitigate this memory cost, directly compressing the states can disrupt the linearity of temporal Δ-Σ convolution, causing accumulated errors in long-term Δ-Σ processing. Thus, we propose MEET, an optimization framework for MEmory-Efficient Temporal Δ-Σ DNNs. MEET transfers the state compression challenge to a well-established weight compression problem by trading fewer activations for more weights and introduces a co-design of network architecture and suppression method to optimize for mixed spatial-temporal execution. Evaluations on three vision applications demonstrate a reduction of 5.113.3 × in total memory compared to the most computation-efficient temporal DNNs, while preserving the computation efficiency and model accuracy in long-term Δ-Σ processing. MEET facilitates the deployment of temporal Δ-Σ DNNs within on-chip memory of embedded event-driven platforms, empowering low-power edge processing.

Live content is unavailable. Log in and register to view live content