Parallel Jacobi Decoding for Fast Autoregressive Image Generation
Abstract
Autoregressive (AR) models have demonstrated remarkable performance in generating high-fidelity images.However, their inherently sequential next-token prediction leads to significantly slows inference.Recent studies have introduced Jacobi-style decoding to accelerate autoregressive image generation.Extending the draft sequence initially improves efficiency, yet the acceleration quickly saturates as error propagation in the one-dimensional sequence hinders convergence.Observing that images exhibit strong local spatial correlations, we propose Parallel Jacobi Decoding (PJD), a training-free decoding approach that expands draft tokens in the two-dimensional spatial domain to enable efficient spatially parallel refinement.PJD adjusts the attention mask to mitigate error accumulation and improve convergence stability.Extensive experiments on diverse datasets show that PJD achieves 4.8×–6.4× acceleration across multiple autoregressive image generation models while preserving image quality.