Beyond Geometry: Artistic Disparity Synthesis for Immersive 2D-to-3D
Abstract
Current 2D-to-3D conversion methods achieve geometric accuracy but are artistically deficient, failing to replicate the immersive and emotionally resonant experience of professional 3D cinema. This is because "geometric reconstruction" paradigms mistake deliberate artistic intent—such as strategic zero-plane shifts for "pop-out" effects and local depth sculpting—for data "noise" or ambiguity. This paper argues for a new paradigm: \textbf{Artistic Disparity Synthesis}, shifting the goal from physically accurate disparity estimation to artistically coherent disparity synthesis. We propose \textbf{Art3D}, a preliminary framework exploring this paradigm. Art3D uses a dual-path architecture to decouple global depth parameters (macro-intent) from local artistic effects (visual brushstrokes) and learns from professional 3D film data via indirect supervision. We also introduce a preliminary evaluation method to quantify cinematic alignment. Experiments show our approach demonstrates potential in replicating key local out-of-screen effects and aligning with the global depth styles of cinematic 3D content, laying the groundwork for a new class of artistically-driven conversion tools.