Sparse–View Localization via Online Neural 3D Regression
Ludvig Dillén ⋅ Magnus Oskarsson ⋅ Viktor Larsson
Abstract
We present ON3R, an online-trained neural regressor addressing sparse-view structureless localization, where database images have limited visual overlap and no prebuilt 3D map. Given any sparse matches between a query and a $K$-tuple of posed database views, ON3R predicts 3D coordinates for matched query keypoints, supervised by database reprojection residuals and a monocular depth prior. Afterwards, the absolute pose of the query is estimated via P3P-RANSAC and refined with lightweight bundle adjustment. Across MegaDepth, Cambridge Landmarks, and a sparsified version of Aachen Day-Night, ON3R outperforms existing methods. ON3R is particularly effective when the data is extremely sparse -- we focus on $K\leq10$ database images. The code, data splits, and SfM models will be made available for full reproducibility.
Successful Page Load