Skip to yearly menu bar Skip to main content


Poster

Reasoning to Attend: Try to Understand How <SEG> Token Works

Rui Qian · Xin Yin · Dejing Dou

ExHall D Poster #353
[ ] [ Project Page ] [ Paper PDF ]
Sun 15 Jun 8:30 a.m. PDT — 10:30 a.m. PDT

Abstract: Current Large Multimodal Models (LMMs) empowered tasks such as visual grounding and segmentation typically rely on $\texttt{}$ token as a text prompting to jointly optimize the vision-language model (e.g., LLaVA) and the downstream task-specified model ($\eg$, SAM). However, we observe that little research has looked into how it works when mapping language vocabulary embedding into corresponding vision codebook space. In this work, we first visualize the similarity maps, $\aka$ pseudo images, which are obtained by computing the dot product similarity between the $\texttt{}$ token and the image token embedings derived from the last hidden layer in both LLaVA and SAM models. Intriguingly, we have found that a striking consistency holds in terms of activation responses in the pseudo images, which reveals that what $\texttt{}$ token contributes to is the semantic correspondences from image-text pairs. Specifically, $\texttt{}$ token, a placeholder expanded in text vocabulary, extensively queries within individual tokenized image patches to map the semantics of an object from text to the paired image while the Large Language Models (LLMs) is being fine tined. Upon above findings, we present READ, which facilitates LMMs' resilient $\textbf{REA}$soning capability of where to atten\textbf{D} under the guidance of highly activated points borrowed from pseudo images. Remarkably, READ features an intuitive design, Similarity as Points module (SasP), which can be seamlessly applied to existing $\texttt{}$-like paradigms with negligible overheads in a plug-and-play fashion. Also, extensive experiments have been conducted on highly challenging reasoning segmentation dataset and widely used RefCOCO(+/g) referring segmentation dataset. To validate whether READ suffers from catastrophic forgetting of previous skills after fine-tuning, as observed in prior works ($\eg$, LISA), we further assess its generation ability on FP-RefCOCO(+/g) dataset. All code, models will be publicly available.

Live content is unavailable. Log in and register to view live content