版权所有:内蒙古大学图书馆 技术提供:维普资讯• 智图
内蒙古自治区呼和浩特市赛罕区大学西街235号 邮编: 010021
作者机构:Peng Cheng Lab Shenzhen 518000 Peoples R China City Univ Hong Kong Dept Comp Sci Hong Kong Peoples R China Xiamen Univ Sch Informat Fujian Key Lab Sensing & Comp Smart City Xiamen 361005 Peoples R China Beihang Univ Sch Comp Sci & Engn State Key Lab Virtual Real Technol & Syst Beijing 100191 Peoples R China
出 版 物:《IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY》 (IEEE Trans Circuits Syst Video Technol)
年 卷 期:2025年第35卷第5期
页 面:4271-4286页
核心收录:
基 金:National Natural Science Foundation of China [62132002, 62202249, 62102206] Major Key Project of Peng Cheng Laboratory(PCL) [PCL2024A04-4] Postdoctoral Science Foundation of China [2022M721732] Postdoctoral Fellowship Program of China Postdoctoral Science Foundation [GZC20233362] Chongqing Postdoctoral Innovative Talents Support Program
主 题:Transformers Target tracking Visualization Semantics Feature extraction Object tracking Natural languages Circuits and systems Fuses Adaptation models Vision-language tracking progressive joint vision-language transformer semantic-aware instance encoder channel communication patch interaction
摘 要:In recent years, vision-language tracking has drawn emerging attention in the tracking field. The critical challenge for the task is to fuse semantic representations of language information and visual representations of vision information. For this purpose, several vision-language tracking methods perform early or late fusion to fuse visual and semantic features. However, these methods cannot take full advantage of the transformer architecture to excavate useful cross-modal context at various levels. To this end, we propose a new progressive joint vision-language transformer (PJVLT) to progressively align and refine visual embedding with semantic embedding for vision-language tracking. Specifically, to align visual signals with semantic signals, we propose to insert a semantic-aware instance encoder layer (SAIEL) into each intermediate layer of transformer encoder to perform progressive alignment of visual and semantic features. Furthermore, to highlight the multi-modal feature channels and patches corresponding to target objects, we propose a unified channel communication patch interaction layer (CCPIL), which is plugged into each intermediate layer of transformer encoder to progressively activate target-aware channels and patches of aligned multi-modal features for fine-grained tracking. In general, by progressively aligning and refining visual features with semantic features in the transformer encoder, our PJVLT can adaptively excavate well-aligned vision-language context at coarse-to-fine levels, therefore highlighting target objects at various levels for more discriminative tracking. Experiments on several tracking datasets show that the proposed PJVLT can achieve favorable performance in comparison with both conventional trackers and other vision-language trackers.