Zikai Song1, Ying Tang1, Run Luo1, Lintao Ma1, Junqing Yu1, Yi-Ping Phoebe Chen2, WeiYang1
1 Huazhong University of Science and Technology
2 La Trobe University
paper
ALTrack is a coherent point tracking framework which designes an autogenic language embedding for visual feature enhancement, strengthens point correspondence in long-term sequences. Unlike existing visual-language schemes, our approach learns text embeddings from visual features through a dedicated mapping network, enabling seamless adaptation to various tracking tasks without explicit text annotations. Additionally, we introduce a consistency decoder that efficiently integrates text tokens into visual features with minimal computational overhead.
- The Code will be released soon!
If you find this repo useful for your research, please consider citing the paper
@inproceedings{song2024altrack,
title={Autogenic Language Embedding for Coherent Point Tracking},
author={Song, Zikai and Tang, Ying and Luo, Run and Ma, Lintao and Yu, Junqing and Chen, Yi-Ping Phoebe and Yang, Wei},
booktitle={Proceedings of the 29th ACM International Conference on Multimedia},
year={2024}
}
