TY - GEN
T1 - Deep bidirectional correlation filters for visual object tracking
AU - Javed, Sajid
AU - Zhang, Xiaoxiong
AU - Seneviratne, Lakmal
AU - Dias, Jorge
AU - Werghi, Naoufel
N1 - Funding Information:
This publication acknowledges the support provided by the Khalifa University of Science and Technology under Award No. RC1-2018-KUCARS.
Funding Information:
This publication is based upon work supported by the Khalifa University of Science and Technology under Award No. RC1-2018-KUCARS.
Publisher Copyright:
© 2020 International Society of Information Fusion (ISIF).
PY - 2020/7
Y1 - 2020/7
N2 - Visual Object Tracking (VOT) is an essential task for many computer vision applications. VOT becomes challenging when a target object faces severe occlusion, drastic illumination changes, and scale variation problems. In the literature, Discriminative Correlation Filters (DCFs)-based tracking methods have achieved promising results in terms of accuracy and efficiency in many complex VOT scenarios. A plethora of DCFs trackers have been proposed which exploit information observed in past frames to create and update DCFs for VOT. To adapt to target appearance variations, the DCFs are enhanced by incorporating spatial and temporal consistency constraints. Nevertheless, the performance degradation is observed for these methods because of the aforementioned limitations. To address these issues, we propose a novel algorithm based on bidirectional DCFs for VOT. In this algorithm, we propose the original idea of leveraging information from both past and future frames. The proposed algorithm first tracks the target object forward in the video sequence and then its uses the predicted location of the last window frame and track the target object backward towards the current frame. We design an appearance consistency loss function by taking the L2 norm between the regression target of the forward tracking and response map of the backward tracking to obtain the resulting response map. Our proposed algorithm realizes a highly accurate DCFs because forward and backward tracking information are fused together for consistent VOT. Although, a result will be output with some small delay because information is taken from a future to the present period, our proposed algorithm has the merit of addressing the drastic appearance variations VOT challenges. We evaluate our proposed tracker using deep features on three publicly available challenging datasets. Our results demonstrate the superior performance of the proposed tracker compared to the existing state-of-the-art trackers.
AB - Visual Object Tracking (VOT) is an essential task for many computer vision applications. VOT becomes challenging when a target object faces severe occlusion, drastic illumination changes, and scale variation problems. In the literature, Discriminative Correlation Filters (DCFs)-based tracking methods have achieved promising results in terms of accuracy and efficiency in many complex VOT scenarios. A plethora of DCFs trackers have been proposed which exploit information observed in past frames to create and update DCFs for VOT. To adapt to target appearance variations, the DCFs are enhanced by incorporating spatial and temporal consistency constraints. Nevertheless, the performance degradation is observed for these methods because of the aforementioned limitations. To address these issues, we propose a novel algorithm based on bidirectional DCFs for VOT. In this algorithm, we propose the original idea of leveraging information from both past and future frames. The proposed algorithm first tracks the target object forward in the video sequence and then its uses the predicted location of the last window frame and track the target object backward towards the current frame. We design an appearance consistency loss function by taking the L2 norm between the regression target of the forward tracking and response map of the backward tracking to obtain the resulting response map. Our proposed algorithm realizes a highly accurate DCFs because forward and backward tracking information are fused together for consistent VOT. Although, a result will be output with some small delay because information is taken from a future to the present period, our proposed algorithm has the merit of addressing the drastic appearance variations VOT challenges. We evaluate our proposed tracker using deep features on three publicly available challenging datasets. Our results demonstrate the superior performance of the proposed tracker compared to the existing state-of-the-art trackers.
KW - Correlation Filters
KW - Deep Learning
KW - Visual Object Tracking
UR - http://www.scopus.com/inward/record.url?scp=85092737727&partnerID=8YFLogxK
U2 - 10.23919/FUSION45008.2020.9190209
DO - 10.23919/FUSION45008.2020.9190209
M3 - Conference contribution
AN - SCOPUS:85092737727
T3 - Proceedings of 2020 23rd International Conference on Information Fusion, FUSION 2020
BT - Proceedings of 2020 23rd International Conference on Information Fusion, FUSION 2020
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 23rd International Conference on Information Fusion, FUSION 2020
Y2 - 6 July 2020 through 9 July 2020
ER -