基于特征正则对抗训练的视觉跟踪对抗鲁棒性提升方法

Improving adversarial robustness of visual trackers via feature regularized adversarial training

  • 摘要: 该文首先基于对抗特征与原始特征在不同卷积尺度下产生分离这一经验观察,提出了特征正则化损失以使二者在特征空间中实现对齐。其次,针对目标跟踪任务的双图像输入特性,创新性地设计了适配深度跟踪网络的对抗训练框架。该框架利用特征正则化损失指导对抗样本生成,有效引导网络在对抗训练中学习鲁棒特征表示,从而显著提升目标跟踪模型的对抗鲁棒性。最后,在公开数据集上的对比实验证明,提出的方法能够在自适应攻击场景下获得最优的性能,同时能够在异构跟踪器间实现有效迁移,且在干净样本上保持有限的精度损失。

     

    Abstract: Visual object tracking (VOT), as a crucial downstream task in computer vision, has consistently garnered significant attention due to its widespread applications. In recent years, adversarial attack methods for VOT have emerged, which disrupt tracker predictions by injecting adversarial perturbations into input data. However, corresponding adversarial defense approaches remain scarce and suffer from multiple limitations: inadequate defense performance against adaptive attacks, excessive computational overhead introduced by preprocessing modules, and poor transferability across heterogeneous trackers. To address these challenges, this paper proposes a feature regularization loss based on the empirical observation that adversarial features and original features exhibit divergence across different convolutional scales, aiming to achieve feature space alignment between them. Second, considering the dual-image input characteristics of visual tracking tasks, an adversarial training framework tailored for visual tracker is designed. This framework effectively guides the network to learn robust feature representations by leveraging the feature regularization loss, thereby enhancing the adversarial robustness of the tracker. Finally, comparative experiments on public benchmarks demonstrate that our method achieves state-of-the-art performance under adaptive attack scenarios while maintaining limited accuracy degradation on clean samples. Notably, the proposed approach exhibits superior transferability across heterogeneous tracking architectures compared to existing defense methods.

     

/

返回文章
返回