Vehicle Re-Identification Based on Wavelet Feature Enhancement and Global-Local Differential Attention Fusion

Authors

  • Bochi Zhu
  • Haifeng Sang

DOI:

https://doi.org/10.54097/v24nha32

Keywords:

Attribute aggregation, Swin transformer, Vehicle re-identification, Wavelet transform

Abstract

In the continuous evolution of intelligent transportation systems, vehicle re-identification technology faces numerous technical challenges, including variations in perspective and equipment resolution. These factors lead to significant intra-class discrepancies in the performance of identical vehicles under varying conditions, as well as inter-class confusion among vehicles with similar appearances. To address these challenges, we integrate vehicle color and type attribute information, enhancing the model’s ability to capture semantic features and improve its discriminative performance. Additionally, we propose a wavelet feature enhancement module that employs wavelet transform to decompose images at multiple scales, effectively capturing fine-grained features such as edges and textures. This enables the model to better represent intricate visual details. Finally, we introduce a differential attention mechanism that combines global and local features, strengthening contextual understanding through interactive feature modeling. Experimental results demonstrate the effectiveness of our approach, achieving a Rank-1 accuracy of 97.0% on the VeRi-776 dataset and 85.2% on the VehicleID dataset, outperforming existing methods and highlighting the efficacy of our proposed framework.

Downloads

Download data is not yet available.

References

[1] He S, Luo H, Wang P, Wang F, Li H, Jiang W. Transreid: Transformer-based object re-identification. In: Proceedings of the IEEE/CVF international conference on computer vision; 2021. p. 15013–15022.

[2] Alexey D. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:201011929. 2020.

[3] Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, et al. Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF international conference on computer vision; 2021. p. 10012–10022.

[4] Qian W, Luo H, Peng S, Wang F, Chen C, Li H. Unstructured feature decoupling for vehicle re-identification. In: European Conference on Computer Vision. Springer; 2022. p. 336–353.

[5] Jaderberg M, Simonyan K, Zisserman A, et al. Spatial transformer networks. Advances in neural information processing systems. 2015;28.

[6] Phan N, Huy TD, Duong ST, Hoang NT, Tran S, Hung DH, et al. Logovit: Local-global vision transformer for object re-identification. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2023. p. 1–5.

[7] Yu Z, Pei J, Zhu M, Zhang J, Li J. Multi-attribute adaptive aggregation transformer for vehicle re-identification. Information Processing & Management. 2022;59(2):102868.

[8] Simonyan K. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556. 2014.

[9] Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2015. p. 1–9.

[10] He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2016. p. 770–778.

[11] Vaswani A. Attention is all you need. Advances in Neural Information Processing Systems. 2017.

[12] Li Z, Zhang X, Tian C, Gao X, Gong Y, Wu J, et al. Tvg-reid: Transformer-based vehicle-graph re-identification. IEEE Transactions on Intelligent Vehicles. 2023.

[13] Lian J, Wang D, Zhu S, Wu Y, Li C. Transformer-based attention network for vehicle re-identification. Electronics. 2022;11(7):1016.

[14] Li H, Xiao Y, Cheng C, Song X. SFPFusion: An improved vision transformer combining super feature attention and wavelet-guided pooling for infrared and visible images fusion. Sensors. 2023;23(18):7870.

[15] Wang C, Wu J, Fang A, Zhu Z, Wang P, Chen H. An efficient frequency domain fusion network of infrared and visible images. Engineering Applications of Artificial Intelligence. 2024;133:108013.

[16] Yao T, PanY, Li Y, Ngo CW, Mei T. Wave-vit: Unifying wavelet and transformers for visual representation learning. In: European Conference on Computer Vision. Springer; 2022. p. 328–345.

[17] Jiang Y, Liu Q, Liu MT. Attribute Feature Fusion Network for Pedestrian Detection and Re-Identification. In: 2023 5th International Conference on Robotics and Computer Vision (ICRCV). IEEE; 2023. p. 36–40.

[18] Khorramshahi P, Kumar A, Peri N, Rambhatla SS, Chen JC, Chellappa R. A dual-path model with adaptive attention for vehicle re-identification. In: Proceedings of the IEEE/CVF international conference on computer vision; 2019. p. 6132–6141.

[19] He B, Li J, Zhao Y, Tian Y. Part-regularized near-duplicate vehicle re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2019. p. 3997–4005.

[20] Lou Y, Bai Y, Liu J, Wang S, Duan L. Veri-wild: A large dataset and a new method for vehicle re-identification in the wild. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2019. p. 3235–3243.

[21] Chu R, Sun Y, Li Y, Liu Z, Zhang C, Wei Y. Vehicle re-identification with viewpoint-aware metric learning. In: Proceedings of the IEEE/CVF international conference on computer vision; 2019. p. 8282–8291.

[22] Qian J, Jiang W, Luo H, Yu H. Stripe-based and attribute-aware network: A two-branch deep model for vehicle re-identification. Measurement Science and Technology. 2020;31(9):095401.

[23] Jin X, Lan C, Zeng W, Chen Z. Uncertainty-aware multi-shot knowledge distillation for image-based object re-identification. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 34; 2020. p. 11165–11172.

[24] Teng S, Zhang S, Huang Q, Sebe N. Multi-view spatial attention embedding for vehicle re-identification. IEEE Transactions on Circuits and Systems for Video Technology. 2020;31(2):816–827.

[25] Zhang F, Ma Y, Yuan G, Zhang H, Ren J. Multiview image generation for vehicle re-identification. Applied Intelligence. 2021;51(8):5665–5682.

[26] Taufique AMN, Savakis A. LABNet: Local graph aggregation network with class balanced loss for vehicle re-identification. Neurocomputing. 2021;463:122–132.

[27] Wang H, Peng J, Jiang G, Xu F, Fu X. Discriminative feature and dictionary learning with part-aware model for vehicle re-identification. Neurocomputing. 2021;438:55–62.

[28] Zhu W, Wang Z, Wang X, Hu R, Liu H, Liu C, et al. A dat self-attention mechanism for vehicle re-identification. Pattern Recognition. 2023;137:109258.

[29] Qian J, Pan M, Tong W, Law R, Wu EQ. URRNet: A Unified Relational Reasoning Network for Vehicle Re-Identification. IEEE Transactions on Vehicular Technology. 2023;72(9):11156–11168.

[30] Huang F, Lv X, Zhang L. Coarse-to-fine sparse self-attention for vehicle re-identification. Knowledge-Based Systems. 2023;270:110526.

[31] Xu Z, Wei L, Lang C, Feng S, Wang T, Bors AG, et al. SSR-Net: A Spatial Structural Relation Network for Vehicle Re-identification. ACM Transactions on Multimedia Computing, Communications and Applications. 2023;19(6):1–22.

Downloads

Published

11-02-2025

Issue

Section

Articles