Coarse-to-fine sparse self-attention for vehicle re-identification

Fuxiang HUANG, Xuefeng LV, Lei ZHANG*

*Corresponding author for this work

Research output: Journal PublicationsJournal Article (refereed)peer-review

18 Citations (Scopus)

Abstract

Existing vehicle re-identification (Re-ID) methods usually combine global features and local features to meet the challenge of inter-class similarity and intra-class variance, but almost all existing methods rely on extra auxiliary networks to locate the vehicle parts for feature mining, which are inefficient, cost-ineffective and time-consuming. Self-attention mechanism, averaging model's attention weights on the similar vehicle parts, can guide the model to identify more discriminative vehicle parts for identification, thus extracting expensive local features becomes unnecessary. However, the computational cost of the original self-attention mechanism is unacceptable and the attention map is less discriminative when processing high-resolution feature maps (i.e. 384*384 or larger), which limits the performance gains of the challenging vehicle Re-ID. In this paper, we propose a lightweight coarse-to-fine sparse self-attention mechanism for vehicle Re-ID. Our method decomposes the self-attention process to a coarse stage and a fine stage. In coarse stage, the pixel-level feature map is transformed to patch-level feature map and the dependencies between similar vehicle parts are captured in a global context. In fine stage, the details of vehicle parts are captured in a local context. In addition, multi-head structure is introduced to capture diverse but robust attention information. Experiments on VeRi and VERI-Wild show that our method not only outperforms recent vehicle Re-ID methods relying on expensive auxiliary networks, but also outperforms the original self-attention mechanism in accuracy with much lower computation and memory cost.
Original languageEnglish
Article number110526
Number of pages10
JournalKnowledge-Based Systems
Volume270
Early online date5 Apr 2023
DOIs
Publication statusPublished - 21 Jun 2023
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2023 Elsevier B.V.

Funding

This work was partially supported by National Natural Science Fund of China (62271090), Chongqing Natural Science Fund (cstc2021jcyj-jqX0023), National Key R&D Program of China (202 1YFB3100800), CCF Hikvision Open Fund, China (CCF-HIKVISION OF 20210002), CAAI-Huawei MindSpore Open Fund, China, and Beijing Aca demy of Artificial Intelligence (BAAI), China.

Keywords

  • Feature mining
  • Neural network
  • Self-attention
  • Vehicle re-identification

Fingerprint

Dive into the research topics of 'Coarse-to-fine sparse self-attention for vehicle re-identification'. Together they form a unique fingerprint.

Cite this