Modeling Inner- and Cross-Task Contrastive Relations for Continual Image Classification

Yuxuan LUO, Runmin CONG*, Xialei LIU, Horace Ho Shing IP, Sam KWONG*

*Corresponding author for this work

Research output: Journal PublicationsJournal Article (refereed)peer-review

2 Citations (Scopus)

Abstract

Existing continual image classification methods demonstrate that samples from all sequences of continual classification tasks contain common (task-invariant) features and class-specific (task-variant) features that can be decoupled for classification tasks. However, the existing feature decomposition strategies only focus on individual tasks while neglecting the essential cues that the relationship between different tasks can provide, thereby hindering the improvement of continual image classification results. To address this issue, we propose an Adversarial Contrastive Continual Learning (ACCL) method that decouples task-invariant and task-variant features by constructing all-round, multi-level contrasts on sample pairs within individual tasks or from different tasks. Specifically, three constraints on the distribution of task-invariant and task-variant features are included, i.e., task-invariant features across different tasks should remain consistent, task-variant features should exhibit differences, and task-invariant and task-variant features should differ from each other. At the same time, we also design an effective contrastive replay strategy to make full use of the replay samples to participate in the construction of sample pairs, further alleviating the forgetting problem, and modeling cross-task relationships. Through extensive experiments on continual image classification tasks on CIFAR100, MiniImageNet and TinyImageNet, we show the superiority of our proposed strategy, improving the accuracy and with better visualized outcomes.

Original languageEnglish
Pages (from-to)10842-10853
Number of pages12
JournalIEEE Transactions on Multimedia
Volume26
Early online date13 Jun 2024
DOIs
Publication statusPublished - 2024

Bibliographical note

Publisher Copyright:
© 2024 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.

Funding

This work was supported in part by National Science and Technology Major Project under Grant 2021ZD0112100, in part by Hong Kong GRF-RGC General Research Fund under Grant 11209819 (CityU 9042816) and Grant 11203820 (CityU 9042598), in part by the Taishan Scholar Project of Shandong Province under Grant tsqn202306079, and in part by Xiaomi Young Talents Program.

Keywords

  • Continual learning
  • contrastive learning
  • feature decomposition
  • image classification

Fingerprint

Dive into the research topics of 'Modeling Inner- and Cross-Task Contrastive Relations for Continual Image Classification'. Together they form a unique fingerprint.

Cite this