Abstract
In this paper, we present a Conv-Attention, a decoder-friendly attention mechanism, in an effort to advancing the practical application of the artificial intelligence-based image coding. More specifically, the proposed method is tailored for JPEG AI, which is the latest advanced neural-network based image coding standard. By identifying the obstacles by profiling the decoding complexity of JPEG AI, the attention module accounts for a significant proportion, which mainly attributes to the intricate network structure and involvement of less efficient operations. Conv-Attention model is composed with plain convolution and activation computations, equipping with sub-scaling and up-scaling design, such that the non-adjacent features can be well captured, leading to the reduction of decoding complexity and maintenance of the synthesis and attentive capability. Simulation results verify the effectiveness of the proposed method with JPEG AI reference software, wherein the decoding complexity is reduced by 80% with negligible coding performance loss. The proposed method was adopted in the 100th JPEG meeting.
Original language | English |
---|---|
Title of host publication | Proceedings : DCC 2024 : 2024 Data Compression Conference |
Editors | Ali BILGIN, James E. FOWLER, Joan SERRA-SAGRISTA, Yan YE, James A. STORER |
Publisher | IEEE |
Pages | 43-52 |
Number of pages | 10 |
ISBN (Electronic) | 9798350385878 |
DOIs | |
Publication status | Published - 2024 |
Externally published | Yes |
Event | 2024 Data Compression Conference, DCC 2024 - Snowbird, United States Duration: 19 Mar 2024 → 22 Mar 2024 |
Publication series
Name | Data Compression Conference Proceedings |
---|---|
ISSN (Print) | 1068-0314 |
Conference
Conference | 2024 Data Compression Conference, DCC 2024 |
---|---|
Country/Territory | United States |
City | Snowbird |
Period | 19/03/24 → 22/03/24 |
Bibliographical note
Publisher Copyright:© 2024 IEEE.