@inproceedings{6724b76ffb424b0c8026aab869623fd7,
title = "AN EFFICIENT END-TO-END IMAGE COMPRESSION TRANSFORMER",
abstract = "Image and video compression received significant research attention and expanded their applications. Existing entropy estimation-based methods combine with hyperprior and local context, limiting their efficacy. This paper introduces an efficient end-to-end transformer-based image compression model, which generates a global receptive field to tackle the long-range correlation issues. A hyper encoder-decoder-based transformer block employs a multi-head spatial reduction self-attention (MHSRSA) layer to minimize the computational cost of the self-attention layer and enable rapid learning of multi-scale and high-resolution features. A Casual Global Anticipation Module (CGAM) is designed to construct highly informative adjacent contexts utilizing channel-wise linkages and identify global reference points in the latent space for end-to-end rate-distortion optimization (RDO). Experimental results demonstrate the effectiveness and competitive performance of the KODAK dataset.",
keywords = "Image compression, encoder-decoder, entropy model, transformer",
author = "Jeny, {Afsana Ahsan} and Junayed, {Masum Shah} and Islam, {Md Baharul}",
note = "Publisher Copyright: {\textcopyright} 2022 IEEE.; 29th IEEE International Conference on Image Processing, ICIP 2022 ; Conference date: 16-10-2022 Through 19-10-2022",
year = "2022",
doi = "10.1109/ICIP46576.2022.9897663",
language = "English",
series = "Proceedings - International Conference on Image Processing, ICIP",
publisher = "IEEE Computer Society",
pages = "1786--1790",
booktitle = "2022 IEEE International Conference on Image Processing, ICIP 2022 - Proceedings",
}