Next Article in Journal
Non-Contact Smart Sensing of Physical Activities during Quarantine Period Using SDR Technology
Next Article in Special Issue
Low-Complexity Multiple Transform Selection Combining Multi-Type Tree Partition Algorithm for Versatile Video Coding
Previous Article in Journal
GRA-PIN: A Graphical and PIN-Based Hybrid Authentication Approach for Smart Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Post-Filtering Using Multi-Head Attention and Multiresolution Feature Fusion for Image and Intra-Video Quality Enhancement

Department of Electronics and Informatics, Vrije Universiteit Brussel, Pleinlaan 2, 1050 Brussels, Belgium
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(4), 1353; https://doi.org/10.3390/s22041353
Submission received: 18 January 2022 / Revised: 2 February 2022 / Accepted: 3 February 2022 / Published: 10 February 2022
(This article belongs to the Special Issue Advances in Image and Video Encoding Algorithm and H/W Design)

Abstract

:
The paper proposes a novel post-filtering method based on convolutional neural networks (CNNs) for quality enhancement of RGB/grayscale images and video sequences. The lossy images are encoded using common image codecs, such as JPEG and JPEG2000. The video sequences are encoded using previous and ongoing video coding standards, high-efficiency video coding (HEVC) and versatile video coding (VVC), respectively. A novel deep neural network architecture is proposed to estimate fine refinement details for full-, half-, and quarter-patch resolutions. The proposed architecture is built using a set of efficient processing blocks designed based on the following concepts: (i) the multi-head attention mechanism for refining the feature maps, (ii) the weight sharing concept for reducing the network complexity, and (iii) novel block designs of layer structures for multiresolution feature fusion. The proposed method provides substantial performance improvements compared with both common image codecs and video coding standards. Experimental results on high-resolution images and standard video sequences show that the proposed post-filtering method provides average BD-rate savings of 31.44 % over JPEG and 54.61 % over HEVC (x265) for RGB images, Y-BD-rate savings of 26.21 % over JPEG and 15.28 % over VVC (VTM) for grayscale images, and 15.47 % over HEVC and 14.66 % over VVC for video sequences.

1. Introduction

In image compression, the main objective is to develop efficient algorithms that minimize the amount of data required to represent the visual information. Nowadays, the consumption of image and video content is constantly increasing which calls for the design of novel compression methods with a highly increased coding performance.
Current compression methods employed on images include conventional image codecs such as JPEG [1] and JPEG2000 [2], denoted here simply J2K. The JPEG codec [1] was developed by the Joint Photographic Experts Group (JPEG), being the most common format for image compression. The J2K standard [2] improves the compression performance over its predecessor and offers several functionalities including resolution and quality scalability and region of interest coding. JPEG and J2K lossy compressed images often suffer from compression artifacts such as blocking artifacts, color bleeding, and ringing effects. JPEG is designed based on the discrete cosine transform (DCT), the block-based nature of this codec being associated with blocking effects at high compression ratios. The global nature of the discrete wavelet transform (DWT) employed in J2K overcomes the blocking effects typical for JPEG compression, but ringing effects and color bleeding are observed at high compression ratios.
A common solution to enhance the quality of lossy images is to employ different filtering techniques which reduce the effect of coding artifacts. In this work, we propose a novel filtering method based on convolutional neural networks (CNNs) designed to enhance the quality of high-resolution images and video sequences by post-processing the decoded images and video frames without applying other modification to the corresponding coding framework.
In our prior work, we propose to replace the built-in filtering module in HEVC [3] with efficient deep learning (DL)-based filtering tools and enhance the quality of HEVC-compressed video and light field (LF) images. In [4], we propose a frame-wise multiscale CNN for quality enhancement of HEVC decoded videos, built based on inception and residual learning blocks (IResLBs) [5], where additional information is extracted from the HEVC decoder to guide the network. In [6], we propose a multiscale CNN to enhance the quality of LF images using macro-pixel volumes. In this work, we introduce a novel DL-based filtering method for enhancing the quality of high-resolution lossy compressed RGB/grayscale images and video sequences. The proposed filtering method follows a simple post-processing approach where a neural network is employed to estimate refinement details which are subsequently added to the decoded images/frames in order to enhance their quality.
In summary, the novel contributions and findings of this paper are as follows: (1) an efficient CNN-based post-filtering method is proposed, where the architecture is designed based on the following concepts: (i) the multi-head attention mechanism for refining the feature maps, (ii) the weight sharing concept for reducing the network complexity, and (iii) novel block designs of layer structures for multiresolution feature fusion; (2) the current resolution patch is processed using feature maps extracted from both input patch and higher resolution patches; (3) the network is trained to estimate the refinement details at full-, half-, and quarter-patch resolutions; (4) the complex experimental evaluation over RGB/grayscale images and video sequences demonstrate that the proposed method offers substantial performance improvements when filtering images encoded by common lossy codecs and video sequences encoded by recent video coding standards, by reducing the coding artifacts specific to each codec.
The remainder of this paper is organized as follows. Section 2 overviews the state-of-the-art methods for image and video quality enhancement. Section 3 describes the proposed filtering method. Section 4 presents the experimental validation on RGB images, grayscale images, and video sequences. Finally, Section 5 draws the conclusions of this work.

2. Related Work

In recent years, machine learning (ML) techniques rapidly gained popularity and novel DL-based tools were proposed to replace, in conventional frameworks, specific coding modules such as transform [7], prediction [5,8,9,10], rate control [11], and interpolation [12], to name a few.
The design of novel DL-based tools to integrate into HEVC became a hot research topic, and filtering tools were proposed. In [13], an artifact reduction CNN (AR-CNN) architecture is proposed to reduce the artifacts in the JPEG compressed images using a sequence of four convolution layers. ARCNN was one of the first architectures that was able to achieve more than 1 dB improvement in PSNR over the JPEG compression on classical test images. In [14], the authors proposed a super-resolution CNN (SRCNN) to obtain high-resolution images from the corresponding low-resolution images in an end-to-end manner that outperforms the state-of-the-art in both reconstruction quality and computational speed. In [15], the authors proposed the variable-filter-size residue-learning CNN (VRCNN) architecture built based on ARCNN [13] and employed to replace the conventional HEVC built-in filters, de-blocking filter (DBF) [16] and sample adaptive offset (SAO) [17], when filtering the HEVC reconstructed intra-frames. In [18], a VRCNN architecture with batch normalization is proposed, where the new method is called VRCNN-BN and provides competitive results when enhancing the quality of HEVC decoded sequences. An iterative post-filtering method based on an RNN design was proposed in [19]. In [20], the authors employ a CNN-based autoencoder to estimate the radial blur and enhance the deblurred image. In [21], an image enhancement algorithm based on image-derived graph for weak illumination images is proposed. In [22], a DL-based hue-correction scheme is proposed based on the constant-hue plane in the RGB color space for image enhancement. In [23], a post-filtering DL-based method for quality enhancement of brain images is proposed, where the method operates on 3D volumetric data with a high dynamic range. In [24], a dual-stream recursive residual network is proposed, which consists of structure and texture streams for separately reducing the specific artifacts related to high-frequency or low-frequency components in JPEG compressed images. The method is called STRNN and it is the current state-of-the-art method for artifact reduction of JPEG compressed grayscale images. In [25], a novel multilevel progressive refinement generative adversarial network (MPRGAN) is employed to filter the intra-coded frames, where the generative network refines the reconstructed frame progressively in order to maximize the error of the adversarial network when distinguishing between the enhanced frame and the original frame. In [26], the authors proposed the frame enhancement CNN (FECNN) architecture which contains nine convolutional layers and introduces the residual learning scheme for filtering intra-frames to achieve a fast convergence speed in network training and also a higher filtering performance compared with VRCNN. In [27], a residual-based video restoration network (residual-VRN) is proposed based on residual learning blocks to enhance the quality of decoded HEVC intra-frames. In [28], a recursive residual CNN (RRCNN) architecture is proposed based on the recursive learning and residual learning schemes, which provides an important advantage over the state-of-the-art methods.
In one approach, DL-based filtering tools are introduced inside the main coding loop to filter the inter-predicted frames. In [29], based on the SRCNN [14] structure, the authors proposed a CNN architecture for in-loop filtering called IFCNN. IFCNN was designed to replace the HEVC built-in SAO [17] filter and to perform frame filtering after DBF [16] in the HEVC framework. In [30], the authors proposed a deep CNN architecture called DCAD employed to perform quality enhancement for all intra- and inter-frames after applying the traditional HEVC built-in DBF [16] and SAO filters [17]. In [31], the authors proposed a CNN architecture called RHCNN which contains residual highway units designed based on residual learning units to enhance the HEVC reconstructed frames. In [32], the authors proposed to perform content-aware CNN-based filtering of the HEVC decoded frames. In [33], a partition-aware architecture is proposed, where the CU partition size is used as additional information. In [34], a GAN-based perceptual training strategy is employed to post-process VVC and AV1 results. In [35], a novel architecture which exploits multilevel feature review residual dense blocks is proposed.
In another approach, DL-based filtering tools were proposed to filter the current inter-coded frame by making use of multiple previously reconstructed frames. In [36], the authors proposed a DL-based multi-frame in-loop filter (MIF) to enhance the current frame using several previously reconstructed frames. In [37], a multi-channel long-short-term dependency residual network (MLSDRN) was proposed to perform quality enhancement in addition to the DBF&SAO filtering. In [38], the authors proposed a filtering method based on the Decoder-side scalable convolutional neural network (DS-CNN), where a DS-CNN-I model is employed to enhance the intra-coded frames. In [39], the authors proposed a DL-based filtering method, QENet, designed to work outside the HEVC coding loop. However, the QENet still makes use of multiple frames considering the temporal correlations among them when filtering the inter-coded frames. In summary, in recent years, several deep-learning-based solutions were proposed to either replace the built-in HEVC in-loop filtering methods or to further refine the reconstructed frame. This is in contrast to the recent post-filtering deep-learning-based methods which prove that an improve coding solution is ordained by simply removing the built-in HEVC in-loop filtering methods.
The research area of neural-network-based loop filtering is intensively studied at the MPEG meetings. Many contributions [40,41,42,43] were recently proposed to enhance the VVC quality. In recent years, the attention mechanism [44] has become a powerful tool for improving the network performance. In [45], the attention module is used for low-light image enhancement. In [46], a new non-local attention module is proposed for low-light image enhancement using multiple exposure image sequences.
In this paper, we follow an approach where no modifications are applied to the codec, where the proposed CNN-based post-filtering method is employed to further filter the reconstructed image or decoded video frame. The quality enhancement experiments are carried out on three types of data: RGB images, grayscale images, and intra-predicted video frames. The proposed filtering method enhances the quality of lossy images encoded using traditional image codes, such as JPEG [1] and J2K [2], and of video sequences encoded using the latest video coding standards, HEVC [3] and VVC [47].

3. Proposed Method

Figure 1 depicts the proposed CNN-based filtering method which is designed to simply post-process a lossy image compressed using a traditional image or video codec. One can note that no modifications are applied to the encoder–decoder framework. The input image/frame is split into h × w blocks which are used as input patches for the deep neural network.
Figure 2 depicts the proposed network architecture called attention-based shared weights quality enhancement convolutional neural network (ASQE-CNN). ASQE-CNN is designed based on the following three important concepts: (i) the attention mechanism, where channel and spatial attention are employed to refine the feature maps, (ii) the weights sharing concept, where a single convolutional layer is employed twice inside a specific block of layer, (iii) the novel multiresolution feature fusion block design used to fuse current resolution feature maps with lower/higher resolution feature maps.
ASQE-CNN operates at three patch resolutions (full, half, and quarter) and introduces in its design new network branches used to extract feature maps from the input patch and to estimate the refinement details at each resolution. ASQE-CNN is designed to operate at h × w , h 2 × w 2 , and h 4 × w 4 patch resolution. In the first part of ASQE-CNN, these three resolution feature maps are obtained by processing the input patch, which lies in contrast to the general approach where the input patch is used only for processing the full-patch resolution. In the last part of ASQE-CNN, the final refinement details are extracted from the full-, half-, and quarter-patch resolutions, in contrast to the general approach where only the full-patch resolution is used.

3.1. Network Design

The ASQE-CNN architecture is built using eight types of blocks of layers that are depicted in Figure 3.
The convolutional block attention module (CBAM) was proposed in [44] and uses both channel and spatial attention. CBAM is employed here in the design of the multi-head attention block (MHA), where the input feature maps are channel split into K sets of feature maps and processed separately by a different CBAM block. Hence, the MHA block is designed based on the observation that the network architectures usually contain many channels to process the current patch. Here, we propose to divide the attention into a few dozen of channels by splitting the input channels into K sets of feature maps.
The convolution block (CB) contains a sequence of a 2D convolution (Conv2D) layer equipped with a 3 × 3 kernel, N filters, and stride s; a batch normalization (BN) layer [48]; and a rectified linear unit (ReLU) activation layer [49]. Similarly, the deconvolution block (DB) contains a sequence of a deconvolution (Deconv2D) layer equipped with a 3 × 3 kernel, N filters, and stride s; a BN layer; and an ReLU activation layer. For simplicity, the strides s = ( 2 , 2 ) and s = ( 4 , 4 ) are denoted as “/2” and “/4”, respectively. DB is used to perform a simple block processing and to increase the patch resolution.
The attention-based shared weights block (ASB) proposes a more efficient patch processing design where an MHA block is inserted between a CB and a Conv2D layer and trainable weights are shared between the two Conv2D layers. The skip connection branch is added to the current branch after Conv2D, while BN and ReLU layers are used to further process the feature maps. Moreover, the attention-based shared weights residual block (ASRB) is introduced as a residual learning [50] bottleneck block build using ARB blocks.
The multiresolution feature fusion block design is implemented using the low-resolution feature fusion (LFF) and high-resolution feature fusion (HFF) blocks. The blocks are designed using a similar strategy as ARB, where the MHA block is replaced by an add block which combines the current feature map with the processed low/high-feature maps.
The ASQE-CNN network contains three parts. The first part is called pre-processing and is used to extract from the input patch N feature maps at three patch resolutions, full, half ( s = / 2 ), and quarter ( s = / 4 ), by employing three CB blocks. The second part is called multiresolution feature fusion (MFF), where the three resolution patches undergo complex processing using multiresolution feature fusion based on ARBS, HFF, and LFF blocks. To reduce the inference time, ASQE-CNN processes the full-resolution blocks using MHA blocks with K = 1 and ARB blocks instead of ASRB blocks. The last part is called multiresolution refinement, where two DB blocks and three Conv2D layers, equipped with 3 × 3 kernels, are used to estimate the final refinement details corresponding to each resolution. One can note that ASQE-CNN adopts a new strategy for processing multiresolution patches, where the number of channels, N , remains constant throughout the network, in contrast to the general approach where the number of channels is doubled when the patch resolution is halved. Our experiments show that the increase in complexity for lower resolution patches does not provide a good improvement in network performance.
The ASQE-CNN model is obtained using N = 64 and K = 4 , and contains around 0.89 million (M) parameters. ASQE-CNN is designed to provide an improved performance using a reduced number of parameters. Note that without applying the weight sharing approach, 1.48 M parameters must be trained. Our experiments show that consecutive Conv2D layers may contain similar weights and a too-complex architecture is affected by the vanishing gradient problem; see Section 4.6.

3.2. Loss Function

The loss function consists of the summation of a mean squared error (MSE)-based term and an 2 regularization term, used to prevent model overfitting. Let us denote as Θ ASQE - CNN the set of all learned parameters of the ASQE-CNN model, X i the ith input patch in the training set, and Y i the corresponding original patch, both of size w × h × c , where c denotes the number of color channels, i.e., c = 3 for RGB images and c = 1 for grayscale images and video sequences. Let F ( · ) be the function which processes X i using Θ ASQE - CNN to compute the enhanced frame Y i ^ = F ( X i , Θ ASQE - CNN ) . The loss function is formulated as follows:
L ( Θ ASQE - CNN ) = 1 c w h i = 1 c w h y i y ^ i 2 2 + λ | | Θ ASQE - CNN | | 2 2 ,
where y i = v e c ( Y i ) and y i ^ = v e c ( Y i ^ ) are the vectorized versions of Y i and Y i ^ , respectively, and λ is the regularization term set here as λ = 0.01 . The Adam optimization algorithm [51] is employed.

4. Experimental Validation

The experimental setup used to validate the proposed DL-based post-filtering method is described in Section 4.1. The experimental results for quality enhancement of decoded RGB images, grayscale images, and video sequences are reported in Section 4.2, Section 4.3 and Section 4.4, respectively. The complexity of the proposed network architecture is discussed in Section 4.5. Finally, an ablation study for the proposed network design is presented in Section 4.6.

4.1. Experimental Setup

The paper aims to provide an efficient post-filtering method for enhancing the quality of high-resolution images and video sequences. The deep neural network models are trained using input patches extracted from the DIV2K dataset [52,53]. The training dataset in DIV2K is denoted DIV2K_train_HR and contains 800 high-resolution images. For each point reported in each rate-distortion curve, 522,240 input patches of resolution w × h = 64 × 64 are extracted from DIV2K_train_HR. The training data is stored using 136 files, where each file contains 3840 input patches. A 125 136 11 136 ( 92 % 8 % ) ratio is used for splitting the training data into training−validation data. Each neural network model is trained during 32 epochs. The learning rate at epoch i is set as η i + 1 = ( f d ) i n s η i , i = 1 , 2 , , 32 , where f d = 0.2 is the decay rate, n s = 5 is the decay step, and η 1 = 5 × 10 4 .
The experimental results for the proposed DL-based post-filtering method for quality enhancement of image and video sequences are reported on the following datasets: (a) DIV2K_valid_HR, the validation dataset from the DIV2K dataset, which contains 100 high-resolution images in the RGB color format (801.png, 802.png, , 900.png), see Table 1; (b) LIVE1 [54,55], which contains 29 images in the RGB color format, denoted here (1.bmp, 2.bmp, , 29.bmp), see Table 1 and Figure 4.; (c) 15 video sequences from the Common-Test Dataset [56], denoted here by HEVC-VTSEQ, which contains four different video classes (see Table 2) in the YUV420 color format. In this work, the experimental validation is performed using the following three different types of input data: (i) RGB images, extracted from DIV2K_valid_HR and LIVE1; (ii) grayscale images (Y800 color format), extracted from DIV2K_valid_HR and LIVE1, where the FFmpeg [57] framework is employed to color-transform the images from RGB to YUV420 format and the grayscale images contains the luminance channel (Y); (iii) video sequences, from HEVC-VTSEQ where the results are reported only for encoding the Y channel.
The proposed method is employed to enhance the quality of both the lossy images encoded using two traditional image codecs, JPEG [1] and J2K [2], and each frame in the video sequence encoded using the previous and ongoing video coding standards, HEVC [3] and VVC [47], respectively. No modifications are applied to the codec as ASQE-CNN is employed as a post-filtering method. HEVC and VVC do apply their built-in post-filtering method. In our prior works [4,6], we proved that such a built-in method can have a negative effect on the HEVC final results if a CNN is employed.
The MATLAB implementation of the JPEG codec [1] is used to generate lossy images for each of the following six quality factors points: q i { 90 , 85 , 70 , 40 , 20 , 10 } , where i = 1 : 6 . Note that the same results can be obtained using the Python Imaging Library (PIL). A neural network model is trained for each q i value and for each color format. Therefore, 6 + 6 = 12 neural network models are trained using DIV2K_train_HR to enhance the quality of RGB and grayscale images in DIV2K_valid_HR and LIVE1.
The MATLAB implementation of the J2K codec [2] is used to generate lossy images at the following six compression ratio points: c r i { 13 , 16 , 20 , 30 , 50 , 100 } . Similarly to JPEG, a neural network model is trained for each c r i value and for each color format. Hence, 12 models are trained using DIV2K_train_HR: six using RGB images and six using grayscale images. All 12 models are employed to enhance the images quality in the DIV2K_valid_HR [52] and LIVE1 [54] datasets.
The x265 library implementation of HEVC [3], available in the FFmpeg [57] framework, is employed to encode single images and video sequences using all-intra profile. For the RGB and grayscale images case, the lossy images are generated for each of the following six constant rate factors (CRF), c r f i { 16 , 20 , 24 , 28 , 32 , 36 } , i.e., the -c:v libx265 -preset veryslow -crf crf_i parameters are used to encode the images. A neural network model is trained for each c r f i value and for each color format. For the video sequences case, the lossy images are generated for each of the following four standard quantization parameter (QP) values, q p j { 22 , 27 , 32 , 37 } , where j = 1 : 4 , i.e., the -x265-params keyint=1:qp=qp_i parameters are used to encode the video sequences. A neural network model is trained for each q p j value. A total of 16 network models are trained: six models are used to enhance the grayscale images and six models to enhance the RGB images in DIV2K_valid_HR and LIVE1, and four models are used to enhance the Y component in HEVC-VTSEQ.
In this work, the VVC Test Model (VTM) [58], which is the reference software implementation of VVC [47], is employed to encode single images and video sequences in all-intra profile. Similarly to HEVC, the lossy images are generated for each of the four standard QP values. Please note that the VVC runtime is extremely large, e.g., a single high-resolution image is encoded in around 15 minutes (min), and 15 ( min ) × 4 ( QP values ) × 800 ( images ) = 48,000 ( min ) 33 ( days ) are needed to generate training data. Four models are trained using input patches extracted from the RGB images in DIV2K_valid_HR [52] and used to report results for quality enhancement of RGB images. Four models are trained using patches extracted from the luminance channel after color transforming the images in DIV2K_valid_HR [52] and used to report results for quality enhancement of both grayscale images and video sequences. Hence, eight network models are trained: four models are used to enhance grayscale images (DIV2K_valid_HR and LIVE1) and video sequences (HEVC-VTSEQ), and four models to enhance RGB images (DIV2K_valid_HR and LIVE1).
The quality enhancement results of the proposed post-filtering method are obtained using a total of 12 ( JPEG ) + 12 ( J 2 K ) + 16 ( HEVC ) + 8 ( VVC ) = 48 trained network models. The results are obtained using the four tested codecs, JPEG [1], J2K [2], HEVC [3], VVC [47], and by employing the proposed method to enhance the quality of the corresponding decode image and video sequences, denoted here as JPEG+ASQE-CNN, J2K+ASQE-CNN, HEVC+ASQE-CNN, and VVC+ASQE-CNN, respectively.
In our previous work [4], a frame-wise CNN-based filtering method is proposed for video quality enhancement. In contrast to this approach, in this work, we propose a block-based approach for quality enhancement. Due to its block-based nature, the enhanced image may still be affected by blocking artifacts. The experiments show that we can further reduce the effect of coding artifacts by applying, for a second time, the proposed method on the same lossy compressed image whereby the top 16 lines and left 16 columns are cropped, i.e., the input patches are extracted a second time from the cropped input image, enhanced, and then concatenated to obtain the second enhanced image. The two enhanced images are then fused to obtain the enhancement results using alpha blending in the RGB image case and mean in the grayscale image case. Note that fusion is performed only over the size of the second enhanced image, while the top 16 lines and left 16 columns area is copied from the first enhanced image.
The image distortion is reported using both peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) [59], where the PSNR between an original image Y ^ and a reconstructed image Y ^ is computed as follows:
P S N R [ Y , Y ^ ] = 20 log 10 255 M S E [ I , I ^ ] ,
where M S E [ Y , Y ^ ] is the corresponding mean squared error. The SSIM index [59] for Y and Y ^ is computed as follows:
S S I M ( Y , Y ^ ) = 2 μ Y μ Y ^ + c 1 2 σ Y Y ^ + c 2 μ Y 2 + μ Y ^ 2 + c 1 σ Y 2 + σ Y ^ 2 + c 2 ,
where μ Y and μ Y ^ are the mean of Y and Y ^ respectively; σ Y 2 and σ Y ^ 2 are the variance of Y and Y ^ respectively; and c 1 = ( k 1 L ) 2 , c 2 = ( k 2 L ) 2 are used to stabilize the division with weak denominator, with default values k 1 = 0.01 ,   k 2 = 0.03 , and L = 2 b 1 for b-bit precision.
The quality enhancement results are compared using the two Bjøntegaard metrics [60], Bjøntegaard delta-rate (BD-Rate) and Bjøntegaard delta-PSNR (BD-PSNR). BD-PSNR was introduced to evaluate the quality gains made by one video codec versus another video codec at the equivalent bitrate. More exactly, when a specific amount of bits are spent to encode the video using each codec, then BD-PSNR measures by how much one codec provides a better quality than the other. Similarly, BD-rate measures the bitrate savings at the equivalent quality. In this paper, we use the Python implementation publicly available here [61]. The bitrate is measured as bits per channel (bpc), which is defined as the ratio between the compresses image file size and the product between the number of image pixels and the number of channels.
We note that in the considered experiments, we evaluate the proposed method on several data types and coding standards. The proposed method functions as a post-processing method, decoupled from the underlying coding methodology. In this sense, in contrast to customized in-loop filtering methods which require modifications of the standard implementation, the proposed method remains standard-compliant and can be used as post-filtering tool on image and video data already encoded with existing coding standards. To address the random access (RA) profile of existing video coding standards, it is expected that the proposed method will need to be performed in-loop, to maximize coding performance. Again, adapting it to in-loop filtering makes it codec-specific, and departs from the basic idea of devising a CNN-based post-processing method followed in this work. Investigating how the proposed post-filtering method can be adapted to the RA configuration is left as a topic of further investigation.

4.2. Quality Enhancement of RGB images

4.2.1. Rate-Distortion Results

Figure 5 presents the rate-distortion results obtained for three randomly selected images from DIV2K_valid_HR (801.png, 821.png, 849.png) and the average results over DIV2K_valid_HR [52]. Figure 6 presents the rate-distortion results obtained for three randomly selected images from LIVE1 (1.bmp, 11.bmp, 20.bmp) and the average results over LIVE1 [54]. The quality enhancement results are measured using PSNR (left column) and SSIM (right column). One can note that the proposed method provides substantial performance improvements for both the image codecs and video coding standards. Impressive results are obtained compared with HEVC. Large BD-rate reduction is obtained compared with the newest video coding standard, VVC.

4.2.2. Bjøntegaard Metrics

Figure 7 presents quality enhancement results computed based on Bjøntegaard metrics [60] for every image in DIV2K_valid_HR [52] and LIVE1 [54]. One can note that for JPEG+ASQE-CNN, the BD-PSNR improvement varies between 0.57 dB and 5.65 dB and the BD-rate savings vary between 12.71 % and 60.34 % . J2K+ASQE-CNN achieves BD-PSNR improvement of up to 4.11 dB and BD-rate savings of up to 45.9 % . HEVC+ASQE-CNN achieves BD-PSNR improvement of up to 5.47 dB and BD-rate savings of up to 65.26 % . VVC+ASQE-CNN achieves BD-PSNR improvement of up to 1.8 dB and BD-rate savings of up to 35 % .
Table 3 presents the average BD-rate savings and BD-PSNR improvement over the two test sets, DIV2K_valid_HR [52] and LIVE1 [54]. In the DIV2K_valid_HR case, ASQE-CNN provides average BD-rate savings of 31.44 % and an average BD-PSNR improvement of 2.006 dB compared to the traditional JPEG codec, and average BD-rate savings of 16.02 % and an average BD-PSNR improvement of 0.905 dB compared to the traditional J2K codec. However, outstanding quality enhancement results are obtained compared with the HEVC standard, average BD-rate savings of 54.61 % and average BD-PSNR improvement of 3.448 dB, and compared with the VVC standard, average BD-rate savings of 19.39 % and average BD-PSNR improvement of 1.201 dB. In the LIVE1 case, ASQE-CNN provides (i) average BD-rate savings of 31.44 % and an average BD-PSNR improvement of 2.006 dB for JPEG; (ii) average BD-rate savings of 16.02 % and an average BD-PSNR improvement of 0.905 dB for J2K; (iii) average BD-rate savings of 52.78 % and an average BD-PSNR improvement of 3.7380 dB for HEVC; and (iv) average BD-rate savings of 18.79 % and an average BD-PSNR improvement of 1.3056 dB for VVC. One can note that the ASQE-CNN provides impressive results over both test sets.

4.2.3. Visual Results

Figure 8 presents a visual comparison between the following methods: JPEG operating at the lowest quality parameter, q 6 = 10 , J2K operating at the highest compression ratio, c r 6 = 13 , HEVC operating at the highest constant rate factor, c r f 6 = 36 , VVC operating at q p 4 = 37 , and the proposed CNN-based post-filtering method, JPEG+ASQE-CNN, J2K+ASQE-CNN, HEVC+ASQE-CNN, and VVC+ASQE-CNN, respectively. The figure shows that the quality of the highly-distorted JPEG image, having a PSNR of only 27.87 dB and SSIM of 0.7825 , was substantially enhanced by JPEG+ASQE-CNN to 29.44 dB in PSNR and 0.8374 in SSIM. The coding artifacts, such as color bleeding and blocking artifacts, were reduced and most of the high details were estimated by the ASQE-CNN models, e.g., see the penguin’s beak and the rocky background. One can note that J2K offers a better compression performance than JPEG; however, in the zoom-in area the coding artifacts are still visible. Figure 8 depicts the J2K compressed image, having a PSNR of 29.77 dB and an SSIM of 0.8548 , while the proposed J2K+ASQE-CNN method is able to enhance the image quality to 30.40 dB PSNR and 0.8676 SSIM, to reduce the coding artifacts and provide an improved visual quality. Moreover, in the case of HEVC compression, the image quality was improved from 25.30 dB in PSNR and 0.6571 in SSIM to 28.55 dB in PSNR and 0.8101 in SSIM, while in the case of VVC compression the image quality was improved from 32.32 dB in PSNR and 0.9046 in SSIM to 33.39 dB in PSNR and 0.9326 in SSIM.

4.3. Quality Enhancement of Grayscale Images

4.3.1. Rate-Distortion Results

Figure 9 presents the rate-distortion results obtained for two images, 801.png from DIV2K_valid_HR and 11.bmp from LIVE1, and over the two datasets, DIV2K_valid_HR [52] and LIVE1 [54]. AEQE-CNN provides an enhanced quality for grayscale images compared to both traditional image codecs, JPEG [1] and J2K [2], and both video coding standards, HEVC [3] and VVC [47]. The image quality is improved over both tested datasets, DIV2K_valid_HR and LIVE1, having different image resolutions, which proves that the models do not overfit on the DIV2K_train_HR dataset. Moreover, impressive results are obtained compared with the most recent video coding standard, VVC [47].
Table 4 presents the comparison with state-of-the-art methods designed to enhance only lossy JPEG images: ARCNN [13], RED30 [62], DRRN [63], ARN [64], and STRNN [24]. Note that the proposed method provides the best results for all three quality factors, q 4 = 40 , q 5 = 20 , and q 6 = 10 .

4.3.2. Bjøntegaard Metrics

Figure 10 presents quality enhancement results computed based on Bjøntegaard metrics [60] for every image in DIV2K_valid_HR [52] and LIVE1 [54]. Note that both the Y-BD-PSNR improvement and Y-BD-rate savings have a high variation over the two datasets. Table 5 presents the average BD-rate savings and Y-BD-PSNR improvement over DIV2K_valid_HR and LIVE1. In the DIV2K_valid_HR case, ASQE-CNN provides average BD-rate savings of 26.21 % and an average Y-BD-PSNR improvement of 1.826 dB compared to the traditional JPEG codec, and average Y-BD-rate savings of 19.31 % and an average Y-BD-PSNR improvement of 0.896 dB compared to the traditional J2K codec. Outstanding quality enhancement results are obtained compared with HEVC, average Y-BD-rate savings of 21.34 % and average Y-BD-PSNR improvement of 1.267 dB, and compared with VVC, average BD-rate savings of 16.18 % and average Y-BD-PSNR improvement of 1.039 dB. Similar results are obtained over the LIVE1 dataset. One can note that the ASQE-CNN provides impressive results over both test sets of grayscale images.

4.3.3. Visual Results

Figure 11 presents a visual comparison between JPEG ( q 6 = 10 ), J2K ( c r 6 = 13 ), HEVC ( c r f 6 = 36 ), VVC ( q p 4 = 37 ), and the proposed CNN-based post-filtering method, JPEG+ASQE-CNN, J2K+ASQE-CNN, HEVC+ASQE-CNN, and VVC+ASQE-CNN, respectively. The quality of the traditional methods is substantially enhanced by ASQE-CNN. The coding artifacts are reduced and most of the high details are estimated by the ASQE-CNN models, e.g., see the penguin’s beak and the rocky background. The bottom row presents the pseudo-color image compression between the above two image quality results, where green marks the pixel positions where the proposed ASQE-CNN method provides an improved performance, and red marks the pixel positions where it fails. The proposed method estimates fine details for around 50 % of the pixels and provides the same result for around 20 % of the pixels.

4.4. Quality Enhancement of Video Sequences

4.4.1. Bjøntegaard Metrics

Table 6 presents the video quality enhancement results computed as Y-BD-rate savings for HEVC-vTEQ [56]. The results are computed for all available frames in each video sequence by employing ASQE-CNN to post-process the HEVC [3]-decoded videos and are compared with the results of the six most recent state-of-the-art methods, VR-CNN [15], FE-CNN [26], MLSDRN [37], RRCNN [28], VRCNN-BN [18], and FQE-CNN [4]. The results obtained for the first 30 frames in each video sequence by employing ASQE-CNN to post-process the VVC [47]-decoded videos are also reported in Table 6. ASQE-CNN achieves an outstanding performance compared with the state-of-the-art methods for all video sequences, except for B5, and provides 15.47 % Y-BD-rate savings compared with HEVC and 14.66 % Y-BD-rate savings compared with VVC.
Table 7 presents quality enhancement results computed as Y-BD-PSNR improvement. ASQE-CNN provides an improvement of around 1.13 dB compared with HEVC [3] and 0.88 dB compared with VVC [47].

4.4.2. Visual Results

Figure 12 presents the pseudo-colored image comparison between HEVC and VVC at q p 4 = 37 and HEVC+ASQE-CNN, and VVC+ASQE-CNN, respectively. ASQE-CNN estimates fine details for ≈45% of the pixels and provides the same result for ≈25% of the pixels.

4.5. Complexity

ASQE-CNN is designed to have a medium complexity and small inference time. It is implemented in the Python programming language using the Keras [65] open-source deep learning library, and is run on machines equipped with NVIDIA TITAN Xp GPU. Each model is trained in around 34 h (1.42 days). The average inference time for a batch of 32 input patches is 67 ms. For example, for filtering a lossy compressed image with a 1356 × 2040 resolution, 704 input patches are extracted and an inference time of around 704 32 · 67 = 1474 ms is required to enhance the lossy compressed image.

4.6. Ablation Study

The ASQE-CNN architecture is designed based on the following three main concepts: (a) the attention mechanism; (b) the weight sharing approach; and (c) the proposed MFF modules. Here, we study how important is to integrate each one of these concepts in ASQE-CNN. The first version is designed without the use of the attention mechanism, where all the MHA blocks are removed. This version is called noAttention. The second version is designed without the use of the weight sharing approach. Therefore, the ASB and LFF and HFF blocks are designed using two convolutional layers with sparely trained weights instead of a single convolutional layer employed twice inside the block. This version is called noWeighSharing. The third version is designed without the use of the MFF modules, i.e., by following the classical multiresolution patch processing design (U-Net). In this case, the following branches in the proposed ASQE-CNN architecture are removed: (1) the half- and quarter-resolution branches which extract feature maps from the input patch in the first part of the architecture; and (2) the half- and quarter-resolution branches which provide the extra refinement in the last part of the architecture. This version is called noMFF (or U-Net).
Figure 13 presents the average performance results over DIV2K_valid_HR [52], while Table 8 presents the average Bjøntegaard metrics and the runtime using batch size (bs) of 100 input patches. All three versions are affected by a performance drop compared with the proposed architecture. The following conclusions can be drawn: (1) the attention mechanism has the highest influence in the network design of around 12 % BD-rate savings compared with the proposed architecture; (2) although noWeighSharing is more complex (contains more parameters), it provides a small improvement of around 11.45 % BD-rate savings compared with the proposed architecture; (3) the classical multiresolution patch processing design (U-Net) was improved using the proposed MFF modules, which provides around 11.54 % BD-rate savings compared with the proposed architecture.

5. Conclusions

The paper introduces a novel CNN-based post-filtering method for enhancing the quality of lossy images encoded using traditional codecs and video sequences encoded using recent video coding standards. A novel deep neural network architecture is proposed to estimate the refinement details. ASQE-CNN incorporates in its design three concepts: the attention mechanism implemented using a multi-head attention block, the weights sharing concept, where a single convolutional layer is employed twice in a specific block, and the proposed MFF blocks, LFF and HFF, designed to fuse current resolution feature maps with lower/higher-resolution feature maps. It operates at three resolutions and uses new network branches to extract feature maps from the input patch and to estimate refinement details at each resolution. Experimental results demonstrate substantial average BD-rate and BD-PSNR improvements over traditional image and video codecs. The paper demonstrates the potential of CNN-based post-filtering methods for widely-used codecs.

Author Contributions

Conceptualization, I.S.; methodology, I.S.; software, I.S.; validation, I.S.; investigation, I.S.; resources, A.M.; writing—original draft preparation, I.S. and A.M.; writing—review and editing, I.S. and A.M.; visualization, I.S.; project administration, I.S. and A.M.; funding acquisition, I.S. and A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by Innoviris within the research project DRIvINg, and by the Research Foundation—Flanders (FWO) within the research project G094122N.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; nor in the decision to publish the results. All authors read and approved the final manuscript.

Abbreviations

The following abbreviations are used in this manuscript:
JPEGJoint Photographic Experts Group
HEVCHigh-efficiency video coding
VVCVersatile video coding
CNNConvolutional neural network
MPEGMoving Picture Experts Group
ASQE-CNNAttention-based shared weights quality enhancement CNN
CBConvolutional block
DBDeconvolutional block
CBAMCovolutional block attention module
ASBAttention-based shared weights block
ASRBAttention-based shared weights residual block
MHAMulti-head attention block
LFFLow-resolution feature fusion
HFFHigh-resolution feature fusion

References

  1. Wallace, G.K. The JPEG Still Picture Compression Standard. Commun. ACM 1991, 34, 30–44. [Google Scholar] [CrossRef]
  2. Taubman, D.S.; Marcellin, M.W. JPEG2000: Standard for interactive imaging. Proc. IEEE 2002, 90, 1336–1357. [Google Scholar] [CrossRef]
  3. Sullivan, G.J.; Ohm, J.; Han, W.; Wiegand, T. Overview of the High Efficiency Video Coding (HEVC) Standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  4. Huang, H.; Schiopu, I.; Munteanu, A. Frame-wise CNN-based Filtering for Intra-Frame Quality Enhancement of HEVC Videos. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 2100–2113. [Google Scholar] [CrossRef]
  5. Schiopu, I.; Munteanu, A. Deep-Learning-Based Lossless Image Coding. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1829–1842. [Google Scholar] [CrossRef]
  6. Huang, H.; Schiopu, I.; Munteanu, A. Macro-pixel-wise CNN-based filtering for quality enhancement of light field images. Electron. Lett. 2020, 56, 1413–1416. [Google Scholar] [CrossRef]
  7. Lan, C.; Xu, J.; Zeng, W.; Shi, G.; Wu, F. Variable Block-Sized Signal-Dependent Transform for Video Coding. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 1920–1933. [Google Scholar] [CrossRef]
  8. Li, J.; Li, B.; Xu, J.; Xiong, R.; Gao, W. Fully Connected Network-Based Intra Prediction for Image Coding. IEEE Trans. Image Process 2018, 27, 3236–3247. [Google Scholar] [CrossRef]
  9. Pfaff, J.; Helle, P.; Maniry, D.; Kaltenstadler, S.; Samek, W.; Schwarz, H.; Marpe, D.; Wieg, T. Neural network based intra prediction for video coding. In Applications of Digital Image Processing XLI; Tescher, A.G., Ed.; International Society Optics and Photonics, SPIE: Bellingham, WA, USA; pp. 359–365.
  10. Schiopu, I.; Huang, H.; Munteanu, A. CNN-Based Intra-Prediction for Lossless HEVC. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1816–1828. [Google Scholar] [CrossRef]
  11. Li, Y.; Li, B.; Liu, D.; Chen, Z. A convolutional neural network-based approach to rate control in HEVC intra coding. In Proceedings of the IEEE Visual Communications and Image Process, St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  12. Wu, C.; Singhal, N.; Krähenbühl, P. Video Compression through Image Interpolation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  13. Dong, C.; Deng, Y.; Loy, C.C.; Tang, X. Compression Artifacts Reduction by a Deep Convolutional Network. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 576–584. [Google Scholar] [CrossRef] [Green Version]
  14. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  15. Dai, Y.; Liu, D.; Wu, F. A Convolutional Neural Network Approach for Post-Processing in HEVC Intra Coding. In Proceedings of the International Conference on Multimedia Model, Reykjavik, Iceland, 4–6 January 2017; pp. 28–39. [Google Scholar]
  16. Norkin, A.; Bjontegaard, G.; Fuldseth, A.; Narroschke, M.; Ikeda, M.; Andersson, K.; Zhou, M.; Van der Auwera, G. HEVC Deblocking Filter. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1746–1754. [Google Scholar] [CrossRef]
  17. Fu, C.M.; Alshina, E.; Alshin, A.; Huang, Y.W.; Chen, C.Y.; Tsai, C.Y.; Hsu, C.W.; Lei, S.M.; Park, J.H.; Han, W.J. Sample Adaptive Offset in the HEVC Standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1755–1764. [Google Scholar] [CrossRef]
  18. Zhao, H.; He, M.; Teng, G.; Shang, X.; Wang, G.; Feng, Y. A CNN-Based Post-Processing Algorithm for Video Coding Efficiency Improvement. IEEE Access 2020, 8, 920–929. [Google Scholar] [CrossRef]
  19. Ororbia, A.G.; Mali, A.; Wu, J.; O’Connell, S.; Dreese, W.; Miller, D.; Giles, C.L. Learned Neural Iterative Decoding for Lossy Image Compression Systems. In Proceedings of the Data Compression Conference, Snowbird, UT, USA, 26–29 March 2019; pp. 3–12. [Google Scholar] [CrossRef] [Green Version]
  20. Hurakadli, V.; Kulkarni, S.; Patil, U.; Tabib, R.; Mudengudi, U. Deep Learning based Radial Blur Estimation and Image Enhancement. In Proceedings of the IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT), Bangalore, India, 26–27 July 2019; pp. 1–5. [Google Scholar] [CrossRef]
  21. Zhuang, L.; Guan, Y. Image Enhancement by Deep Learning Network Based on derived image and Retinex. In Proceedings of the IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 11–13 October 2019; pp. 1670–1673. [Google Scholar] [CrossRef]
  22. Kinoshita, Y.; Kiya, H. Hue-Correction Scheme Based on Constant-Hue Plane for Deep-Learning-Based Color-Image Enhancement. IEEE Access 2020, 8, 9540–9550. [Google Scholar] [CrossRef]
  23. Gao, S.; Xiong, Z. Deep Enhancement for 3D HDR Brain Image Compression. In Proceedings of the IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; pp. 714–718. [Google Scholar] [CrossRef]
  24. Jin, Z.; Iqbal, M.Z.; Zou, W.; Li, X.; Steinbach, E. Dual-Stream Multi-Path Recursive Residual Network for JPEG Image Compression Artifacts Reduction. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 467–479. [Google Scholar] [CrossRef]
  25. Jin, Z.; An, P.; Yang, C.; Shen, L. Quality Enhancement for Intra Frame Coding Via CNNs: An Adversarial Approach. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Calgary, AB, Canada, 15–20 April 2018; pp. 1368–1372. [Google Scholar]
  26. Li, F.; Tan, W.; Yan, B. Deep Residual Network for Enhancing Quality of the Decoded Intra Frames of HEVC. In Proceedings of the IEEE International Conference on Image Processing, Athens, Greece, 7–10 October 2018; pp. 3918–3922. [Google Scholar]
  27. Ma, L.; Tian, Y.; Huang, T. Residual-Based Video Restoration for HEVC Intra Coding. In Proceedings of the IEEE International Conference on Multimedia Big Data (BigMM), Xi’an, China, 13–16 September 2018; pp. 1–7. [Google Scholar]
  28. Zhang, S.; Fan, Z.; Ling, N.; Jiang, M. Recursive Residual Convolutional Neural Network-Based In-Loop Filtering for Intra Frames. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 1888–1900. [Google Scholar] [CrossRef]
  29. Park, W.; Kim, M. CNN-based in-loop filtering for coding efficiency improvement. In Proceedings of the IEEE Image, Video, and Multidimensional Signal Processing Workshop (IVMSP), Bordeaux, France, 11–12 July 2016; pp. 1–5. [Google Scholar]
  30. Wang, T.; Chen, M.; Chao, H. A Novel Deep Learning-Based Method of Improving Coding Efficiency from the Decoder-End for HEVC. In Proceedings of the Data Compression Conference, Snowbird, UT, USA, 4–7 April 2017; pp. 410–419. [Google Scholar]
  31. Zhang, Y.; Shen, T.; Ji, X.; Zhang, Y.; Xiong, R.; Dai, Q. Residual Highway Convolutional Neural Networks for in-loop Filtering in HEVC. IEEE Trans. Image Process. 2018, 27, 3827–3841. [Google Scholar] [CrossRef]
  32. Jia, C.; Wang, S.; Zhang, X.; Wang, S.; Liu, J.; Pu, S.; Ma, S. Content-Aware Convolutional Neural Network for In-Loop Filtering in High Efficiency Video Coding. IEEE Trans. Image Process. 2019, 28, 3343–3356. [Google Scholar] [CrossRef]
  33. Lin, W.; He, X.; Han, X.; Liu, D.; See, J.; Zou, J.; Xiong, H.; Wu, F. Partition-Aware Adaptive Switching Neural Networks for Post-Processing in HEVC. IEEE Trans. Multimed. 2020, 22, 2749–2763. [Google Scholar] [CrossRef] [Green Version]
  34. Zhang, F.; Ma, D.; Feng, C.; Bull, D.R. Video Compression with CNN-based Post Processing. IEEE Multimed. 2021, 28, 74–83. [Google Scholar] [CrossRef]
  35. Ma, D.; Zhang, F.; Bull, D.R. MFRNet: A New CNN Architecture for Post-Processing and In-loop Filtering. IEEE J. Sel. Top. Signal Process. 2021, 15, 378–387. [Google Scholar] [CrossRef]
  36. Li, T.; Xu, M.; Zhu, C.; Yang, R.; Wang, Z.; Guan, Z. A Deep Learning Approach for Multi-Frame In-Loop Filter of HEVC. IEEE Trans. Image Process. 2019, 28, 5663–5678. [Google Scholar] [CrossRef] [Green Version]
  37. Meng, X.; Chen, C.; Zhu, S.; Zeng, B. A New HEVC In-Loop Filter Based on Multi-channel Long-Short-Term Dependency Residual Networks. In Proceedings of the Data Compression Conference, Snowbird, UT, USA, 27–30 March 2018; pp. 187–196. [Google Scholar] [CrossRef]
  38. Yang, R.; Xu, M.; Wang, Z. Decoder-side HEVC quality enhancement with scalable convolutional neural network. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 817–822. [Google Scholar]
  39. Lu, M.; Cheng, M.; Xu, Y.; Pu, S.; Shen, Q.; Ma, Z. Learned Quality Enhancement via Multi-Frame Priors for HEVC Compliant Low-Delay Applications. In Proceedings of the IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; pp. 934–938. [Google Scholar]
  40. Yin, H.; Yang, R.; Fang, X.; Ma, S. CE10-1.7: Adaptive convolutional neural network loop filter. In Proceedings of the Joint Video Experts Team (JVET) Meeting, Gothenburg, Sweden, 3–12 July 2019; No. JVET-O0063, ITU-T, ISO/IEC: Gothenburg, Sweden, 2019. [Google Scholar]
  41. Wan, S.; Wang, M.; Ma, Y.; Huo, J.; Gong, H.; Zou, C. CE10: Integrated in-loop filter based on CNN. In Proceedings of the Joint Video Experts Team (JVET) Meeting, Gothenburg, Sweden, 3–12 July 2019; No. JVET-O0079, ITU-T, ISO/IEC: Gothenburg, Sweden, 2019. [Google Scholar]
  42. Kidani, Y.; Kawamura, K.; Unno, K.; Naito, S. CE10- 1.10/CE10-1.11: Evaluation results of CNN-based filtering with on-line learning model. In Proceedings of the Joint Video Experts Team (JVET) Meeting, Gothenburg, Sweden, 3–12 July 2019; No. JVET-O0131, ITU-T, ISO/IEC: Gothenburg, Sweden, 2019. [Google Scholar]
  43. Kidani, Y.; Kawamura, K.; Unno, K.; Naito, S. CE10- 2.10/CE10-2.11: Evaluation results of CNN-based filtering with off-line learning model. In Proceedings of the Joint Video Experts Team (JVET) Meeting, Gothenburg, Sweden, 3–12 July 2019; No. JVET-O0132, ITU-T, ISO/IEC: Gothenburg, Sweden, 2019. [Google Scholar]
  44. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  45. Huang, W.; Zhu, Y.; Huang, R. Low Light Image Enhancement Network With Attention Mechanism and Retinex Model. IEEE Access 2020, 8, 74306–74314. [Google Scholar] [CrossRef]
  46. Chen, Q.; Fan, J.; Chen, W. An improved image enhancement framework based on multiple attention mechanism. Displays 2021, 70, 102091. [Google Scholar] [CrossRef]
  47. Bross, B.; Chen, J.; Ohm, J.R.; Sullivan, G.J.; Wang, Y.K. Developments in International Video Coding Standardization After AVC, With an Overview of Versatile Video Coding (VVC). Proc. IEEE 2021, 109, 1463–1493. [Google Scholar] [CrossRef]
  48. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the International Conference on Machine Learning (ICML), Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  49. Nair, V.; Hinton, G. Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the International Conference on Machine Learning (ICML), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  51. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  52. Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  53. Ignatov, A.; Timofte, R.; Van Vu, T.; Minh Luu, T.; XPham, T.; Van Nguyen, C.; Kim, Y.; Choi, J.S.; Kim, M.; Huang, J.; et al. PIRM challenge on perceptual image enhancement on smartphones: Report. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2019; 2019. [Google Scholar]
  54. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process 2006, 15, 3440–3451. [Google Scholar] [CrossRef]
  55. Sheikh, H.; Wang, Z.; Cormack, L.; Bovik, A. LIVE Image Quality Assessment Database. Available online: https://live.ece.utexas.edu/research/quality/subjective.htm (accessed on 1 July 2021).
  56. Bossen, F. Common HM test conditions and software reference Configurations. In Proceedings of the Joint Collaborative Team on Video Coding (JCT-VC) Meeting, Report No. JCTVC-G1100, San Jose, CA, USA, 11–20 July 2012. [Google Scholar]
  57. FFmpeg. Available online: http://ffmpeg.org/documentation.html (accessed on 1 February 2021).
  58. Fraunhofer Institute, HHI. VVC Test Model (VTM) Source Code. Available online: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftware_VTM (accessed on 1 July 2021).
  59. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  60. Bjøntegaard, G. Calculation of average PSNR differences between RD-Curves. In Proceedings of the joint International Telecommunication Union Telecommunication Standardization Sector (ITU-T) and Video Coding Experts Group (VCEG) Meeting, Pattaya, Thailand, 4–7 December 2001. [Google Scholar]
  61. Bjontegaard Metrics Computation. Available online: https://github.com/Anserw/Bjontegaard_metric (accessed on 1 February 2021).
  62. Mao, X.J.; Shen, C.; Yang, Y.B. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2802–2810. [Google Scholar]
  63. Tai, Y.; Yang, J.; Liu, X. Image Super-Resolution via Deep Recursive Residual Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2790–2798. [Google Scholar] [CrossRef]
  64. Zhang, Y.; Sun, L.; Yan, C.; Ji, X.; Dai, Q. Adaptive Residual Networks for High-Quality Image Restoration. IEEE Trans. Image Process 2018, 27, 3150–3163. [Google Scholar] [CrossRef]
  65. Chollet, F. Keras. Available online: https://keras.io (accessed on 1 July 2017).
Figure 1. Flowchart of the proposed post-filtering approach.
Figure 1. Flowchart of the proposed post-filtering approach.
Sensors 22 01353 g001
Figure 2. Proposed attention-based shared weights quality enhancement convolutional neural network, which is built using the following blocks: convolutional Block (CB), deconvolutional block (DB), attention-based shared weights block (ASB), attention-based shared weights residual block (ASRB), multi-head attention block (MHA), low-resolution feature fusion (LFF), and high-resolution feature fusion (HFF). The layer structure of each building block is presented in Figure 3.
Figure 2. Proposed attention-based shared weights quality enhancement convolutional neural network, which is built using the following blocks: convolutional Block (CB), deconvolutional block (DB), attention-based shared weights block (ASB), attention-based shared weights residual block (ASRB), multi-head attention block (MHA), low-resolution feature fusion (LFF), and high-resolution feature fusion (HFF). The layer structure of each building block is presented in Figure 3.
Sensors 22 01353 g002
Figure 3. The layer structure of the eight types of blocks used to build ASQE-CNN. (Top) The attention mechanism implemented using the convolutional block attention module (CBAM) proposed in [44]. (Middle) Simple building blocks (convolutional and deconvolutional blocks) and novel attention-based blocks with weight sharing. (Bottom) Novel multiresolution feature fusion blocks: (left) low-resolution feature fusion (LFF), (right) high-resolution feature fusion (HFF).
Figure 3. The layer structure of the eight types of blocks used to build ASQE-CNN. (Top) The attention mechanism implemented using the convolutional block attention module (CBAM) proposed in [44]. (Middle) Simple building blocks (convolutional and deconvolutional blocks) and novel attention-based blocks with weight sharing. (Bottom) Novel multiresolution feature fusion blocks: (left) low-resolution feature fusion (LFF), (right) high-resolution feature fusion (HFF).
Sensors 22 01353 g003
Figure 4. Sample images from LIVE1 [54] dataset. (Left to right) 1.bmp, 2.bmp, 3.bmp, and 4.bmp.
Figure 4. Sample images from LIVE1 [54] dataset. (Left to right) 1.bmp, 2.bmp, 3.bmp, and 4.bmp.
Sensors 22 01353 g004
Figure 5. Rate-distortion results for three images, (a,b) Image 801.png, (c,d) Image 821.png, (e,f) Image 849.png, and (g,h) over DIV2K_valid_HR [52]. (Left) bpc vs. PSNR results. (Right) bpc vs. SSIM results.
Figure 5. Rate-distortion results for three images, (a,b) Image 801.png, (c,d) Image 821.png, (e,f) Image 849.png, and (g,h) over DIV2K_valid_HR [52]. (Left) bpc vs. PSNR results. (Right) bpc vs. SSIM results.
Sensors 22 01353 g005
Figure 6. Rate-distortion results for three images, (a,b) Image 1.bmp, (c,d) Image 3.bmp, (e,f) Image 15.bmp, and over LIVE1 [54], (g,h). (Left) bpc vs. PSNR results. (Right) bpc vs. SSIM results.
Figure 6. Rate-distortion results for three images, (a,b) Image 1.bmp, (c,d) Image 3.bmp, (e,f) Image 15.bmp, and over LIVE1 [54], (g,h). (Left) bpc vs. PSNR results. (Right) bpc vs. SSIM results.
Sensors 22 01353 g006
Figure 7. BD-PSNR results for every image in (a) DIV2K_valid_HR [52] and (c) LIVE1 [54]. BD-rate (%) results for every image in (b) DIV2K_valid_HR [52] and (d) LIVE1 [54].
Figure 7. BD-PSNR results for every image in (a) DIV2K_valid_HR [52] and (c) LIVE1 [54]. BD-rate (%) results for every image in (b) DIV2K_valid_HR [52] and (d) LIVE1 [54].
Sensors 22 01353 g007
Figure 8. Visual results for images 801.png from DIV2K_valid_HR [52] and 15.bmp from LIVE1 [54]. (1st and 4th row) Original image and visual results of image and video codecs, JPEG [1], J2K [2], HEVC [3], and VVC [47]. (2nd and 5th row) Quality enhancement results corresponding to each codec. (3rd and 6th row) Pseudo-colored image comparison. Green marks the pixel positions where ASQE-CNN achieves better performance. Blue marks the pixel positions where the methods have the same performance. Red marks the pixel positions where the image/video codec (JPEG/J2K/HEVC/VVC) achieves better performance.
Figure 8. Visual results for images 801.png from DIV2K_valid_HR [52] and 15.bmp from LIVE1 [54]. (1st and 4th row) Original image and visual results of image and video codecs, JPEG [1], J2K [2], HEVC [3], and VVC [47]. (2nd and 5th row) Quality enhancement results corresponding to each codec. (3rd and 6th row) Pseudo-colored image comparison. Green marks the pixel positions where ASQE-CNN achieves better performance. Blue marks the pixel positions where the methods have the same performance. Red marks the pixel positions where the image/video codec (JPEG/J2K/HEVC/VVC) achieves better performance.
Sensors 22 01353 g008
Figure 9. Rate-distortion results: (a,b) for Image 801.png from DIV2K_valid_HR [52], (c,d) over DIV2K_valid_HR [52] dataset, (e,f) for Image 15.bmp from LIVE1 [54], (g,h) over LIVE1 [54] dataset. (Left) bpc vs. PSNR results. (Right) bpc vs. SSIM results.
Figure 9. Rate-distortion results: (a,b) for Image 801.png from DIV2K_valid_HR [52], (c,d) over DIV2K_valid_HR [52] dataset, (e,f) for Image 15.bmp from LIVE1 [54], (g,h) over LIVE1 [54] dataset. (Left) bpc vs. PSNR results. (Right) bpc vs. SSIM results.
Sensors 22 01353 g009
Figure 10. BD-PSNR results for every image in (a) DIV2K_valid_HR [52] and (c) LIVE1 [54]. BD-rate (%) results for every image in (b) DIV2K_valid_HR [52] and (d) LIVE1 [54].
Figure 10. BD-PSNR results for every image in (a) DIV2K_valid_HR [52] and (c) LIVE1 [54]. BD-rate (%) results for every image in (b) DIV2K_valid_HR [52] and (d) LIVE1 [54].
Sensors 22 01353 g010
Figure 11. Visual results for images 801.png in DIV2K_valid_HR [52]. (Top) Original image and the results of traditional codecs. (Middle) Quality enhancement results corresponding to each codec. (Bottom) Pseudo-colored image comparison. Green marks the pixel positions where ASQE-CNN achieves better performance. Blue marks the pixel positions where the methods have the same performance. Red marks the pixel positions where traditional codecs achieves better performance.
Figure 11. Visual results for images 801.png in DIV2K_valid_HR [52]. (Top) Original image and the results of traditional codecs. (Middle) Quality enhancement results corresponding to each codec. (Bottom) Pseudo-colored image comparison. Green marks the pixel positions where ASQE-CNN achieves better performance. Blue marks the pixel positions where the methods have the same performance. Red marks the pixel positions where traditional codecs achieves better performance.
Sensors 22 01353 g011
Figure 12. Pseudo-colored image comparison. Green marks the pixel positions where ASQE-CNN achieves better performance. Blue marks the pixel positions where the methods have the same performance. Red marks the pixel positions where HEVC/VVC achieves better performance. The bottom-right corner of each image presents the zoom in of the RoI marked by the cyan rectangle.
Figure 12. Pseudo-colored image comparison. Green marks the pixel positions where ASQE-CNN achieves better performance. Blue marks the pixel positions where the methods have the same performance. Red marks the pixel positions where HEVC/VVC achieves better performance. The bottom-right corner of each image presents the zoom in of the RoI marked by the cyan rectangle.
Sensors 22 01353 g012
Figure 13. Ablation study of the ASQE-CNN architecture. Average performance over the DIV2K_valid_HR [52] dataset: (a) bpc-vs-PSNR results; (b) bpc vs. SSIM results.
Figure 13. Ablation study of the ASQE-CNN architecture. Average performance over the DIV2K_valid_HR [52] dataset: (a) bpc-vs-PSNR results; (b) bpc vs. SSIM results.
Sensors 22 01353 g013
Table 1. Image datasets.
Table 1. Image datasets.
UsageDataset NameNumber of ImagesMegapixels
TrainingDIV2K_train_HR [52]8001.3–4 MP
TestingDIV2K_valid_HR [52]1001.6–4 MP
TestingLIVE1 [54]290.28–0.39 MP
Table 2. HEVC-VTSEQ: the Common-Test Dataset [56].
Table 2. HEVC-VTSEQ: the Common-Test Dataset [56].
IDVideo Sequence NameNumber of FramesClassFrame Resolution
A1Traffic150A 2560 × 1600
A2PeopleOnStreet150
B1Kimono240B 1920 × 1080
B2ParkScene240
B3Cactus500
B4BQTerrace600
B5BasketballDrive500
C1RaceHorses300C 832 × 480
C2BQMall600
C3PartyScene500
C4BasketballDrill500
D1RaceHorses300D 416 × 240
D2BQSquare600
D3BlowingBubbles500
D4BasketballPass500
Table 3. Average Bjøntegaard metrics for RGB images.
Table 3. Average Bjøntegaard metrics for RGB images.
MethodBD-Rate (%)BD-PSNR (dB)
DIV2K [52]LIVE1 [54]DIV2K [52]LIVE1 [54]
JPEG+ASQE-CNN 31.44 26.59 2.006 1.7442
J2K+ASQE-CNN 16.02 14.26 0.905 0.8190
HEVC+ASQE-CNN 54.61 52.78 3.448 3.7380
VVC+ASQE-CNN 19.39 18.79 1.201 1.3056
Table 4. Average PSNR (dB) results over LIVE1 [54] for grayscale images.
Table 4. Average PSNR (dB) results over LIVE1 [54] for grayscale images.
q i JPEG
[1]
ARCNN
[13]
RED30
[62]
DRRN
[63]
ARN
[64]
STRNN
[24]
JPEG+
ASQE-CNN
1027.7728.7329.3229.2129.2729.3829.42
2030.0730.8931.6931.1931.3431.7231.79
4032.3533.6333.1334.0634.16
Table 5. Average Bjøntegaard metrics for grayscale images.
Table 5. Average Bjøntegaard metrics for grayscale images.
MethodY-BD-Rate (%)Y-BD-PSNR (dB)
DIV2K [52]LIVE1 [54]DIV2K [52]LIVE1 [54]
JPEG+ASQE-CNN 26.21 24.29 1.826 1.7415
J2K+ASQE-CNN 19.31 19.23 0.896 0.8436
HEVC+ASQE-CNN 21.34 17.32 1.367 1.3539
VVC+ASQE-CNN 16.18 15.28 1.039 1.1171
Table 6. Video quality enhancement results computed as Y-BD-rate (%) savings for HEVC-vTSEQ.
Table 6. Video quality enhancement results computed as Y-BD-rate (%) savings for HEVC-vTSEQ.
IDHEVC [3] CodecVVC [47]
(All Frames)(30 Frames)
VR-CNNFE-CNNMLSDRNRRCNNVRCNN-BNFQE-CNNProposed
[15][26][37][28][18][4]ASQE-CNN
A1−5.6−6.6−9.8−8.7−11.0−18.46−14.04
A2−5.4−5.8−8.7−9.3−9.6−17.09−11.85
B1−2.5−5.7−6.2−6.6−11.8−8.4−19.02−16.32
B2−4.4−5.1−6.5−7.6−8.7−8.0−18.12−20.43
B3−4.6−4.9−5.8−7.0−10.5−7.7−14.10−15.46
B4−2.6−2.5−3.7−3.3−7.7−4.3−12.87−12.39
B5−2.5−5.2−4.4−6.7−14.5−8.5−13.37−14.35
C1−4.2−4.2−4.7−6.3−6.7−8.4−14.65−14.59
C2−5.1−5.4−6.0−9.9−7.5−11.8−15.94−15.87
C3−3.6−3.6−4.3−5.9−4.0−6.5−13.43−14.45
C4−6.9−8.6−8.3−15.1−8.5−14.7−19.54−16.34
D1−7.6−7.7−8.9−10.0−6.6−9.6−14.20−14.93
D2−3.8−4.2−4.3−7.4−3.4−6.8−11.61−12.82
D3−4.9−5.0−5.7−6.8−4.3−7.0−12.32−15.05
D4−5.3−5.9−6.6−10.0−7.4−10.6−12.74−15.55
A−5.5−6.2−9.3−9.0−10.3−17.77−12.95
B−3.3−4.7−5.3−6.2−10.6−7.4−15.51−15.79
C−5.0−5.5−5.8−9.3−6.7−10.4−15.89−15.31
D−5.4−5.7−6.4−8.6−5.4−8.5−12.72−14.59
All−4.8−5.5−8.4−8.3−9.2−15.47−14.66
Table 7. Video quality enhancement results computed as Y-BD-PSNR improvement over HEVC-vTSEQ [56].
Table 7. Video quality enhancement results computed as Y-BD-PSNR improvement over HEVC-vTSEQ [56].
ClassHEVC [3]VVC [47]
(Average)(All Frames)(30 Frames)
A1.2007 dB0.7551 dB
B0.7827 dB0.6743 dB
C1.2006 dB1.0315 dB
D1.3279 dB1.0689 dB
ALL1.1280 dB0.8824 dB
Table 8. Average Bjøntegaard metrics for RGB images.
Table 8. Average Bjøntegaard metrics for RGB images.
MethodBD-RateBD-PSNR#paramTime (bs = 100)
noAttention 19.51 % 1.1400 dB 0.83 M144 ms
noWeightSharing 20.10 % 1.1772 dB 1.48 M169 ms
noMFF (U-Net) 20.01 % 1.1701 dB 0.56 M 113 ms
Proposed 31.55 % 2.0166 dB 0.89 M165 ms
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Schiopu, I.; Munteanu, A. Deep Learning Post-Filtering Using Multi-Head Attention and Multiresolution Feature Fusion for Image and Intra-Video Quality Enhancement. Sensors 2022, 22, 1353. https://doi.org/10.3390/s22041353

AMA Style

Schiopu I, Munteanu A. Deep Learning Post-Filtering Using Multi-Head Attention and Multiresolution Feature Fusion for Image and Intra-Video Quality Enhancement. Sensors. 2022; 22(4):1353. https://doi.org/10.3390/s22041353

Chicago/Turabian Style

Schiopu, Ionut, and Adrian Munteanu. 2022. "Deep Learning Post-Filtering Using Multi-Head Attention and Multiresolution Feature Fusion for Image and Intra-Video Quality Enhancement" Sensors 22, no. 4: 1353. https://doi.org/10.3390/s22041353

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop