Next Article in Journal
Online Learning for Wearable EEG-Based Emotion Classification
Previous Article in Journal
Quality Management System for an IoT Meteorological Sensor Network—Application to Smart Seoul Data of Things (S-DoT)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vision Transformers in Image Restoration: A Survey

1
Robotics and Internet-of-Things Laboratory, Prince Sultan University, Riyadh 12435, Saudi Arabia
2
Department of Electronics and Electrical Communications Engineering, Faculty of Electronic Engineering, Menoufia University, Menouf 32952, Egypt
3
SE & ICT Laboratory, LR18ES44, ENICarthage, University of Carthage, Tunis 1054, Tunisia
4
Security Engineering Laboratory, Computer Science Department, Prince Sultan University, Riyadh 11586, Saudi Arabia
5
RIADI Laboratory, University of Manouba, Manouba 2010, Tunisia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(5), 2385; https://doi.org/10.3390/s23052385
Submission received: 18 January 2023 / Revised: 14 February 2023 / Accepted: 17 February 2023 / Published: 21 February 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
The Vision Transformer (ViT) architecture has been remarkably successful in image restoration. For a while, Convolutional Neural Networks (CNN) predominated in most computer vision tasks. Now, both CNN and ViT are efficient approaches that demonstrate powerful capabilities to restore a better version of an image given in a low-quality format. In this study, the efficiency of ViT in image restoration is studied extensively. The ViT architectures are classified for every task of image restoration. Seven image restoration tasks are considered: Image Super-Resolution, Image Denoising, General Image Enhancement, JPEG Compression Artifact Reduction, Image Deblurring, Removing Adverse Weather Conditions, and Image Dehazing. The outcomes, the advantages, the limitations, and the possible areas for future research are detailed. Overall, it is noted that incorporating ViT in the new architectures for image restoration is becoming a rule. This is due to some advantages compared to CNN, such as better efficiency, especially when more data are fed to the network, robustness in feature extraction, and a better feature learning approach that sees better the variances and characteristics of the input. Nevertheless, some drawbacks exist, such as the need for more data to show the benefits of ViT over CNN, the increased computational cost due to the complexity of the self-attention block, a more challenging training process, and the lack of interpretability. These drawbacks represent the future research direction that should be targeted to increase the efficiency of ViT in the image restoration domain.

1. Introduction

Image Restoration is an umbrella term incorporating many computer vision tasks. These approaches seek to generate a better version of an image that is retrieved in a lower-quality format. The leading seven tasks of image restoration are: Super-Resolution, Image Denoising, General Image Enhancement, JPEG Compression Artifact Reduction, Image Deblurring, Removing Adverse Weather Conditions, and Image Dehazing. All these tasks share among them many common characteristics. First of all, they are image generation tasks. This means that we generate new data that tries to mimic the authentic representation of the information being captured inside the image. Second, they all try to remove a specific type of corruption applied to the image. This corruption may be a noise, a bad quality image sensor, a capturing imperfection, a weather condition that hides some information, etc. Third, all of them are now becoming based on deep learning techniques. This architectural change is due primarily to their effectiveness compared to old techniques. Hence, any advancement in deep learning theory directly affects these tasks and brings new ideas for any image restoration researcher.

1.1. Techniques Used in Image Restoration

Since deep learning popularity began in the computer vision field in 2012 [1], Convolutional Neural Networks (CNN) have been the default feature extractor used to learn patterns in the data. An overwhelming number of models and architectures have been designed and improved over the years to increase the learning capability of CNNs. Generic architectures, as well as task-specific architectures, were designed. Before 2020, it was the rule for any image restoration researcher to customize a model based on CNN to increase the efficiency of the state-of-the-art works on his task [2,3,4,5]. This approach is adopted in most computer vision tasks [6,7,8,9,10,11,12,13,14].
Since 2017, the Transformer architecture has made a significant breakthrough in the Natural Language Processing (NLP) domain. The self-attention mechanism and the new proposed architecture have proven to have a considerable ability to manage sequential data efficiently. Nevertheless, adoption in the computer vision domain was delayed until 2020, when the Vision Transformer (ViT) architecture was first introduced [15]. Up to now, ViT has proven efficient and competed very well with CNN. Here also, an overwhelming number of architectures inspired by ViT have been introduced with an attractive efficiency. Moreover, ViT has been demonstrated to solve many issues traditionally associated with CNN. Therefore, a debate between these two architectures’ advantages and drawbacks has emerged to decide where to go in future research directions.
Many techniques have been developed for image restoration, including those based on CNNs and ViTs. However, CNNs have several limitations regarding their ability to learn a mapping from degraded images to reach their original counterparts. In addition, the quality of the images generated by CNNs is often far from the original images. In contrast, some techniques address these challenges of CNN. These techniques include the Generative Adversarial Networks (GANS) [16] and the Diffusion model [17], as shown in Figure 1. GANs are a class of deep learning models that involve training two neural networks, a generator, and a discriminator, to work together to generate new, synthetic data similar to the real data. Despite the success of GANs in various tasks, their application in image restoration methods has often been found to produce suboptimal results due to issues such as pattern collapse, excessive smoothing, artifacts, and training instability. The Diffusion model, on the other hand, is a generative model that uses a diffusion process to generate new images by modeling the image formation process. However, while diffusion models have been used in image generation and denoising tasks, they may struggle to generate images with accurate global features and require a high computational cost. This is because the diffusion process models the image formation process, but they may not be as effective as other models in capturing global features. Additionally, the high computational cost is a limitation that needs to be considered when using these models.
The current study aimed to answer the question: what is the impact of ViT on the Image Restoration domain? The study is based on a study of around 70 papers. First, every research work has been classified into the seven fields of image restoration. Then, a description of the state-of-the-art techniques in these seven fields is presented. Later, the advantages, drawbacks, and future research directions are discussed.

1.2. Related Surveys

To our knowledge, no literature survey has targeted this problematic. The existing surveys about ViT are more generic in scope and do not consider the image restoration domain’s peculiarities. For example, Khan et al. [18] have explored the use of vision transformers in popular recognition tasks, generative modeling, multi-modal tasks, video processing, low-level vision, and three-dimensional analysis. Han et al. [19] undertook a generic review of ViT for different computer vision tasks; they compared their advantages and drawbacks. Then, they classified the tasks into four classes: low-level vision, high/mid-level vision, backbone network, and video processing. They also discussed the application of vision transformers to real devices. Islam et al. [20] provided an overview of the best-performing modern vision Transformer methods and compared the strengths, weaknesses, and computational costs between vision Transformers and CNN methods on the most benchmark datasets. Shamshad et al. [21] were more specific and focused on one domain. They presented a comprehensive review of the application of vision transformers in medical imaging, including medical image segmentation, classification, detection, clinical report generation, synthesis, registration, reconstruction, and other tasks. They also discussed unresolved problems in the architectural designs of vision transformers.
Looking for surveys that focused only on image restoration tasks, no one has elaborated a study about ViT architectures in this regard. For example, Su et al. [22] generally tackled deep learning algorithms without focusing on ViT specifically. They presented a comparative study of deep learning techniques used in image restoration, including image super-resolution, dehazing, deblurring, and denoising. They also discussed the deep network structures involved in these tasks, including network architectures, skip connection or residual, and receptive field in autoencoder mechanisms. They also focused on presenting an effective network to eliminate errors caused by multi-tasking training in the super-resolution task. In Table 1, the related surveys to our research are described based on two main characteristics. The first is the algorithm studied (CNN and/or ViT). The second is the area of the survey. As concluded, a specific survey that studies the effect of ViT in the Image Restoration domain is still lacking in the current state of the art.

1.3. Main Contributions and Organization of the Survey

A gap is noted in these related surveys. A specific survey that analyses the impact of only ViT-based architectures on the image restoration domain is still needed, considering all of its peculiarities. The main contributions elaborated in this research are:
  • The study lists the most important ViT-based architectures introduced in the image restoration domain, classified by each of the seven subtasks: Image Super-Resolution, Image Denoising, General Image Enhancement, JPEG Compression Artifact Reduction, Image Deblurring, Removing Adverse Weather Conditions, and Image Dehazing.
  • It describes the impact of using ViT, its advantages, and drawbacks in relation to these image restoration tasks.
  • It compares the efficiency metrics, such as The Peak Signal-to-Noise Ratio (PSNR) and the Structural Similarity Index Metric (SSIM) of the ViT-based architectures, on the main benchmarks and datasets used in every task of image restoration.
  • It discusses the most critical challenges facing ViT in image restoration and presents some solutions and future work.
This study is organized as follows, Section 2 introduces the ViT architecture and presents its success keys, including self-attention, sequential feature encoding, and feature positioning. Section 3 details the evaluation metrics used to assess the performance of image restoration algorithms. These evaluation metrics will be used later for the whole paper, and their definitions are worth describing from the beginning. Section 4 describes the state of the art of ViT-based architectures in each of the seven image restoration tasks. Section 5 presents a discussion of the advantages, drawbacks, and the main challenges of ViT in image restoration. Finally, in Section 6, our work is concluded alongside a description of the next research problematics to be addressed in future works.

2. Vision Transformer Model

In this section, we introduce ViT and the most important principles upon which it is built, including structure, self-attention, multi-headed self-attention, and the mathematical background behind ViT. The ViT was introduced by [15] in 2020. The ViT is a deep neural network architecture for image recognition tasks based on the Transformer architecture initially developed for natural language processing tasks. The main idea behind ViT is to treat an image as a sequence of tokens (typically, image patches). Then, the transformer architecture is used to process this sequence. The transformer architecture, on which ViT is based, has been adapted to many tasks and is effective in many of them, such as image restoration and object detection [23,24].
The ViT is based on the Transformer architecture. Figure 2 shows the main stages of VIT architecture. where the image is tokenized and embedded by dividing it into a grid of non-overlapping patches, flattening each patch, and mapping it to a high-dimensional space through a linear transformation followed by normalization. This process, called tokenization and embedding, allows the model to learn both global and local information from an image.
The transformer architecture is designed to process any sequence, but it does not explicitly consider each token’s position in the sequence. To address this limitation, the ViT architecture uses pre-defined positional embeddings. These positional embeddings are additional vectors that encode the position of each token in the sequence, and they are added to the token embeddings before they are passed through the transformer layers. This mechanism allows the model to capture the relative position of the tokens and to extract spatial information from the image.
The core of the ViT architecture lies in the Multi-head Self-Attention (MSA). This mechanism allows the model to simultaneously attend to different parts of the image. It is composed of several different “heads”, each of which computes attention independently. These attention heads can attend to different regions of the image and produce different representations, which are then concatenated to form a final image representation. This allows the ViT to capture more complex relationships between the input elements, as it can attend to multiple parts of the input simultaneously. However, it also increases the complexity and computational cost of the model, as it requires more attention to heads and more processing to combine the outputs from all the heads. MSA can be expressed as follows:
M u l t i H e a d Q , K , V = C o n c a t H e a d 1 ,   H e a d 2 ,   H e a d n
where H e a d n = s e l f   A t t e n t i o n Q . W n Q ,   K . W n K ,   V . W n V , n is the number of heads. MSA depends on the self-attention mechanism introduced by the authors of [25]. The basic idea of self-attention is to estimate how closely one element relates to the other elements in the sequence of elements. In digital images, the image is divided into several patches, then each patch is converted into a sequence, and then self-attention estimates the relevance of one sequence to the rest of the sequences.
The self-attention mechanism is the backbone of transformers, which explicitly model the interactions and connections between all sequences of prediction tasks. The self-attention layer also collects global information and features from the entire input sequence, distinguishing it from CNN. It is for this reason that transformers have a larger model capacity. To understand more about the mechanism of self-attention, we may consider a sequence of 𝓃 elements X 1 ,   X 2 ,   X 3 ,   X 𝓃 by X 𝓃 × d , where d is the embedding dimension to represent each element. The purpose of the attention mechanism is to estimate the interactions and connections between all n elements by encoding each element and then capturing global information and features. In order to implement the attention mechanism, three learnable weight matrices must first be defined, including Queries W Q d × d q , Keys W K d × d k , and Values W V d × d v , where d q = d k . The X input sequence is projected by weights matrices Q = X . W q , K = X . W k , V = X . W v . The output of attention layer Z is Equation (2):
Z = s o f t m a x Q K T d q V ,
The attention mechanism computes the dot product of Q (query) with all K (keys) for a given element in the sequence. Then the SoftMax operator is used for normalization to get the attention outputs. Each element in the sequence becomes the sum weight of all elements in the entire sequence; attention outputs generate these weights. Then, using the dot product, the attention output is multiplied by the V (value) matrix. Figure 3 shows an example of the self-attention model.
The main advantage of the self-attention mechanism compared to the convolution mechanism is that the values of the filters are calculated dynamically instead of static filters, as in the case of the convolution operation, where the filters remain the same for any input. In addition, self-attention is distinguished from standard convolution in the stability of permutations and changes in the number of insertion points. As a result, it can process irregular data and input. Based on literature reviews, it appears theoretically that the process of self-attention with positional encodings is more flexible in extracting local features than convolutional models [26]. Cordonnier et al. [27] investigated the connections between self-attention and convolution operations. Their results show that self-attention with sufficient parameters is a very flexible and more general process that can extract local and global features. Furthermore, based on previous studies and research findings on various computer vision tasks, self-attention can learn both local and global features. It also provides adaptive learning of kernel weights, as in research on deformable convolutions [28].
The ViT architecture requires a large amount of data for training to achieve optimal performance. This is a common challenge encountered with transformer models. To address this limitation, a two-stage training approach is often employed. In the first stage, supervised or self-supervised [24,29] pre-training is performed on a large dataset or a combination of several datasets [30]. In the second stage, the pre-trained weights are used to fine-tune the model on smaller or medium-sized datasets for specific tasks such as object detection, classification, and image restoration. This approach has been demonstrated to be effective in previous research, where it was shown that the accuracy of the Vision Transformer model decreased by 13% when trained only on the ImageNet dataset and increased when trained on the JFT dataset containing 300 million images. This highlights the importance of pre-training transformer models for vision and language on large-scale datasets.
Obtaining labeled datasets for training AI models can be a significant challenge due to the high cost and difficulties associated with the process. To address this, self-supervised learning (SSL) has emerged as a promising alternative to traditional supervised learning. SSL enables training models on a large number of parameters, up to a trillion, as demonstrated in [31]. The basic principle of SSL is to train a model on unlabeled datasets by applying various transformations to the images, such as slight adversarial changes or replacing one object with another in the same scene, without altering the semantics of the base class. This allows the model to learn from the data without needing manual labeling. For more information and a comprehensive survey of SSL technology, readers can refer to references [32,33].

3. Evaluation Metrics

This section introduces the most widely used image quality measurement methods in image restoration. In addition, we present the most modern methods for measuring image quality based on ViT. The quality of digital images can be defined based on the measurement method. Measurement methods depend on viewers’ perceptions and visual attributes. Image Quality Assessment (IQA) methods are classified into subjective and objective methods [34]. Subjective methods rely on image quality evaluators, but these methods take a lot of time, effort, and cost due to large datasets. Therefore, objective methods are more appropriate in image restoration tasks, according to [35]. Objective methods need the ground-truth image and the predicted image. PSNR is one of the most widely used image quality measures. PSNR is maximum signal to maximum noise. In image restoration, the PSNR results from divisibility between the maximum pixel value and Mean Square Error (MSE) between the ground-truth image and the predicted image. The PSNR is defined as:
P S N R i g t ,   i p = 10   log 10 m a x 2 M S E ,
where m a x is the maximum pixel value, i.e., the max value of 255 for 8-bit color depth and M S E is defined as:
M S E i g t ,   i p = 1   t k = 1 t i g t k i p k 2 ,
where i g t is the ground truth image, and i p   is the predicted image. The PSNR approach can be misleading because the predicted image may not be perceptually similar to a ground truth image [36,37]. However, the PSNR approach is still used in all image restoration research, and research results are also compared using it. Human perception is the best measure of image quality for its ability to extract structural information [38]. However, ensuring human visual validation of the data remains very expensive, especially with large datasets. An image quality metric that measures the structural similarity between ground truth images and predicted images by comparing luminance, contrast, and structural information is called SSIM [39].
S S I M i g t ,   i p = 2 μ i g t μ i p + 𝒸 1 2 σ i g t i p + 𝒸 2 μ i g t 2 + μ i p 2 + 𝒸 1 σ i g t 2 + σ i p 2 + 𝒸 2 ,
where μ i g t is the average value for the ground truth image, μ i p is the average value of the predicted image, σ i g t is the standard deviation for the ground truth image, σ i p is the standard deviation of the predicted image, and σ i g t i p = μ i g t i p   μ i g t μ i p is the co-variation. c 2   i s   k 2 L 2 ,   and   c 1   i s   k 1 L 2 are two variables to avoid division by zero, k 2 = 0.03 , and k 1 = 0.01 .
In image restoration research, two of the most widely used image quality metrics are SSIM and PSNR. SSIM, in particular, is based on the human perception of structural information in images and is commonly used for generic images, as discussed in [40]. However, when applied to medical images, where brightness and contrast may not be consistent, the SSIM metric can be unstable, as reported in [41]. To address this issue, the authors in [42] proposed a model for measuring the perceptual quality of images. They tested this model using the Restormer model [43] in various noise removal and adverse weather conditions. They then applied an object detection model to the same images before and after using the Restormer model. They found that while SSIM and PSNR results are often unstable, the results from the object detection model were more consistent. This led the authors to propose a Grad-CAM model to measure image quality, which bridges the gap between human perception and machine evaluation [44].
Recently, Cheon et al. [45] proposed a new Image Quality Transformer (IQT) which utilizes the transformer architecture. IQT extracts perceptual features from both ground-truth and predicted images using a CNN backbone. These features are then fed to the transformer encoder and decoder, which compares the ground-truth images and the predicted images. The transformer’s output is then passed to a prediction head, which predicts the quality of the images. In further research, Conde et al. [46] replaced the IQT encoder-decoder with the Conformer architecture, which utilizes convolution layers from the Inception-ResNet-v2 model [47] along with attention operations to extract both global and local features.

4. Image Restoration Tasks

In this section, we will introduce all transformer-based image restoration tasks, in addition to presenting a comparative study between all models of vision transformers in each sub-task. Figure 4 presents a taxonomy of all vision tasks with the most common head transformer models.

4.1. Image Super-Resolution

Super-resolution is a technique for reconstructing high-resolution images from low-resolution input images. SR has been widely applied in image restoration tasks due to its value in various applications and its ability to overcome the limitations imposed by imaging systems [48]. SR can be broadly classified into two categories: multi-image super-resolution (MISR) and single-image super-resolution (SISR). MISR generates a single high-resolution image from multiple low-resolution images, while SISR generates a high-resolution image from a single low-resolution input. Vision transformer (VT)-based SR models, such as HAT [49], SwinIR [50], and SwinFIR [51], have gained widespread adoption in image restoration and are depicted in Figure 5 as a general block diagram.
In most applications, obtaining multiple images of the same scene is often complex and costly. Therefore, using only a single image in super-resolution technology is a suitable and effective solution for many applications. In most research papers, low-resolution images are typically obtained by applying a degradation model to high-resolution images, which involves blurring followed by downsampling. Traditional methods use interpolation and blur to enhance low-resolution images and rely on iterative optimization frameworks. Yang et al. [52] conducted an investigative comparison of ten high-resolution papers that utilize iterative optimization techniques and sparse representation, including [53,54,55].
Traditional methods for image restoration are often ineffective and time-consuming. In contrast, deep neural networks can learn from large amounts of training data rather than relying on complex statistical models. With the advancement of computing devices and GPUs, neural networks have become more prevalent. In recent years, the vision transformer has emerged as a strong contender in the field of computer vision, including in the area of image SR. According to the results of many research papers, vision transformers have shown high accuracy and efficiency in SR techniques. In this survey, we can classify super-resolution transformer networks based on the input size, the type of images, and the type of task. A low-resolution image X ^   can be expressed as
X ^ = D X B + N ,
where X is a high-resolution image, B is effect blur, and N is random noise. SR models based on transformers can be divided based on the nature of the images. Using a generic image to train super-resolution models is the most common and comprehensive method in most super-resolution applications. One of the most famous models is SwinIR [50], which is widely popular. SwinIR is an image restoration model that does more than one task depending on the type of training. SwinIR is inspired by the swin v1 [56] transformer used in image classification. SwinIR consists of a convolutional layer at the beginning called shallow feature extraction. In the middle, SwinIR consists of 36 consecutive swin blocks divided into six stages; each stage consists of 6 blocks and is called deep feature extraction. Finally, the reconstruction layer, which contains more than one type, depending on the task, in super-resolution, is an enlargement layer for the image dimensions. Because of the high efficiency of the swin model in extracting global and local features, it has also been used in many super-resolution models such as SWCGAN [57], SwinFIR [51] and 3DT-Net [58].
The HAT [49] model is a model that combines self-attention and channel attention, so it has a high ability to aggregate features. HAT also proposed an overlapping cross-attention module to collect information through cross-windows in a new and highly efficient way. Based on the results shown in Table 2, Table 3 and Table 4, HAT is the highest model in PSNR and SSIM scores across all scales. The SwinFIR model is a recent variation of the SwinIR model that utilizes Fast Fourier Convolution (FFC) components to extract global information suitable for the task of super-resolution. This is achieved by combining global and local features extracted by the FFC and residual modules.
Recently, authors in [65] proposed the first transformer-based light field super-resolution (LFSR) processing model, called SA-LSA. The model divides each light field into a set of image sequences and utilizes a combination of convolutional layers and a self-attention network to reinforce non-local spatial angular dependencies in each sequence. Table 5 compares SA-LSA with other super-resolution models on light field images.
SWCGAN uses a swin-inspired switched window self-attention mechanism combined with the GAN model to combine the advantages of swin switches with convolutional layers. SWCGAN has been applied to remote-sensing images. Table 6 compares super-resolution models on remote sensing images.
Transformers have also been applied to the super-resolution of medical images, such as in the case of the ASFT model [68]. ASFT consists of three branches, two of which transmit similar features from MRI slices, and the third branch builds the super-resolution slice. The performance of ASFT is compared with other super-resolution models on MIR medical images in Table 7 of the corresponding paper.
The 3DT-Net architecture is a transformer-based approach that leverages the spatial information in the hyperspectral images. It utilizes multiple layers of the Swin Transformer in place of 2D-CNN layers and employs 3D convolutional layers to take advantage of the spatial spectrum of the data. This architecture has been applied to hyperspectral image processing, which inherently possesses multi-dimensional characteristics. It is effective in this task, and it is worth noting that the performance of 3DT-Net has been evaluated and compared against other state-of-the-art models on hyperspectral images, as shown in Table 8.
Table 9 presents a collection of various super-resolution models that do not belong to a standardized dataset. The authors in [73] propose a transformer-based super-resolution texture network, called TTSR. This model uses reference images to create a super-resolution image from multiple low-resolution images and reference images. The VGG model [74] extracts features from the reference images. Additionally, a self-attention-based transformer model is proposed for extracting and transferring texture features from reference and low-resolution images.
A transformer-based super-resolution architecture, called CTCNet, was proposed by Gao et al. [75] for the task of face image super-resolution. CTCNet is composed of an encoder-decoder architecture and a novel global-local feature extraction module that can extract fine-grained facial details from low-resolution images. The proposed architecture demonstrated promising results and potential for the task of face image super-resolution. BN-CSNT [76] is a network that combines channel splitting and the Swin transformer to extract context and spatial information. Additionally, it includes a feature fusion module with an attention mechanism, and it has been applied to thermal images to address the super-resolution problem. One of the main advantages of transformers in super-resolution is their adaptability to different types of images, similar to CNNs.

4.2. Image Denoising

Noise reduction, or denoising, is crucial in image processing and restoration. Removing noise from images is often used as a preprocessing step for many computer vision tasks, and it is also used to evaluate optimization methods and image prior models from a Bayesian perspective [77]. There are various approaches to reducing noise, including traditional methods such as BM3D [78], which enhances image contrast by assembling blocks of similar images from 3D images. There are also machine learning-based methods, and in [79], there is a comprehensive study on the use of machine learning for noise reduction in images. The wide range of noise reduction models is due to their widespread use in signal processing. In recent years, deep learning methods have been dominant in terms of perceptual efficiency and quality for noise reduction. For an overview of deep learning models based on CNNs for noise reduction, see [80]. In this section, we will focus on vision transformer-based noise reduction algorithms. Noisy images production can be expressed mathematically as
y = x + n ,
where y is a noisy image, x is a clean image, and n is the random noise. Noise removal algorithms based on deep learning are divided into two methods.
There are two main approaches for removing noise from images using ViT, as shown in Figure 6. The first approach, known as “noise-to-noise,” involves using deep learning networks to estimate the noise in an image and then subtracting the estimated noise from the noisy image to generate a noise-free image. The second approach involves using deep learning networks to estimate noise-free images from noisy images directly. For example, Prayuda et al. [81] proposed a convolutional vision transformer (CVT) for noise removal using the noise-to-noise approach, which employs residual learning to reduce noise from the noisy image by estimating the relationship between the noisy image and its noise map. Several previous research studies, including SwinIR, Uformer, Restormer, and IPT, have used a single encoder for all image restoration tasks but have utilized multiple decoders or reconstruction layers depending on the specific task.
Liu et al. [82] have proposed a new network called DnT for unsupervised noise removal from images. DnT combines a CNN and Transformer to estimate clean images from noisy images, using a loss function that is measured by pairs of noisy independent images generated from the input images. This method, known as R2R [83], effectively removes noise from images. In addition, the authors of [84] propose an efficient model called TECDNet, which utilizes a transformer for data encoding and a convolutional network for the decoder. Additionally, a convolutional network is incorporated to reduce computational complexity. Another transformer-based model for noise removal is TC-Net, which was introduced by Xue et al. [85]. This network consists of extra skip-connections, window multi-head self-attention, a convolution-based forward network, and a deep residual shrinkage network. The components of TC-Net work together to integrate features between layers, reduce computational complexity, and extract local features. The swin transformer is also frequently used in various shapes and structures for noise removal in image restoration tasks due to its ability to effectively extract important features from input images and its adaptability to different structures. Table 10 presents a comparison of various ViT models for the task of denoising in both generic and medical images. It can be observed that when the level of noise is high, the performance of the models, as measured by the SSIM/PSNR metrics, decreases.
Fan et al. [87] have used a swin transformer as a basic block inside the layers of the U-net model, and it is called SUNet. Li et al. [88] collected a real dataset of noisy polarized color images to remove the disturbing noise and used a conventional camera to take the same images and use these images as ground truth. Then they proposed a hybrid transformer model based on the attention mechanism to remove the disturbing noise from the polarized color images, called Pocoformer. The authors in [89] have proposed a model to improve image quality and reduce noise in Low-Dose Computed Tomography (LDCT). Their proposed model, called TransCT analyzes distorted images into High-Frequency (HF) and Low-Frequency (LF) components. Several convolutional layers extract content and texture features from LF and texture features from HF. The extracted features are then fed into a modified transformer with three encoders and decoders blocks to obtain well-polished features. Finally, the refined features are combined to restore high-quality LDCT images. LDCT images are the most common in diagnosing diseases. However, LDCT suffers from loud noises more than normal CT images. Therefore, the authors of [90] have proposed a new convolution-free Token-to-Token (T2T) transformer model to remove noise from LDCT images. Luthra et al. [91] have proposed a transformer that improves edge-to-edge using a Sobel filter called Eformer. In addition, non-overlapping self-attention based on a shifted window is used to reduce computational costs while using residual learning to remove natural noise in LDCT images.

4.3. Image Deblurring

Blurry images can be caused by various factors, such as the random movement of objects or the movement of the camera itself and physical limitations. This makes the problem of blurry images challenging to describe and solve. Blurry images are common in many image capture devices due to factors such as camera movement or scene changes [92]. It is not feasible to develop a mathematical formula that accurately captures the blur in an image, as it depends on a range of variables. In the past, various approaches have been proposed for deblurring images, including traditional methods. Wang et al. [93] made a comprehensive review of the traditional methods that occurred in common imaging and divided them into five frames according to the characteristics of each of them. Lay et al. [94] evaluated 13 single-image deblurring algorithms. Zhang et al. [95] presented a recent study on deep learning and CNN methods in terms of structure, loss function, and various applications. This paper will focus on transformer-based deep learning models to treat blurry images. Figure 7 presents some examples of blurry image processing. The leftmost image is a full, blurry image, the middle image is a cropped version of the blurry image, and the rightmost image is the reconstruction of the blurry image using the model.
Wang et al. [86] introduced a new transformer-based network called Uformer, consisting of a hierarchical U-net architecture. Uformer is based on two main designs: The locally enhanced window (LeWin) block significantly reduces computational cost when using it with high-resolution feature maps rather than window-based self-attention. Secondly, they designed a learnable multi-scale restoration modulator to adjust the spatial features at the multi-layer level. Zamir et al. [43] introduced a novel transformer called Restormer Transformer (Restormer). Transformers are characterized by high image quality accuracy but suffer from significant computational complexity. Restormer is a practical image restoration transformer that models global features and can be applied to large images. Restormer consists of a multi-Dconv head ‘transposed’ attention (MDTA) block. MDTA is distinct from the vanilla multi-headed [25] in calculating the inter-channel covariance to obtain an optimal attention map. An advantage of MDTA is that it uses global relationships of pixels and optimizes the local context for feature mapping. Table 11 illustrates various image deblurring models that are based on the ViT architecture. It is worth noting that all the datasets used in this demonstration are real-world data and not synthetic.
In order to reduce the computational complexity of transformers, the authors of [96] proposed a deep network called Multi-scale Cubic-Mixer. Besides, to reduce computational costs, they did not use any self-attention mechanism. The proposed model uses the fast Fourier transform to calculate the Fourier coefficients to use the real and imaginary components and thus obtain the image without blurring. The proposed new network extracts the long-range and local features from blurred images in the frequency domain. The authors [97] have proposed a hybrid transformer called Stripformer, which relies on the attention of inter-strip and intra-strip [98] and which estimates motion blur from a blurred image by projecting motion blur in vertical and horizontal directions in the Cartesian coordinate. In addition, the vertical and horizontal features extracted at each pixel are stored to provide more information for subsequent layers in order to deduce the blur pattern. This method enables the sequential extraction of multi-scale local features. Therefore, the stripformer can remove long- and short-range blurred artifacts.

4.4. Image Dehazing

Image dehazing involves the removal of non-linear noise produced by a turbid medium, such as dust and fog, which impairs the visibility of images and objects. The presence of hazing in images adds complex and difficult-to-predict noise, making it a challenging problem in image restoration. In addition, hazing reduces the accuracy and efficiency of computer vision algorithms. The generation of hazy images can be described using the atmospheric scattering model:
y = x e β 𝒹 + a 1 e β 𝒹 ,
where y is a hazy image, x is an image without hazing, a is global atmospheric light, β is the scattering coefficient, and 𝒹 is the distance between the camera and the object. Many computer vision algorithms degrade in hazing scenes. Figure 8 describes an atmospheric dispersion model describing the hazing formation.
The RESIDE dataset is commonly used for training and testing models for dehazing images. It includes 13,990 images in the Indoor Training Set (ITS), 500 images in the Synthetic Object Testing Set (SOTS), and 20 images in the Hybrid Subjective Testing Set (HSTS). To address the inconsistency problem of various features from CNN and transformer models, the authors of [99] propose a feature modulation module for combining hierarchical global features from transformers and hierarchical local features from CNN for image dehazing. Furthermore, Gao et al. [100] proposed a model that combines residual and parallel convolutional networks with a transformer-based channel attention module for detailed feature extraction called TCAM. TCAM consists of a channel attention module and a spatial attention module, and the channel attention module improves features along the channel.
Table 12 presents a comparison of different ViT models for the task of image dehazing. It can be observed that many of these models utilize the RESIDE dataset, which consists of both indoor and outdoor real-world images, for their evaluation. Li et al. [101] proposed a new two-phase de-hazing network. The first phase is based on the swin transformer and extracts key features from blurred images. In addition, the swin architecture has been improved by adding an Inter-block Supervision Mechanism (ISM). In the second phase, convolutional layers are used to extract local features and are also merged. The authors [102] have proposed a dual-stream network consisting of a CNN and transformers to extract global and local features. In addition, they proposed an atmospheric light estimation model based on atmospheric veils and partial derivatives. Song et al. [103] introduced a DehazeFormer model for removing hazing from generic images. Besides, they use multi-head self-attention (MHSA) that is dynamically trainable to adapt to different generics of images. Additionally, they suggest a soft reconstruction module based on SoftReLU. The contributions of Song et al. are to propose a modified normalization layer, spatial feature aggregation system, and soft activation function. Zhao et al. [104] suggest an improved framework for complementary features that relies on forcing the model to learn weak complementary features rather than iteratively learning ineffective features of the model. In addition, they proposed the Complementary Feature Selection Module (CFSM) to identify the most valuable features in learning.

4.5. Image JPEG Compression Artifact Reduction

With the significant development in cameras, most devices have become highly dependent on images, including airplanes, cars, and smartphones. Images and videos require huge storage space and high internet speed. This leads to the need for image compression and storage systems in all applications. However, most compression methods result in compression noises, such as compression artifacts [105]. JPEG is the most popular image compression technology on the Internet. JPEG compression technology consists of four stages: discrete cosine transformation (DCT), entropy coding, block division, and quantization. JPEG suffers from image dropouts at spatial boundaries because JPEG ignores the spatial connections between image blocks. Therefore, there are approaches based on filters [106,107,108] to improve the quality of images after compression, but they suffer from image blur. Deep learning techniques have emerged in the past few years to solve the JPEG compression artifact problem by mapping compressed images to ground-truth images [109,110,111]. In Table 13, a comparison of transformer-based models for removing JPEG compression artifacts is presented. The evaluation is done in terms of quality, the number of parameters, and the datasets used for a single image and the pairs of stereo images.
Modern techniques are divided into two parts in JPEG compression artifact reduction; the first is using one image to solve the problem, and the second is using two or more images. Each method has advantages and disadvantages. In this work, we present the latest JPEG compression artifact reduction techniques based on transformers. The most famous transformer-based model is SwinIR [50] which was mentioned before. Transformers are distinguished in capturing global and local features from CNN, which only extracts local features. SwinIR is a model that addresses more than one problem in image restoration, and the last layers of it are changed to suit the solution of each problem. A single convolution layer is used in the last block of the SwinIR model to remove the JPEG compression artifact. In addition, Charbonnier loss [113] was used in training the SwinIR model. The second type of JPEG artifact removal is the use of stereo image pairs.

4.6. Removing Adverse Weather Conditions

Many computer vision algorithms, including those used for detection, depth estimation, and segmentation [114,115,116], are sensitive to the surrounding conditions and can be affected by factors such as adverse weather. These algorithms have important applications in various systems that are integral to our daily lives, such as navigation and surveillance systems and Unmanned Aerial Vehicles (UAVs) [117,118,119]. As a result, it is necessary to address hostile weather conditions such as fog, rain, and snow that can negatively impact the reliability of vision systems. Traditional methods for removing adverse weather are often based on empirical observations [120,121,122]. However, these methods must be tailored to each weather condition, as an all-weather method is generally ineffective. Additionally, there are algorithms based on CNNs that have been developed for removing adverse weather, including deraining [123,124,125,126], desnowing [127,128,129], and raindrop removal [130,131]. In Figure 9, it can be observed that there are several transformer-based approaches for removing adverse weather conditions, which can be broadly categorized into two groups: methods that focus on removing a specific type of adverse weather and methods that are capable of removing a variety of adverse weather conditions.
Qin et al. [132] proposed ETDNet, which utilizes transformers to extract features from individual images to remove rain streaks. ETDNet employs expansion filters at different rates to estimate the appropriate kernel for rainy images, thereby allowing for the extraction of features in an approximate to exact manner. The authors of [133] utilized a Swin-transformer [56] for rain removal from single images, utilizing a Swin block in three parallel branches referred to as Mswt. Deep and shallow features are extracted through the use of four consecutive Mswt and skip-connections between Mswt. The authors [134] proposed a two-stage training and fine-tuning approach called Task-Agnostic Pre-Embedding (TAPE). TAPE is trained on natural images, and then the knowledge is used to help remove rain, snow, and noise from images in a single model. The authors [135] proposed a model that eliminates adverse weather conditions at once called TransWeather, which consists of a single decoder. The weather-type queries are identified by a multi-headed attention mechanism and matched with the values and keys taken from the features extracted from the transformer encoder. To reconstruct the images after identifying the type of deterioration, hierarchical features from the encoder and features from the decoder are combined and then projected by a convolutional block on the image space. Therefore, TransWeather consists of one encoder and one decoder to eliminate adverse weather and generate a pure image. Transformers can extract important global features compared to CNN. However, when patches are large, transformers fail to pay attention to small information and detail.
Table 14 presents a comparison of transformer-based models for removing adverse weather conditions. The models can be divided into two categories: those that remove a specific type of adverse weather and those that are capable of removing multiple types of adverse weather. It is important to note that it is generally preferable to use models that can handle multiple types of adverse weather conditions in real-world applications. For example, Liu et al. [136] designed an image restoration model called SiamTrans, inspired by Siamese transformers. SiamTrans is trained on an extensive dataset in a noise reduction task. Then knowledge transfer is used to train SiamTrans on many low-level image restoration tasks, including demoireing, deraining, and desnowing. In addition, the SamTrans consists of the encoder, decoder, self-attention units, and temporal-attention units. Two Uformer and Restormer models were previously introduced in detail in the image deblurring section.

4.7. General Image Enhancement

Image restoration is a technique used to restore missing information in images and has been applied to various tasks such as super-resolution, denoising, deblurring, dehazing, JPEG compression artifact reduction, and the removal of adverse weather conditions. However, some research involving image datasets may not fit neatly into any specific image restoration task. These datasets may take the form of images captured of different qualities or images that have undergone significant degradation due to natural or artificial factors, as shown in Figure 10. The datasets that need improvement have two forms. The first form is images taken twice, once with high quality and once with low quality. The second form is images suffering from a lot of image degradation, including noise, downsample, blur, hazing, and adverse weather conditions.
Deng et al. [137] have created a clinical dataset (real fundus). This dataset consists of 120 real pairs of low-resolution and high-resolution images. They also proposed a new transformer-based GAN called RFormer. RFormer is a true degradation restoration model for fundus images. RFormer consists of a generator and a discriminator. The generator is based on the U-Net architecture and is mainly based on the Window-based Self-Attention Block (WSAB) but has been replaced by the Multilayer Perceptual (MLP) with the Feed-Forward Network (FFN). The WSAB generates more realistic images that capture long-range dependencies and non-local self-similarities. The discriminator model relies on a transformer to distinguish between real and fake images to monitor the quality of the generator. The discriminator is based on PatchGAN [138], which has a final layer as N × N × 1.
Boudiaf et al. [139] have employed an Image Processing Transformer (IPT) model [59] to address the problem of noise and distortion in underwater images. The IPT model was trained on a large number of images for tasks such as super-resolution and noise removal, and it has shown promising results in this context. In a separate study, the authors in [140] focused on restoring and improving degraded facial images, due to their significance in various applications. They proposed the FRGAN model, which can restore facial features and enhance accuracy. The FRGAN comprises three stages: a head position estimation network, a Postural Transformer Network, and a Face Generative Adversarial Network. Additionally, a new loss function called Swish-X was also proposed for use in the FRGAN model. Souibgui et al. [141] presented a flexible document image enhancement model called DocEnTr, based on transformers. DocEnTr is the first encoder-decoder-based image document restoration model that uses transformers without any CNNs. The encoder extracts the most salient features from the positional pixels and converts them into encoded patches, without the use of convolution layers, while the decoder reconstructs the images from the encoded patches. Table 15 presents the results of various ViT models in enhancing the quality of images. It is important to note the diversity of the datasets used in these evaluations, which include scenarios that do not require mathematical degradation models.
Puyang et al. [142] have created two datasets for blind face restoration, known as BFR128 and BFR512. They also present a comprehensive model for facial image degradation, which includes noise, down-sampling, blur, and JPEG compression artifacts, and combine them in a complete degradation scenario. In addition, they propose the U-net model and replace the down-sampling and up-sampling layers with swin-transformer blocks. In [143], they propose the Sinogram Restoration Transformer (SRT) model, which is inspired by the swin transformer and long-range dependency modeling. They also introduce the Residual Image Reconstruction Module (RIRM) for restoring X-ray sinograms and a differentiable DuDo Consistency Layer to preserve consistency in the sinograms, leading to the creation of high-quality Computed Tomography (CT) images. In [144], the authors propose the U2-Former transformer based on a nested U-shaped architecture to facilitate feature transfer between different layers. The model utilizes two overlapping or nested structures: an inner U-shape built on the self-attention block and an external U-shape to create a deep encoding and decoding space. This model is complex, but its strength lies in its large transformer depth, leading to high efficiency in image restoration tasks such as reflection removal, rain streak removal, and dehazing.
Yan et al. [145] propose a medical MRI model called SMIR, which combines the advantages of residual contact with those of the transformer. SMIR consists of a feature extraction unit and a reconstruction unit, with the Swin transformer used as the feature extractor. The proposed model is applied to different ratio radial filling trajectory samples. The authors in [146] proposed a U-net-based transformer to combine the advantages of convolutional layers and transformers. They used several convolution layers in the encoder, the encoder vector, the transformer encoder block, and the same structure in the decoder. In addition, they used self-learning and contrastive loss.

5. Discussion and Future Work

In this section, we present the advantages and limitations of ViT in image restoration as well as the new possible research directions arising from these limitations. The current study has analyzed the use of transformer models, specifically the ViT, in the field of image restoration. Our review of the literature found that ViT has been widely applied to various tasks within image restoration, including Super-Resolution, Denoising, General Image Enhancement, JPEG Compression Artifact Reduction, Image Deblurring, Removing Adverse Weather Conditions and Image Dehazing. It has generally achieved strong results compared to traditional methods and other deep learning models such as CNNs and GANs.
One of the key advantages of using ViT in image restoration is its ability to extract global features and its stability during training. This allows ViT to effectively model complex relationships within the data and generalize well to new examples. In addition, ViT has been shown to be flexible and perform well on various types of data, including generic images, medical images, and remote sensing data. However, there are also some limitations to using ViT in image restoration. One major issue is the high computational cost of training ViT models, which can make them impractical for some applications. However, ViT performs better than CNN models when trained on large datasets. The performance of different versions of ViTs in the task of image classification shows that its performance increases while increasing the number of training images compared to the CNN-based ResNet (BiT) model [15]. In contrast, ViT suffers from slow training (large training time) due to the self-attention process, which requires higher computational costs than CNN models [15]. In addition, ViT is less effective at extracting local features and may struggle to capture close-range dependencies within the data. This can be a drawback in tasks requiring detailed, fine-grained image analysis. Figure 11 summarizes the advantages and drawbacks of ViT in image restoration.
Despite these limitations, our review suggests that ViT has strong potential for use in image restoration and has already demonstrated impressive results in a variety of tasks. However, further research is needed to address the challenges of using ViT in image restoration, such as reducing the computational cost and improving its ability to extract local features. This could involve developing more efficient self-attention mechanisms, adapting the model to the specific characteristics of image restoration tasks, or exploring hybrid approaches that combine both ViT and other deep learning models.
Several directions for future research could help to further advance the use of ViT in image restoration. One potential avenue is to explore ways to reduce the computational cost of training ViT models, such as using more efficient self-attention mechanisms or adapting the model to the specific characteristics of image restoration tasks. Another area for investigation is the development of more interpretable ViT models, which could help better understand and explain the model’s decisions during the restoration process. This could involve incorporating additional interpretability mechanisms into the model architecture or designing more transparent loss functions. In addition, further research could focus on developing hybrid approaches that combine the strengths of ViT with other deep learning models, such as CNNs. This could allow for the use of the best features of both models in a single image restoration pipeline, potentially resulting in improved performance.

6. Conclusions

In conclusion, ViT models have gained widespread attention in computer vision. They have demonstrated superior performance to traditional models such as CNNs in various tasks, including image classification, object detection, action recognition, segmentation, and image restoration. This survey has explicitly focused on the use of ViT models, in the field of image restoration. We have reviewed a range of approaches based on vision transformers, including image super-resolution, denoising, enhancement, JPEG compression artifact reduction, deblurring, removal of adverse weather conditions, and dehazing. We have also presented a comparative study of the state-of-the-art vision transformer models in each of these tasks and have discussed the main strengths and drawbacks of their adoption in image restoration. Overall, our review suggests that vision transformers offer a promising approach to addressing the challenges of image restoration and have demonstrated strong performance in many tasks. However, there are also limitations to their use, including a high computational cost and difficulty in extracting local features. To further advance the use of vision transformers in image restoration, it will be important for future research to address these limitations and explore ways to improve the model’s performance and efficiency.

Author Contributions

Conceptualization, A.M.A., B.B. and A.K.; methodology, A.M.A., B.B. and A.K.; investigation, A.M.A. and B.B.; writing—original draft preparation, A.M.A. and, B.B.; writing—review and editing, A.M.A., B.B., A.K., W.E.-S., Z.K. and W.B.; supervision, B.B., A.K. and W.E.-S.; project administration, B.B. and A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Prince Sultan University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to acknowledge the support of Prince Sultan University for paying the Article Processing Charges (APC) of this publication.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  2. El-Shafai, W.; Ali, A.M.; El-Rabaie, E.S.M.; Soliman, N.F.; Algarni, A.D.; Abd El-Samie, F.E. Automated COVID-19 Detection Based on Single-Image Super-Resolution and CNN Models. Comput. Mater. Contin. 2021, 70, 1141–1157. [Google Scholar] [CrossRef]
  3. El-Shafai, W.; El-Nabi, S.A.; El-Rabaie, E.S.M.; Ali, A.M.; Soliman, N.F.; Algarni, A.D.; Abd El-Samie, F.E. Efficient Deep-Learning-Based Autoencoder Denoising Approach for Medical Image Diagnosis. Comput. Mater. Contin. 2021, 70, 6107–6125. [Google Scholar] [CrossRef]
  4. El-Shafai, W.; Mohamed, E.M.; Zeghid, M.; Ali, A.M.; Aly, M.H. Hybrid Single Image Super-Resolution Algorithm for Medical Images. Comput. Mater. Contin. 2022, 72, 4879–4896. [Google Scholar] [CrossRef]
  5. Lu, D.; Weng, Q. A Survey of Image Classification Methods and Techniques for Improving Classification Performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  6. Noor, A.; Benjdira, B.; Ammar, A.; Koubaa, A. DriftNet: Aggressive Driving Behaviour Detection Using 3D Convolutional Neural Networks. In Proceedings of the 1st International Conference of Smart Systems and Emerging Technologies, SMART-TECH 2020, Riyadh, Saudi Arabia, 3–5 November 2020; pp. 214–219. [Google Scholar] [CrossRef]
  7. Varone, G.; Boulila, W.; lo Giudice, M.; Benjdira, B.; Mammone, N.; Ieracitano, C.; Dashtipour, K.; Neri, S.; Gasparini, S.; Morabito, F.C.; et al. A Machine Learning Approach Involving Functional Connectivity Features to Classify Rest-EEG Psychogenic Non-Epileptic Seizures from Healthy Controls. Sensors 2022, 22, 129. [Google Scholar] [CrossRef]
  8. Benjdira, B.; Koubaa, A.; Boulila, W.; Ammar, A. Parking Analytics Framework Using Deep Learning. In Proceedings of the 2022 2nd International Conference of Smart Systems and Emerging Technologies, SMARTTECH 2022, Riyadh, Saudi Arabia, 9–11 May 2022; pp. 200–205. [Google Scholar] [CrossRef]
  9. Benjdira, B.; Koubaa, A.; Azar, A.T.; Khan, Z.; Ammar, A.; Boulila, W. TAU: A Framework for Video-Based Traffic Analytics Leveraging Artificial Intelligence and Unmanned Aerial Systems. Eng. Appl. Artif. Intell. 2022, 114, 105095. [Google Scholar] [CrossRef]
  10. Benjdira, B.; Ouni, K.; al Rahhal, M.M.; Albakr, A.; Al-Habib, A.; Mahrous, E. Spinal Cord Segmentation in Ultrasound Medical Imagery. Appl. Sci. 2020, 10, 1370. [Google Scholar] [CrossRef] [Green Version]
  11. Benjdira, B.; Ammar, A.; Koubaa, A.; Ouni, K. Data-Efficient Domain Adaptation for Semantic Segmentation of Aerial Imagery Using Generative Adversarial Networks. Appl. Sci. 2020, 10, 1092. [Google Scholar] [CrossRef] [Green Version]
  12. Khan, A.R.; Saba, T.; Khan, M.Z.; Fati, S.M.; Khan, M.U.G. Classification of Human’s Activities from Gesture Recognition in Live Videos Using Deep Learning. Concurr. Comput. 2022, 34, e6825. [Google Scholar] [CrossRef]
  13. Ubaid, M.T.; Saba, T.; Draz, H.U.; Rehman, A.; Ghani, M.U.; Kolivand, H. Intelligent Traffic Signal Automation Based on Computer Vision Techniques Using Deep Learning. IT Prof. 2022, 24, 27–33. [Google Scholar] [CrossRef]
  14. Delia-Alexandrina, M.; Nedevschi, S.; Fati, S.M.; Senan, E.M.; Azar, A.T. Hybrid and Deep Learning Approach for Early Diagnosis of Lower Gastrointestinal Diseases. Sensors 2022, 22, 4079. [Google Scholar] [CrossRef]
  15. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020. [Google Scholar] [CrossRef]
  16. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  17. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  18. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M.; Khan, S.; Naseer, M.; City, M.; Dhabi, A.; et al. Transformers in Vision: A Survey. ACM Comput. Surv. 2022, 54, 200. [Google Scholar] [CrossRef]
  19. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Vision Transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef]
  20. Islam, K. Recent Advances in Vision Transformer: A Survey and Outlook of Recent Work. arXiv 2022. [Google Scholar] [CrossRef]
  21. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in Medical Imaging: A Survey. arXiv 2022. [Google Scholar] [CrossRef]
  22. Su, J.; Xu, B.; Yin, H. A Survey of Deep Learning Approaches to Image Restoration. Neurocomputing 2022, 487, 46–65. [Google Scholar] [CrossRef]
  23. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V.; et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019. [Google Scholar] [CrossRef]
  24. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL HLT 2019–2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; Volume 1, pp. 4171–4186. [Google Scholar] [CrossRef]
  25. Vaswani, A.; Brain, G.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  26. Pérez, J.; Marinković, J.; Barceló, P. On the Turing Completeness of Modern Neural Network Architectures. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar] [CrossRef]
  27. Cordonnier, J.-B.; Loukas, A.; Jaggí, M.J. On the Relationship between Self-Attention and Convolutional Layers. arXiv 2019. [Google Scholar] [CrossRef]
  28. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  29. Li, X.; Yin, X.; Li, C.; Zhang, P.; Hu, X.; Zhang, L.; Wang, L.; Hu, H.; Dong, L.; Wei, F.; et al. Oscar: Object-Semantics Aligned Pre-Training for Vision-Language Tasks. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 121–137. [Google Scholar] [CrossRef]
  30. Su, W.; Zhu, X.; Cao, Y.; Li, B.; Lu, L.; Wei, F.; Dai, J. VL-BERT: Pre-Training of Generic Visual-Linguistic Representations. arXiv 2019. [Google Scholar] [CrossRef]
  31. Fedus, W.; Zoph, B.; Shazeer, N. Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity. J. Mach. Learn. Res. 2021, 23, 1–40. [Google Scholar] [CrossRef]
  32. Jing, L.; Tian, Y. Self-Supervised Visual Feature Learning with Deep Neural Networks: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4037–4058. [Google Scholar] [CrossRef]
  33. Liu, X.; Zhang, F.; Hou, Z.; Mian, L.; Wang, Z.; Zhang, J.; Tang, J. Self-Supervised Learning: Generative or Contrastive. IEEE Trans. Knowl. Data Eng. 2021, 35, 857–876. [Google Scholar] [CrossRef]
  34. Thung, K.H.; Raveendran, P. A Survey of Image Quality Measures. In Proceedings of the International Conference for Technical Postgraduates 2009, TECHPOS 2009, Kuala Lumpur, Malaysia, 14–15 December 2009. [Google Scholar] [CrossRef]
  35. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind Image Quality Assessment: A Natural Scene Statistics Approach in the DCT Domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef]
  36. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs SSIM. In Proceedings of the International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  37. Almohammad, A.; Ghinea, G. Stego Image Quality and the Reliability of PSNR. In Proceedings of the 2010 2nd International Conference on Image Processing Theory, Tools and Applications, IPTA 2010, Paris, France, 7–10 July 2010; pp. 215–220. [Google Scholar] [CrossRef]
  38. Rouse, D.M.; Hemami, S.S. Analyzing the Role of Visual Structure in the Recognition of Natural Image Content with Multi-Scale SSIM. In Human Vision and Electronic Imaging XIII; SPIE: Bellingham, WA, USA, 2008; Volume 6806, pp. 410–423. [Google Scholar] [CrossRef]
  39. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  40. Sara, U.; Akter, M.; Uddin, M.S.; Sara, U.; Akter, M.; Uddin, M.S. Image Quality Assessment through FSIM, SSIM, MSE and PSNR—A Comparative Study. J. Comput. Commun. 2019, 7, 8–18. [Google Scholar] [CrossRef] [Green Version]
  41. Pambrun, J.F.; Noumeir, R. Limitations of the SSIM Quality Metric in the Context of Diagnostic Imaging. In Proceedings of the International Conference on Image Processing, ICIP 2015, Quebec City, QC, Canada, 27–30 September 2015; pp. 2960–2963. [Google Scholar] [CrossRef]
  42. Marathe, A.; Jain, P.; Walambe, R.; Kotecha, K. RestoreX-AI: A Contrastive Approach Towards Guiding Image Restoration via Explainable AI Systems. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 3030–3039. [Google Scholar]
  43. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H. Restormer: Efficient Transformer for High-Resolution Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  44. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision 2017, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  45. Cheon, M.; Yoon, S.-J.; Kang, B.; Lee, J. Perceptual Image Quality Assessment with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA, 20–25 June 2021; pp. 433–442. [Google Scholar]
  46. Conde, M.V.; Burchi, M.; Timofte, R. Conformer and Blind Noisy Students for Improved Image Quality Assessment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 940–950. [Google Scholar]
  47. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Available online: https://dl.acm.org/doi/10.5555/3298023.3298188 (accessed on 30 October 2022).
  48. Park, S.C.; Park, M.K.; Kang, M.G. Super-Resolution Image Reconstruction: A Technical Overview. IEEE Signal Process. Mag. 2003, 20, 21–36. [Google Scholar] [CrossRef] [Green Version]
  49. Chen, X.; Wang, X.; Zhou, J.; Dong, C. Activating More Pixels in Image Super-Resolution Transformer. arXiv 2022. [Google Scholar] [CrossRef]
  50. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; van Gool, L.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2021, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  51. Zhang, D.; Huang, F.; Liu, S.; Wang, X.; Jin, Z. SwinFIR: Revisiting the SwinIR with Fast Fourier Convolution and Improved Training for Image Super-Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  52. Yang, C.Y.; Ma, C.; Yang, M.H. Single-Image Super-Resolution: A Benchmark. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; pp. 372–386. [Google Scholar] [CrossRef]
  53. Yang, C.-Y.; Yang, M.-H. Fast Direct Super-Resolution by Simple Functions. In Proceedings of the IEEE International Conference on Computer Vision 2013, Sydney, Australia, 2–8 December 2013; pp. 561–568. [Google Scholar]
  54. Shan, Q.; Li, Z.; Jia, J.; Tang, C.K. Fast Image/Video Upsampling. ACM Trans. Graph. 2008, 27, 153. [Google Scholar] [CrossRef] [Green Version]
  55. Sun, J.; Sun, J.; Xu, Z.; Shum, H.Y. Image Super-Resolution Using Gradient Profile Prior. In Proceedings of the 26th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Anchorage, Alaska, 23–28 June 2008. [Google Scholar] [CrossRef]
  56. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2021, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  57. Tu, J.; Mei, G.; Ma, Z.; Piccialli, F. SWCGAN: Generative Adversarial Network Combining Swin Transformer and CNN for Remote Sensing Image Super-Resolution. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 5662–5673. [Google Scholar] [CrossRef]
  58. Ma, Q.; Jiang, J.; Liu, X.; Ma, J. Learning A 3D-CNN and Transformer Prior for Hyperspectral Image Super-Resolution. arXiv 2021. [Google Scholar] [CrossRef]
  59. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-Trained Image Processing Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA, 20–25 June 2021; pp. 12299–12310. [Google Scholar]
  60. He, D.; Wu, S.; Liu, J.; Xiao, G. Cross Transformer Network for Scale-Arbitrary Image Super-Resolution Lecture Notes in Computer Science. In Proceedings of the Knowledge Science, Engineering and Management: 15th International Conference, KSEM 2022, Singapore, 6–8 August 2022; pp. 633–644. [Google Scholar] [CrossRef]
  61. Liu, H.; Shao, M.; Wang, C.; Cao, F. Image Super-Resolution Using a Simple Transformer Without Pretraining. Neural Process. Lett. 2022, 1–19. [Google Scholar] [CrossRef]
  62. Zhang, X.; Zeng, H.; Guo, S.; Zhang, L. Efficient Long-Range Attention Network for Image Super-Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  63. Cai, Q.; Qian, Y.; Li, J.; Lv, J.; Yang, Y.-H.; Wu, F.; Zhang, D. HIPA: Hierarchical Patch Transformer for Single Image Super Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  64. Yoo, J.; Kim, T.; Lee, S.; Kim, S.H.; Lee, H.; Kim, T.H. Enriched CNN-Transformer Feature Aggregation Networks for Super-Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  65. Wang, S.; Zhou, T.; Lu, Y.; Di, H. Detail-Preserving Transformer for Light Field Image Super-Resolution. In Proceedings of the AAAI Conference on Artificial Intelligence 2022, Virtual, 22 February–1 March 2022; Volume 36, pp. 2522–2530. [Google Scholar] [CrossRef]
  66. Liang, Z.; Wang, Y.; Wang, L.; Yang, J.; Zhou, S. Light Field Image Super-Resolution with Transformers. IEEE Signal Process. Lett. 2022, 29, 563–567. [Google Scholar] [CrossRef]
  67. Lei, S.; Shi, Z.; Mo, W. Transformer-Based Multistage Enhancement for Remote Sensing Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 3136190. [Google Scholar] [CrossRef]
  68. Wang, L.; Zhu, H.; He, Z.; Jia, Y.; Du, J. Adjacent Slices Feature Transformer Network for Single Anisotropic 3D Brain MRI Image Super-Resolution. Biomed. Signal Process. Control. 2022, 72, 103339. [Google Scholar] [CrossRef]
  69. Zhang, W.; Wang, L.; Chen, W.; Jia, Y.; He, Z.; Du, J. 3D Cross-Scale Feature Transformer Network for Brain mr Image Super-Resolution. In Proceedings of the ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing, Singapore, 23–27 May 2022; pp. 1356–1360. [Google Scholar] [CrossRef]
  70. Fang, C.; Zhang, D.; Wang, L.; Zhang, Y.; Cheng, L.; Lab, Z.; Han, J. Cross-Modality High-Frequency Transformer for MR Image Super-Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  71. Hu, J.F.; Huang, T.Z.; Deng, L.J.; Dou, H.X.; Hong, D.; Vivone, G. Fusformer: A Transformer-Based Fusion Network for Hyperspectral Image Super-Resolution. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6012305. [Google Scholar] [CrossRef]
  72. Liu, Y.; Hu, J.; Kang, X.; Luo, J.; Fan, S. Interactformer: Interactive Transformer and CNN for Hyperspectral Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531715. [Google Scholar] [CrossRef]
  73. Yang, F.; Yang, H.; Fu, J.; Lu, H.; Guo, B. Learning Texture Transformer Network for Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2020, Seattle, WA, USA, 13–19 June 2020; pp. 5791–5800. [Google Scholar]
  74. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar] [CrossRef]
  75. Gao, G.; Xu, Z.; Li, J.; Yang, J.; Zeng, T.; Qi, G.-J. CTCNet: A CNN-Transformer Cooperation Network for Face Image Super-Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  76. Yan, B.; Cao, L.; Qi, F.; Wang, H. Bilateral Network with Channel Splitting Network and Transformer for Thermal Image Super-Resolution. arXiv 2022. [Google Scholar] [CrossRef]
  77. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible Solution for CNN-Based Image Denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [Green Version]
  78. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  79. Thakur, R.S.; Chatterjee, S.; Yadav, R.N.; Gupta, L. Image De-Noising with Machine Learning: A Review. IEEE Access 2021, 9, 93338–93363. [Google Scholar] [CrossRef]
  80. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep Learning on Image Denoising: An Overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef]
  81. Prayuda, A.W.H.; Prasetyo, H.; Guo, J.M. AWGN-Based Image Denoiser Using Convolutional Vision Transformer. In Proceedings of the 2021 International Symposium on Electronics and Smart Devices: Intelligent Systems for Present and Future Challenges, ISESD 2021, Bandung, Indonesia, 29–30 June 2021. [Google Scholar] [CrossRef]
  82. Liu, X.; Hong, Y.; Yin, Q.; Zhang, S. DnT: Learning Unsupervised Denoising Transformer from Single Noisy Image. In Proceedings of the 4th International Conference on Image Processing and Machine Vision, Hong Kong, China, 25–27 March 2022; pp. 50–56. [Google Scholar] [CrossRef]
  83. Pang, T.; Zheng, H.; Quan, Y.; Ji, H. Recorrupted-to-Recorrupted: Unsupervised Deep Learning for Image Denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA, 20–25 June 2021; pp. 2043–2052. [Google Scholar]
  84. Zhao, M.; Cao, G.; Huang, X.; Yang, L. Hybrid Transformer-CNN for Real Image Denoising. IEEE Signal Process. Lett. 2022, 29, 1252–1256. [Google Scholar] [CrossRef]
  85. Xue, T.; Ma, P. TC-Net: Transformer Combined with Cnn for Image Denoising. Appl. Intell. 2022, 1–10. [Google Scholar] [CrossRef]
  86. Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A General U-Shaped Transformer for Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 17683–17693. [Google Scholar]
  87. Fan, C.-M.; Liu, T.-J.; Liu, K.-H. SUNet: Swin Transformer UNet for Image Denoising. arXiv 2022. [Google Scholar] [CrossRef]
  88. Li, Z.; Jiang, H.; Zheng, Y. Polarized Color Image Denoising Using Pocoformer. arXiv 2022. [Google Scholar] [CrossRef]
  89. Zhang, Z.; Yu, L.; Liang, X.; Zhao, W.; Xing, L. TransCT: Dual-Path Transformer for Low Dose Computed Tomography. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; pp. 55–64. [Google Scholar] [CrossRef]
  90. Wang, D.; Wu, Z.; Yu, H. TED-Net: Convolution-Free T2T Vision Transformer-Based Encoder-Decoder Dilation Network for Low-Dose CT Denoising. In Proceedings of the Machine Learning in Medical Imaging: 12th International Workshop, MLMI 2021, Held in Conjunction with MICCAI 2021, Strasbourg, France, 27 September 2021; pp. 416–425. [Google Scholar] [CrossRef]
  91. Luthra, A.; Sulakhe, H.; Mittal, T.; Iyer, A.; Yadav, S. Eformer: Edge Enhancement Based Transformer for Medical Image Denoising. arXiv 2021. [Google Scholar] [CrossRef]
  92. Gupta, A.; Joshi, N.; Lawrence Zitnick, C.; Cohen, M.; Curless, B. Single Image Deblurring Using Motion Density Functions. In Proceedings of the Computer Vision–ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; pp. 171–184. [Google Scholar] [CrossRef]
  93. Wang, R.; Tao, D. Recent Progress in Image Deblurring. arXiv 2014. [Google Scholar] [CrossRef]
  94. Lai, W.-S.; Huang, J.-B.; Hu, Z.; Ahuja, N.; Yang, M.-H. A Comparative Study for Single Image Blind Deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1701–1709. [Google Scholar]
  95. Zhang, K.; Ren, W.; Luo, W.; Lai, W.S.; Stenger, B.; Yang, M.H.; Li, H. Deep Image Deblurring: A Survey. Int. J. Comput. Vis. 2022, 130, 2103–2130. [Google Scholar] [CrossRef]
  96. Zheng, Z.; Jia, X. UHD Image Deblurring via Multi-Scale Cubic-Mixer. arXiv 2022. [Google Scholar] [CrossRef]
  97. Tsai, F.-J.; Peng, Y.-T.; Lin, Y.-Y.; Tsai, C.-C.; Lin, C.-W. Stripformer: Strip Transformer for Fast Image Deblurring. arXiv 2022. [Google Scholar] [CrossRef]
  98. Sun, J.; Cao, W.; Xu, Z.; Ponce, J. Learning a Convolutional Neural Network for Non-Uniform Motion Blur Removal. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 769–777. [Google Scholar]
  99. Guo, C.-L.; Yan, Q.; Anwar, S.; Cong, R.; Ren, W.; Li, C. Image Dehazing Transformer with Transmission-Aware 3D Position Embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 5812–5820. [Google Scholar]
  100. Gao, G.; Cao, J.; Bao, C.; Hao, Q.; Ma, A.; Li, G. A Novel Transformer-Based Attention Network for Image Dehazing. Sensors 2022, 22, 3428. [Google Scholar] [CrossRef] [PubMed]
  101. Li, X.; Hua, Z.; Li, J. Two-Stage Single Image Dehazing Network Using Swin-Transformer. IET Image Process 2022, 16, 2518–2534. [Google Scholar] [CrossRef]
  102. Jiao, Q.; Liu, M.; Ning, B.; Zhao, F.; Dong, L.; Kong, L.; Hui, M.; Zhao, Y. Image Dehazing Based on Local and Non-Local Features. Fractal Fract. 2022, 6, 262. [Google Scholar] [CrossRef]
  103. Song, Y.; He, Z.; Qian, H.; Du, X. Vision Transformers for Single Image Dehazing. arXiv 2022. [Google Scholar] [CrossRef]
  104. Zhao, D.; Li, J.; Member, S.; Li, H.; Xu, L. Complementary Feature Enhanced Network with Vision Transformer for Image Dehazing. arXiv 2021. [Google Scholar] [CrossRef]
  105. Dong, C.; Deng, Y.; Loy, C.C.; Tang, X. Compression Artifacts Reduction by a Deep Convolutional Network. In Proceedings of the IEEE International Conference on Computer Vision 2015, Santiago, Chile, 7–13 December 2015; pp. 576–584. [Google Scholar]
  106. Liu, X.; Cheung, G.; Ji, X.; Zhao, D.; Gao, W. Graph-Based Joint Dequantization and Contrast Enhancement of Poorly Lit JPEG Images. IEEE Trans. Image Process. 2019, 28, 1205–1219. [Google Scholar] [CrossRef]
  107. Foi, A.; Katkovnik, V.; Egiazarian, K. Pointwise Shape-Adaptive DCT for High-Quality Denoising and Deblocking of Grayscale and Color Images. IEEE Trans. Image Process. 2007, 16, 1395–1411. [Google Scholar] [CrossRef]
  108. Chen, T.; Wu, H.R.; Qiu, B. Adaptive Postifiltering of Transform Coefficients for the Reduction of Blocking Artifacts. IEEE Trans. Circuits Syst. Video Technol. 2001, 11, 594–602. [Google Scholar] [CrossRef]
  109. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  110. Jiang, J.; Zhang, K.; Timofte, R. Towards Flexible Blind JPEG Artifacts Removal. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2021, Montreal, BC, Canada, 11–17 October 2021; pp. 4997–5006. [Google Scholar]
  111. Fu, X.; Zha, Z.-J.; Wu, F.; Ding, X.; Paisley, J. JPEG Artifacts Reduction via Deep Convolutional Sparse Coding. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2019, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2501–2510. [Google Scholar]
  112. Jiang, X.; Tan, W.; Cheng, R.; Zhou, S.; Yan, B. Learning Parallax Transformer Network for Stereo Image JPEG Artifacts Removal. In Proceedings of the 30th ACM International Conference on Multimedia, Lisbon, Portugal, 10 October 2022; pp. 6072–6082. [Google Scholar] [CrossRef]
  113. Charbonnier, P.; Blanc-Féraud, L.; Aubert, G.; Barlaud, M. Two Deterministic Half-Quadratic Regularization Algorithms for Computed Imaging. In Proceedings of the International Conference on Image Processing, ICIP 1994, Austin, TX, USA, 13–16 November 1994; Volume 2, pp. 168–172. [Google Scholar] [CrossRef]
  114. Wang, L.; Zhang, J.; Wang, O.; Lin, Z.; Lu, H. SDC-Depth: Semantic Divide-and-Conquer Network for Monocular Depth Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2020, Seattle, WA, USA, 13–19 June 2020; pp. 541–550. [Google Scholar]
  115. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision 2018, Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  116. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar] [CrossRef]
  117. Perera, A.G.; Wei Law, Y.; Chahl, J. UAV-GESTURE: A Dataset for UAV Control and Gesture Recognition. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops 2018, Munich, Germany, 8–14 September 2018. [Google Scholar]
  118. Liang, M.; Yang, B.; Wang, S.; Urtasun, R. Deep Continuous Fusion for Multi-Sensor 3D Object Detection. In Proceedings of the European Conference on Computer Vision 2018, Munich, Germany, 8–14 September 2018; pp. 641–656. [Google Scholar]
  119. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum PointNets for 3D Object Detection From RGB-D Data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 918–927. [Google Scholar]
  120. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear Total Variation Based Noise Removal Algorithms. Phys. D 1992, 60, 259–268. [Google Scholar] [CrossRef]
  121. Roth, S.; Black, M.J. Fields of Experts: A Framework for Learning Image Priors. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2005, San Diego, CA, USA, 20–25 June 2005; Volume II, pp. 860–867. [Google Scholar] [CrossRef]
  122. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [CrossRef]
  123. Zhang, H.; Patel, V.M. Density-Aware Single Image De-Raining Using a Multi-Stream Dense Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 695–704. [Google Scholar]
  124. Wei, W.; Meng, D.; Zhao, Q.; Xu, Z.; Wu, Y. Semi-Supervised Transfer Learning for Image Rain Removal. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 15–20 June 2019; pp. 3877–3886. [Google Scholar]
  125. Qian, R.; Tan, R.T.; Yang, W.; Su, J.; Liu, J. Attentive Generative Adversarial Network for Raindrop Removal from a Single Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2482–2491. [Google Scholar]
  126. Fu, X.; Huang, J.; Zeng, D.; Huang, Y.; Ding, X.; Paisley, J. Removing Rain from Single Images via a Deep Detail Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 3855–3863. [Google Scholar]
  127. Zhang, K.; Li, R.; Yu, Y.; Luo, W.; Li, C. Deep Dense Multi-Scale Network for Snow Removal Using Semantic and Depth Priors. IEEE Trans. Image Process. 2021, 30, 7419–7431. [Google Scholar] [CrossRef]
  128. Ren, W.; Tian, J.; Han, Z.; Chan, A.; Tang, Y. Video Desnowing and Deraining Based on Matrix Decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 4210–4219. [Google Scholar]
  129. Liu, Y.F.; Jaw, D.W.; Huang, S.C.; Hwang, J.N. DesnowNet: Context-Aware Deep Network for Snow Removal. IEEE Trans. Image Process. 2018, 27, 3064–3073. [Google Scholar] [CrossRef] [Green Version]
  130. You, S.; Tan, R.T.; Kawakami, R.; Mukaigawa, Y.; Ikeuchi, K. Adherent Raindrop Modeling, Detectionand Removal in Video. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1721–1733. [Google Scholar] [CrossRef]
  131. Quan, Y.; Deng, S.; Chen, Y.; Ji, H. Deep Learning for Seeing Through Window with Raindrops. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2019, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2463–2471. [Google Scholar]
  132. Qin, Q.; Yan, J.; Wang, X.; Wang, Q.; Li, M.; Wang, Y. ETDNet: An Efficient Transformer Deraining Model. IEEE Access 2021, 9, 119881–119893. [Google Scholar] [CrossRef]
  133. Tan, F.; Kong, Y.; Fan, Y.; Liu, F.; Zhou, D.; Zhang, H.; Chen, L.; Gao, L.; Qian, Y. SDNet: Mutil-Branch for Single Image Deraining Using Swin. arXiv 2021. [Google Scholar] [CrossRef]
  134. Liu, L.; Xie, L.; Zhang, X.; Yuan, S.; Chen, X.; Zhou, W.; Li, H.; Tian, Q. TAPE: Task-Agnostic Prior Embedding for Image Restoration. arXiv 2022. [Google Scholar] [CrossRef]
  135. Maria, J.; Valanarasu, J.; Yasarla, R.; Patel, V.M. TransWeather: Transformer-Based Restoration of Images Degraded by Adverse Weather Conditions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 2353–2363. [Google Scholar]
  136. Liu, L.; Yuan, S.; Liu, J.; Guo, X.; Yan, Y.; Tian, Q. SiamTrans: Zero-Shot Multi-Frame Image Restoration with Pre-Trained Siamese Transformers. In Proceedings of the AAAI Conference on Artificial Intelligence 2022, Virtual, 22 February–1 March 2022; Volume 36, pp. 1747–1755. [Google Scholar] [CrossRef]
  137. Deng, Z.; Cai, Y.; Chen, L.; Gong, Z.; Bao, Q.; Yao, X.; Fang, D.; Yang, W.; Zhang, S.; Ma, L. RFormer: Transformer-Based Generative Adversarial Network for Real Fundus Image Restoration on a New Clinical Benchmark. IEEE J. Biomed. Health Inform. 2022, 26, 4645–4655. [Google Scholar] [CrossRef]
  138. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A.; Research, B.A. Image-To-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  139. Boudiaf, A.; Guo, Y.; Ghimire, A.; Werghi, N.; de Masi, G.; Javed, S.; Dias, J. Underwater Image Enhancement Using Pre-Trained Transformer. In Proceedings of the Image Analysis and Processing–ICIAP 2022: 21st International Conference, Lecce, Italy, 23–27 May 2022; pp. 480–488. [Google Scholar] [CrossRef]
  140. Wei, T.; Li, Q.; Chen, Z.; Liu, J. FRGAN: A Blind Face Restoration with Generative Adversarial Networks. Math. Probl. Eng. 2021, 2021, 2384435. [Google Scholar] [CrossRef]
  141. Souibgui, M.A.; Biswas, S.; Jemni, S.K.; Kessentini, Y.; Fornés, A.; Lladós, J.; Pal, U. DocEnTr: An End-to-End Document Image Enhancement Transformer. arXiv 2022. [Google Scholar] [CrossRef]
  142. Zhang, P.; Zhang, K.; Luo, W.; Li, C.; Wang, G. Blind Face Restoration: Benchmark Datasets and a Baseline Model. arXiv 2022. [Google Scholar] [CrossRef]
  143. Wang, C.; Shang, K.; Zhang, H.; Li, Q.; Hui, Y.; Zhou, S.K. DuDoTrans: Dual-Domain Transformer Provides More Attention for Sinogram Restoration in Sparse-View CT Reconstruction. arXiv 2021. [Google Scholar] [CrossRef]
  144. Ji, H.; Feng, X.; Pei, W.; Li, J.; Lu, G. U2-Former: A Nested U-Shaped Transformer for Image Restoration. arXiv 2021. [Google Scholar] [CrossRef]
  145. Yan, C.; Shi, G.; Wu, Z. SMIR: A Transformer-Based Model for MRI Super-Resolution Reconstruction. In Proceedings of the 2021 IEEE International Conference on Medical Imaging Physics and Engineering, ICMIPE 2021, Hefei, China, 13–14 November 2021. [Google Scholar] [CrossRef]
  146. Wang, H.; Jiang, K. Research on Image Super-Resolution Reconstruction Based on Transformer. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence and Industrial Design, AIID 2021, Guangzhou, China, 28–30 May 2021; pp. 226–230. [Google Scholar] [CrossRef]
Figure 1. The main techniques used in Image Restoration.
Figure 1. The main techniques used in Image Restoration.
Sensors 23 02385 g001
Figure 2. The general form of the ViT architecture.
Figure 2. The general form of the ViT architecture.
Sensors 23 02385 g002
Figure 3. The self-attention model used in ViT.
Figure 3. The self-attention model used in ViT.
Sensors 23 02385 g003
Figure 4. The most popular image restoration techniques based on ViT.
Figure 4. The most popular image restoration techniques based on ViT.
Sensors 23 02385 g004
Figure 5. The block diagram of the best ViT-based super-resolution models.
Figure 5. The block diagram of the best ViT-based super-resolution models.
Sensors 23 02385 g005
Figure 6. The main method for removing noise from images using transformer-based approaches. The top image shows a traditional technique that focuses on restoring clear image details from a noisy image. The bottom image depicts a more efficient method that estimates the random noise present in the noisy image and then subtracts this estimated noise from the image to produce a noise-free result.
Figure 6. The main method for removing noise from images using transformer-based approaches. The top image shows a traditional technique that focuses on restoring clear image details from a noisy image. The bottom image depicts a more efficient method that estimates the random noise present in the noisy image and then subtracts this estimated noise from the image to produce a noise-free result.
Sensors 23 02385 g006
Figure 7. Some examples of blurry images processing.
Figure 7. Some examples of blurry images processing.
Sensors 23 02385 g007
Figure 8. The atmospheric dispersion model by which the haze formation process is described.
Figure 8. The atmospheric dispersion model by which the haze formation process is described.
Sensors 23 02385 g008
Figure 9. Illustration of the two common types of adverse weather removal algorithms. The top image shows a technique that removes a single type of adverse weather, such as rain or snow. The bottom image depicts a more efficient approach that is capable of removing multiple types of adverse weather at the same time.
Figure 9. Illustration of the two common types of adverse weather removal algorithms. The top image shows a technique that removes a single type of adverse weather, such as rain or snow. The bottom image depicts a more efficient approach that is capable of removing multiple types of adverse weather at the same time.
Sensors 23 02385 g009
Figure 10. Illustration for general image enhancement, which does not fall under any specific sub-task of the image restoration.
Figure 10. Illustration for general image enhancement, which does not fall under any specific sub-task of the image restoration.
Sensors 23 02385 g010
Figure 11. The main advantages and drawbacks of ViT in Image Restoration.
Figure 11. The main advantages and drawbacks of ViT in Image Restoration.
Sensors 23 02385 g011
Table 1. Comparison between our survey and the related surveys in image restoration.
Table 1. Comparison between our survey and the related surveys in image restoration.
MethodYearAlgorithmRegion of Interest
[18]2022ViTGenerative modeling, low-level vision (e.g., image super-resolution, image enhancement, and colorization), Recognition tasks (e.g., image classification, object detection, action recognition, and segmentation), video processing (e.g., activity recognition, video forecasting), multi-modal tasks (e.g., visual question answering, visual reasoning, and visual grounding), and three-dimensional analysis (e.g., point cloud classification and segmentation)
[19]2022ViTBackbone network (e.g., supervised pretraining and self-supervised pretraining), high/mid-level vision (e.g., object detection, segmentation, and pose estimation), low-level vision (e.g., image generation, and image enhancement), video processing (e.g., video inpainting, and video captioning), multimodality tasks (e.g., classification, image generation, and multi-task), and efficient transformer (e.g., decomposition, distillation, quantization, and architecture design)
[20]2022ViT, CNNFundamental concepts, a background of the self-attention mechanism, strengths and weaknesses, computational cost, and comparison of ViT and CNN performance on benchmark datasets
[21]2022ViTMedical image segmentation, detection, classification, reconstruction, synthesis, registration, and clinical report generation
[22]2022CNN, ViTGeneral deep learning techniques in super-resolution, dehazing, deblurring, and image denoising. Then, a summary of the leading architectural components involved in these tasks, such as residual or skip connection, receptive field, and unsupervised autoencoder mechanisms.
Our-ViTSeven image restoration tasks are deeply studied: image super-resolution, image denoising, general image enhancement, JPEG compression artifact reduction, image deblurring, removing adverse weather conditions, and image dehazing.
Table 2. Generic Image Super-Resolution-Based on ViT for scale ×2. Best records are emphasized in bold font.
Table 2. Generic Image Super-Resolution-Based on ViT for scale ×2. Best records are emphasized in bold font.
MethodTraining DatasetSet14
PSNR/SSIM
BSD100
PSNR/SSIM
Manga109
PSNR/SSIM
Parameters
HAT [49], 2022DF2K35.29/0.929332.74/0.906641.01/0.983120.8M
SwinFIR [51], 2022DF2K34.93/0.927632.64/0.905440.61/0.981613.99M
IPT [59], 2021ImageNet34.4332.48-46.0M
SwinIR [50], 2021DF2K34.61/0.926032.55/0.904340.02/0.980011.8M
CrossSR [60], 2022DIV2K33.99/0.921832.27/0.9000-18.3M
SRT [61], 2022DIV2K33.95/0.920732.35/0.9018-11M
ELAN [62], 2022DIV2K33.94/0.920732.30/0.901239.11/0.9782582K
HIPA [63], 2022DIV2K34.25/0.923532.48/0.904039.75/0.979419.2M
ACT [64], 2022ImageNet34.68/0.926032.60/0.905240.11/0.980746.0M
Table 3. Generic Image Super-Resolution-Based on ViT for scale ×3. Best records are emphasized in bold font.
Table 3. Generic Image Super-Resolution-Based on ViT for scale ×3. Best records are emphasized in bold font.
MethodTraining DatasetSet14
PSNR/SSIM
BSD100
PSNR/SSIM
Manga109
PSNR/SSIM
Parameters
HAT [49], 2022DF2K31.47/0.858429.63/0.819136.02/0.957620.8M
SwinFIR [51], 2022DF2K31.24/0.856629.55/0.816935.77/0.956313.99M
IPT [59], 2021ImageNet30.8529.38-46.0M
SwinIR [50], 2021DF2K31.00/0.854229.49/0.815035.28/0.954311.8M
CrossSR [60], 2022DIV2K30.43/0.843329.15/0.806333.95/0.94550.68M
SRT [61], 2022DIV2K30.53/0.846029.21/0.8082-18.3M
ELAN [62], 2022DIV2K30.55/0.846329.21/0.808134.00/0.9478590K
HIPA [63], 2022DIV2K30.38/0.841729.13/0.806133.82/0.9460365K
ACT [64], 2022DIV2K30.80/0.850429.45/0.812734.86/0.952119.2M
HAT [49], 2022ImageNet31.17/0.854929.55/0.817135.47/0.954846.0M
Table 4. Generic Image Super-Resolution-Based on ViT for scale ×4. Best records are emphasized in bold font.
Table 4. Generic Image Super-Resolution-Based on ViT for scale ×4. Best records are emphasized in bold font.
MethodTraining DatasetSet14
PSNR/SSIM
BSD100
PSNR/SSIM
Manga109
PSNR/SSIM
Parameters
HAT [49], 2022DF2K29.47/0.801528.09/0.755133.09/0.933520.8M
SwinFIR [51], 2022DF2K29.36/0.799328.03/0.752032.83/0.931413.99M
IPT [59], 2021ImageNet29.0127.82-46.0M
SwinIR [50], 2021DF2K29.15/0.795827.95/0.749432.22/0.927311.8M
CrossSR [60], 2022DIV2K28.11/0.784227.54/0.746430.09/0.9077-
SRT [61], 2022DIV2K28.69/0.783327.69/0.737930.75/0.91000.68M
ELAN [62], 2022DIV2K28.79/0.785627.70/0.740532.46/0.897518.3M
HIPA [63], 2022DIV2K28.87/0.788027.75/0.7429-18.3M
ACT [64], 2022DIV2K28.78/0.785827.69/0.740630.92/0.9150601K
HAT [49], 2022DIV2K28.68/0.783227.62/0.738230.76/0.9111365K
SwinFIR [51], 2022DIV2K29.02/0.794527.94/0.746331.77/0.923119.2M
IPT [59], 2021ImageNet29.27/0.796828.00/0.751632.44/0.928246.0M
Table 5. Light Field Super-Resolution (LFSR). The best records are in bold font.
Table 5. Light Field Super-Resolution (LFSR). The best records are in bold font.
MethodScaleEPFL
PSNR/SSIM
HCInew
PSNR/SSIM
HCIold
PSNR/SSIM
INRIA
PSNR/SSIM
Parameters
SA-LSA [65], 2022×234.48/0.975937.35/0.977044.31/0.994336.40/0.98433.78M
LFT-transformer [66], 202234.80/0.97837.84/0.97944.52/0.99536.59/0.9861.16M
SA-LSA [65], 2022×428.93/0.916731.19/0.918637.39/0.972030.96/0.95023.78M
LFT-transformer [66], 202229.25/0.92131.46/0.92237.63/0.97431.20/0.9521.16M
Table 6. Image Super-Resolution based on ViT in the Remote Sensing field.
Table 6. Image Super-Resolution based on ViT in the Remote Sensing field.
MethodScaleTraining DatasetAID
PSNR/SSIM
UCMerced
PSNR/ SSIM
Parameters
TransENet [67], 2022×2UCMerced, AID35.28/0.937434.03/0.9301-
×429.38/0.790927.77/0.7630-
SWCGAN [57], 2022×4-27.633.8M
Table 7. MRI Images Super-Resolution. The best records are in bold font.
Table 7. MRI Images Super-Resolution. The best records are in bold font.
MethodScaleTraining DatasetKirby21
PSNR/SSIM
ANVIL-adult
PSNR/SSIM
MSSEG
PSNR/SSIM
BraTS2018
PSNR/SSIM
Parameters
ASFT [68], 2022×2Kirby2143.68 ± 2.08/
0.9965 ± 0.0014
40.96 ± 1.00/
0.9906 ± 0.0013
41.22 ± 1.37/
0.9978 ± 0.0004
-1.85M
CFTN [69], 202239.70 0.9847---21.93M
ASFT [68], 2022×340.19 ± 2.04/0.9882 ± 0.003437.54 ± 1.10/0.9703 ± 0.004136.82 ± 1.43 /0.9868 ± 0.0021-1.85M
CFTN [69], 202236.03 0.9612---21.93M
Cohf-T [70], 2022×3----34.84/0.9507152M
×4---33.26/0.9425
Table 8. Hyperspectral Image Super-resolution ×4. Best records are in bold font.
Table 8. Hyperspectral Image Super-resolution ×4. Best records are in bold font.
MethodTraining
Dataset
Harvard
PSNR/SSIM
CAVE
PSNR/SSIM
Houston
PSNR/SSIM
Pavia Centre
PSNR/SSIM
Parameters
3DT-Net [58], 2021Harvard, CAVE50.93/0.99648.05/0.991--3.46M
Fusformer [71], 2022Harvard, CAVE48.56/0.99544.42/0.984--0.10M
Interactformer [72] 2022---29.74/0.918128.51/0.88974465 K
Table 9. Different Models of Image Super-Resolution for Different Datasets.
Table 9. Different Models of Image Super-Resolution for Different Datasets.
MethodSub TaskDatasetPSNR/SSIMParameters
TTSR [73], 2020 Reference-based SR ×4CUFED527.09/0.8049.10M
Sun8030.02/0.814
Urban10025.87/0.784
Manga10930.09/0.907
CTCNet [75], 2022 Face Image SR ×8CelebA28.37/0.8115-
Helen27.08/0.8077
BN-CSNT [76], 2022Thermal Image SRPBVS-2022 ×221.08/0.7803-
PBVS-2022 ×433.64/0.9263
Table 10. Image denoising.
Table 10. Image denoising.
MethodTrain DatasetDatasetNoise Factor (NF)PSNR/SSIMParameters
CVT [81], 2021DIV2KSET 121534.548-
2531.865
3027.676
BSD681533.790
2530.828
3026.688
SwinIR [50], 2021DIV2KKodak241535.3411.8M
2532.89
5029.79
McMaster1535.61
2533.20
5030.22
Urban1001535.13
2532.90
5029.82
CBSD681534.42
2531.78
5028.56
IPT [59], 2021ImageNetCBSD683032.3246.0M
5029.88
Uformer [86], 2022SIDD and DNDSIDDReal world Noise39.89/0.96050.88M
DNDReal world Noise40.04/0.956
Restormer [43], 2022DIV2K and SIDDCBSD681534.4025.31M
2531.79
5028.60
Urban1001535.13
2532.96
5030.02
SIDDReal world Noise40.02/0.960
DnT [82], 2022-Set92532.18-
5029.29/
7527.62/
10026.31/
CBSD682528.78/8.16
5025.72/7.02
TECDNet [85], 2022-SIDDReal world Noise39.77/0.97020.87M
DNDReal world Noise39.92/0.955
TC-Net [86], 2022-SIDDReal world Noise39.69/0.970-
DNDReal world Noise39.88/0.954
SUNet [87], 2022DIV2KCBSD681035.94 0.95899M
3030.28 0.870
5027.85 0.799
Kodak241036.79 0.953
3031.82 0.899
5029.54 0.810
Pocoformer [88], 2022Real-world polarized color imageReal-world polarized color imageReal world Noise39.33/0.96626.26M
TransCT [89], 2021Mayo Clinic Low-Dose CTMayoReal world Noise0.923 ± 0.024-
Pig Real world Noise0.87 ± 0.029
TED-Net [90], 2021Mayo Clinic Low-Dose CTMayoReal world Noise/0.914418.88M
Eformer [91], 2021Mayo Clinic Low-Dose CTMayoReal world Noise43.487/0.9861-
Table 11. Image deblurring.
Table 11. Image deblurring.
MethodDataset TrainingGoPro
PSNR/SSIM
HIDE
PSNR/SSIM
RealBlur-R
PSNR/SSIM
Parameters
Uformer [86], 2022GoPro33.06/0.96730.90/0.95336.19/0.95650.88M
Restormer [43], 2022GoPro32.92/0.96131.22/0.94236.19/0.95725.31M
Multi-scale Cubic-Mixer [96], 20224KRD33.79/0.962-39.66/0.96940M
Stripformer [97], 2022RealBlur33.08/0.96231.03/0.94039.84/0.97420M
Table 12. Image Dehazing.
Table 12. Image Dehazing.
MethodDataset TrainingSOTS-Indoor
PSNR/SSIM
SOTS-Outdoor
PSNR/SSIM
HSTS
PSNR/SSIM
O-Hazy
PSNR/SSIM
Parameters
DeHamer [99], 2022RESIDE36.63/0.988135.18/0.9860---
TCAM [100], 2022RESIDE21.44/0.8851-23.83/0.9022--
ISM [101], 2022RESIDE36.34/0.983630.85/0.962830.40/0.9696--
Jiao et al. [102], 2022O-hazy and I-hazy---15.89/0.56-
Song et al. [103], 2022RESIDE38.46/0.99434.29/0.983--4.634M
Zhao et al. [104], 2021RESIDE32.17/0.970--29.87/0.758-
Table 13. JPEG compression artifact reduction.
Table 13. JPEG compression artifact reduction.
MethodTypeDatasetQuality FactorPSNR/SSIMParameters
SwinIR [50], 2021Single ImageClassic51030.27/0.824911.8M
2032.52/0.8748
3033.73/0.8961
4034.52/0.9082
LIVE11029.86/0.8287
2032.25/0.8909
3033.69/0.9174
4034.67/0.9317
PTNet [112], 2022Pair Stereo imagesFlickr10241028.05/0.84030.91 M
2030.39/0.9017
3031.83/0.9264
KITTI20121031.43/0.8786
2033.85/0.9231
3035.18/0.9404
KITTI20151031.42/0.8730
2034.07/0.9245
3035.57/0.9449
Middlebury1032.05/0.8676
2034.51/0.9200
3035.85/0.9400
Table 14. Removing Adverse Weather Conditions.
Table 14. Removing Adverse Weather Conditions.
MethodDataset TrainingRain100L
PSNR/SSIM
Rain100H
PSNR/SSIM
SPAD
PSNR/SSIM
Raindrop800
PSNR/SSIM
Snow100K
PSNR/SSIM
Parameters
ETDNet [132], 2021Rain100L and Rain100H41.09/0.98632.35/0.9299---32.97M
SDNet [133], 2021Rain100L and Rain100H37.92/0.984328.26/0.8957---2.14M
IPT [59], 2021ImageNet-41.62/0.9880---46.0M
Uformer [86], 2022SPAD--47.84/0.9925--50.88M
Restormer [43], 2022Rain100L and Rain100H38.99/0.97831.46/0.904---25.31M
TAPE [134], 2022Rain200H,
Raindrop800 and Snow100K
33.17--27.6926.331.07M
TransWeather [135], 2022Raindrop800 and Snow100K---34.55/0.950233.78/0.928731 M
SiamTrans [136], 2022NTURain27.02/0.9024---26.05/0.8605-
Table 15. Image Restoration and Enhancement.
Table 15. Image Restoration and Enhancement.
MethodTypeTraining DatasetPSNR/SSIMParameters
RFormer [137], 2022Medical Real Fundus restoration120 pair Real Fundus28.38/0.86321.11M
UIE-IPT [139], 2022Underwater ImagesUFO-12023.14/0.9046.0M
FRGAN [140], 2021Face RestorationVGGFace2 and CASIA-WebFace23.54/0.8199-
DocEnTr [141], 2022Enhance Handwritten Document ImagesDIBCO-/20.81-
H-DIBCO-/22.29
STUNet [142], 2022Blind Face RestorationEDFace-Celeb-1M (BFR128)24.5500/0.6978-
EDFace-Celeb-150K (BFR512)27.1833/0.7346
DuDoTrans [143], 2021Medical CT ReconstructionNIH-AAPM32.68/0.9047-
COVID-1937.83/0.9727
U2-Former [144], 2021Image Reflection RemovalPLNetReal2023.67/0.835-
Nature24.75/0.848
Solid25.27/0.907
Wild25.68/0.905
Postcard22.43/0.889
Image Rain RemovalRain100L39.31/0.982-
Rain100H30.87/0.899
Image Hazing RemovalRESIDEIndoor36.42/0.988-
Outdoor31.10/0.976
SMIR [145], 2021MRI image reconstruction
different sampling ratios
HCP10%-/0.7211M
20%-/0.86
30%-/0.87
40%-/0.89
50%-/0.91
Wang et al. [146], 2021Image ReconstructionImageNet 20kSet5-/32.61-
Set14-/28.92
B100-/27.78
Urban100-/26.82
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ali, A.M.; Benjdira, B.; Koubaa, A.; El-Shafai, W.; Khan, Z.; Boulila, W. Vision Transformers in Image Restoration: A Survey. Sensors 2023, 23, 2385. https://doi.org/10.3390/s23052385

AMA Style

Ali AM, Benjdira B, Koubaa A, El-Shafai W, Khan Z, Boulila W. Vision Transformers in Image Restoration: A Survey. Sensors. 2023; 23(5):2385. https://doi.org/10.3390/s23052385

Chicago/Turabian Style

Ali, Anas M., Bilel Benjdira, Anis Koubaa, Walid El-Shafai, Zahid Khan, and Wadii Boulila. 2023. "Vision Transformers in Image Restoration: A Survey" Sensors 23, no. 5: 2385. https://doi.org/10.3390/s23052385

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop