Next Article in Journal
Special Issue on New Trends in Nonlinear Optics in Nanostructures and Plasmonics
Next Article in Special Issue
Automated Brain Tumor Identification in Biomedical Radiology Images: A Multi-Model Ensemble Deep Learning Approach
Previous Article in Journal
Study on the Nonlinear Permeability Mechanism and Pore Structure Characteristics of Deep Confined Aquifers
Previous Article in Special Issue
Object Detection for Brain Cancer Detection and Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Smoothing for Visual Improvement of Image Quality via the p(x)-Laplacian Operator Effects of the p(x)-Laplacian Smoothing Operator on Digital Image Restoration: Contribution to an Adaptive Control Criterion

1
Department of Mathematics, Computer Sciences, 97110 Pointe a Pitre, France
2
LAMIA Laboratory, EA 4540, 97157 Pointe-a-Pitre, France
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(20), 11600; https://doi.org/10.3390/app132011600
Submission received: 28 March 2023 / Revised: 25 September 2023 / Accepted: 10 October 2023 / Published: 23 October 2023

Abstract

:
This article concerns the improvement of digital image quality using mathematical tools such as nonlinear partial differential operators. In this paper, to perform smoothing on digital images, we propose to use the p(x)-Laplacian operator. Its smoothing power plays a main role in the restoration process. This enables us to dynamically process certain areas of an image. We used a mathematical model of image regularisation that is described by a nonlinear diffusion Equation (this diffusion is modelled by the p(x)-Laplacian operator). We implemented the continuous model in order to observe the steps of the regularisation process to understand the effects of the different parameters of the model on an image. This will enable parameters to be used and adapted in order to provide a proposed solution.

1. Introduction

The smoothing of digital images is an important theme in the field of image correction. It provides a method for erasing unwanted effects caused by various factors, such as the quality of the acquisition material, human error or the degradation induced in order to reduce its storage space.
There are many processing operators for smoothing digital images. Their aim is to maintain highlights in the relevant areas and smooth the rest.
Among these operators, we can mention the p(x)-Laplacian operator, which mathematically models the diffusion of a signal (in this case, a digital image), taking into account the local information of this signal. This operator, by its formulation, is able to perform smoothing according to a criterion. Generally, in the literature, the criterion defining the degree of smoothing for the p(x)-Laplacian operator is calculated at the beginning of processing. In this way, the amount of smoothing is set to the same level. In the present work, contrary to what is generally carried out, we wanted to make the processing adaptive by locally adjusting the criterion taken into account in the definition of the p(x)-Laplacian operator. Thus, this procedure allows us to select, from the observed image, the areas that will undergo the necessary smoothing while preserving those that we consider more important.
In recent years, the most common version is the lossy compression, which often results in image degradation. In this context, we propose the adaptation of a nonlinear mathematical model for the improvement of these images. Currently, most image restoration methods are based on partial differential Equations (see [1,2,3,4]) that involve two parameters, one of which is always considered to be fixed.
This paper is organised as follows:
  • Section 1 introduces the subject matter of this work and the setting of the problem;
  • In Section 2, we present the main mathematical tools associated with a continuous version of the problem and its formulation. The use of graph theory proposes a discretised version of the initial problem by employing an adaptive algorithm. In addition, we recall some current properties of the p ( x ) -Laplacian operator;
  • Section 3 discusses the simulation approaches;
  • Section 4 proposes some illustrations of usually referenced images which are provided in well-known databases;
  • Section 5 addresses some comparative studies between our proposed approach and others.

2. Context and Problems

2.1. Pictures

Images may have minor changes that are in need of smoothing in order to improve their readability. The image shown in Figure 1 is an example of the type of degradation that can be seen in an image.
All equipment has sensors and acquisition algorithms that are more or less efficient, which means that quality is lost. To this, one must add the human factor [5], which is often the reason why the work of an inexperienced photographer has lower quality.

2.2. Adaptive Smoothing Using the p(x)-Laplacian

Mathematical modelling is used to model phenomena from various fields, including physics, chemistry, biology and, in particular, digital image processing. In some cases, computer science makes it possible to apply some of these patterns to improve the quality of digital images. In this paper, we use the results from a class governed by a nonlinear operator of the pseudo-Laplacian-type with a variable exponent. In our case, this function denotes an image (see [6]).
Our aim is to improve the quality of the images by homogenising areas where the gradient is low, while preserving the amount of detail in the areas where the gradient is high.

3. The Method

3.1. Presentation of Some Models with a Diffusion Term

First of all, we must consider that an observed digital image f 0 is the noisy image. This digital image could be interpreted as a real image f (the desired one) with one (or several) noise(s) or residual(s) of image r (see [7], p. 6, (18) r = N(0; σ ) is considered as a Gaussian noise with mean = 0 and deviation σ ). The effects of the degradation (caused by the camera or object motion, atmospheric turbulence optics, etc.) on the piecewise smooth image f are modelled by a linear or a nonlinear operator K. Usually, the operator K is named the degradation operator or degradation function, and it is also known as a point spread function (PSF) or kernel.
This is modelled by the following equation
f 0 = K ( f ) + r .
The smoothing process consists of approaching f from f 0 by means of a minimisation problem of the energy function E ( f ; f 0 ; λ ) as defined below
E ( f , f 0 , λ ) = E r e g ( f ) + λ E a p p r o x ( f , f 0 )
where
  • E ( f ; f 0 ; λ ) stands for the energy that allows one to go from f 0 to the desired image;
  • E r e g ( f ) denotes the regularisation energy to the real image f;
  • E a p p r o x ( f , f 0 ) interprets the energy of approximation between the real image f and the observed image f 0 , also named the fitting term or the fidelity term;
  • λ is the fidelity parameter or the penalty parameter.
The main goal is to recover an image corrupted by the noise while preserving edges and some details on the observed image and small scale structures such as textures and other details on the observed image.
Among examples of energies functions, we can cite the following continuous energies functions.
E ( f , f 0 , λ ) = Ω f d x + λ 2 Ω f f 0 2 d x ,
E ( f , f 0 , λ ) = Ω f 2 d x + λ 2 Ω f f 0 2 d x .
Models (3) and (4) were introduced by Rudin et al. in [3].
We can also cite the model associated with the following energy function
E ( f , f 0 , λ ) = Ω f p ( x ) d x + λ 2 Ω f f 0 2 d x .
here, the variable exponent p is defined as follows
p ( x ) = 1 if | x | > M , 2 if x = 0
where M > 0 is a positive constant. Precisely, M > 0 is either the index of the preservation (i.e., x > M ) or the index of degradation (i.e., x M ) . For more details, a reader interested in the field can refer to [1].
A variant of Model (5) is the following
E ( f , f 0 , ε ) = 1 ε f ( x ) < ε f 2 d x + 1 2 ε f ε f ε 2 d x + λ 2 Ω f f 0 2 d x .
This energy function was introduced by Chambolle and Lions [2]. In this model, we can observe that p = 2 for f ( x ) < ε , while p = 1 for f ( x ) ε .

3.2. The Formulation of the Present Model

To date, the previous models have been considered with a degree of smoothness p independent of the location x in the domain Ω . The aim of our paper is to observe the effects of the degree of smoothness depending on the location x. Taking into account the spatial location x, the smoothness is considered a dynamic process linked to the following energy function.
E ( f , f 0 , λ ) = Ω 1 p ( x ) f p ( x ) d x + λ 2 Ω f f 0 2 d x
where
p ( x ) = 1 + 1 1 + k | G σ f 0 ( x ) | 2 , k , σ > 0 ,
G σ ( x ) = σ 1 exp x 2 / 4 σ 2 represents the Gaussian filter function. Models (7)–(8) were suggested and investigated by Chen, Levine, and Rao in [4] in the special context of BV p ( · ) functions. For more details and properties on BV p ( · ) functions, an interested reader can consult [8].
The variable exponent function p ( · ) belongs in the interval [ 1 ; + [
 It represents a regularisation degree. It also improves the quality of the image resulting from the smoothing process. More precisely, this function encodes the information on the regions where the gradient is near zero (the homogeneous regions). Therefore, the model avoids the staircase effect by always keeping the edges. When the p ( · ) function is constant ( p ( · ) = p ) , this situation is namely called static restoration while the opposite situation is called dynamic restoration.
In addition, λ is the data fidelity parameter. This parameter specifies the compromise between the two competing terms and determines the amount of noise removed.
It is not difficult to see that the objective function is strictly convex, and so using standard convex analysis arguments demonstrates that the optimisation problem has a unique solution.
f represents the local variation of the function f over the domain Ω . Let us note that, in Model (7), the parameters p ( · ) and λ strike a balance between the regularisation and approximation energies.

3.3. Preliminaries and Results for the Minimising Problem Associated with (7)

Before dealing with the results of the insurance of the minimisation problem, we recall some preliminary tools.
Let Ω be an open-bounded domain in R N with a sufficient regular boundary. Throughout this section, we assume that p is a measurable function from Ω to ] 1 ; + [ : We write
p + : = sup x Ω p ( x ) , p : = inf x Ω p ( x ) .
L p ( · ) Ω , the Lebesgue space with a variable exponent p ( · ) , which is the so-called Nakano space and a special sort of Musielak–Orlicz spaces (see [9]). It is defined as follows:
L p ( · ) ( Ω ) : = { u : Ω R ; measurable in Ω and Ω | u ( x ) | p ( x ) d x < + }
This space is equipped with the Luxemburg-type norm
u p ( x ) = inf s > 0 ; Ω u ( x ) s p ( x ) d x 1 .
W 1 , p ( · ) ( Ω ) designates the variable exponent Sobolev spaces on the open-bounded domain Ω . It is defined as follows:
W 1 , p ( · ) ( Ω ) : = u L p ( · ) ( Ω ) ; u x i L p ( · ) ( Ω ) for all i = 1 , 2 , , N
with the norm
u W 1 , p ( · ) ( Ω ) : = u p ( x ) + u p ( x )
where u p ( x ) defines the norm of u on the Lebesgue space L p ( · ) ( Ω ) .
The complete and full results on the Lebesgue space and Sobolev spaces with a variable exponent can be found in the monograph in [10].
Now, let us consider the energy function E ( f , f 0 , λ ) defined as in (7), and then the smoothing process corresponds to the optimal problem. Below, we give the existence and uniqueness results.
Proposition 1. 
Under the assumptions 1 < p p ( x ) < p + < 2 and f 0 L 2 ( Ω ) , the minimising problem
inf f W 1 , p ( Ω ) Ω f p ( x ) d x + λ 2 Ω f f 0 2 d x .
has a unique solution f in W 1 , p ( Ω ) .
Proof. 
We present the most important points to ensure that the desired result (11) is achieved. Throughout the proof, we denote by Φ the application defined for any f in W 1 , p ( x ) ( Ω ) by
Φ ( f ) = Ω 1 p ( x ) f ( x ) p ( x ) d x + λ 2 Ω | f ( x ) f 0 ( x ) | 2 d x .
The proof will be divided into two steps.
  • Step 1 In the first step, we establish the existence of an element f of W 1 , p ( x ) ( Ω ) such that Φ ( f ) = inf f W 1 , p ( x ) ( Ω ) Φ ( f ) .
    For this, we can already notice that, for any f W 1 , p ( x ) ( Ω ) , Φ ( f ) is strictly positive. This implies that there exists a real number α , such that
    inf f B V ( Ω ) Φ ( f ) = α .
    To do this, we consider a minimising sequence ( f n ) n of Φ .
    || f n || 1 , p ( x ) C .
    Indeed, since f n is a minimising sequence of Φ , then we specifically have
    Φ ( f ) α + 1 n , for all n N { 0 } .
    Consequently, by the definition of Φ given in (12), we successively deduce
    Ω 1 p ( x ) f n ( x ) p ( x ) d x α + 1
    and
    Ω | f n ( x ) f 0 ( x ) | 2 d x α + 1 .
    By using the triangular inequality and (17), it follows that there exists a constant K 1 > 0 , such that
    || f n || L 2 ( Ω ) K 1 .
    On the other hand, since p + : = sup · x Ω p ( x ) < 2 , the embedding of L 2 ( Ω ) into L p ( x ) ( Ω ) is continuous (i.e., there exists a constant C ˜ > 0 such that || f n || L p ( x ) ( Ω ) C ˜ || f n || L 2 ( Ω ) ). Thus, combined with (18), we find a constant K 2 > 0 , such that
    || f n || L p ( x ) ( Ω ) K 2 .
    To continue, we observe that after combining the equivalence norm properties [11] (Theorem 2.3) with the estimate in (16), we can deduce that there exists a positive constant K 3 such that
    || f n || L p ( x ) ( Ω ) K 3 , for all n N .
    Hence, let us return to the definition in (10) of the norm · W 1 , p ( x ) ( Ω ) . By using the estimates from (19) and (20), we obtain
    || f n || W 1 , p ( x ) ( Ω ) C .
    In other words, we obtain the desired estimate (14). Moreover, since W 1 , p ( x ) ( Ω ) is a reflexive space, we can extract from the sequence f n a sub-sequence that converges weakly in W 1 , p ( x ) ( Ω ) . We denote this weak limit by f . Since the application Φ is strictly convex (as a sum of two convex functions on W 1 , p ( x ) ( Ω ) ) and continuous for the topology of the norm · W 1 , p ( x ) ( Ω ) , we can apply [12] (Corollary 3.9). In this case, we obtain
    Φ ( f ) lim inf n + Φ ( f n ) .
    Consequently, from the existence of inf f W 1 , p ( x ) ( Ω ) , Φ ( f ) = α leads to f on one side
    α Φ ( f ) .
    On the other side, taking into account that the sequence f n has the property from (15), this implies that, after passing to the lower limit on n, the following yields
    lim inf n + Φ ( f n ) α .
    Therefore, it is possible to conclude that
    Φ ( f ) α .
    Then, estimates from (21) and (22) lead to the conclusion that
    Φ ( f ) = α = inf f W 1 , p ( x ) ( Ω ) Φ ( f )
    which establishes that the minimisation problem has a solution in W 1 , p ( x ) ( Ω ) . The uniqueness of f is due to the strict convexity property of the function Φ . Indeed, let us reason by the opposite. This means that, suppose that the application Φ reaches its minimum in f 1 and that f 2 has two distinct elements. As Φ is strictly convex on W 1 , p ( x ) ( Ω ) , particularly for all β ] 0 , 1 [ , we have
    Φ ( β f 1 + ( 1 β ) f 2 ) < β Φ ( f 1 ) + ( 1 β ) Φ ( f 2 ) .
    Thus, because α = inf f W 1 , p ( x ) ( Ω ) Φ ( f ) , we specifically have
    α Φ ( β f 1 + ( 1 β ) f 2 ) .
    By observing (23), we can thus conclude that a contradiction yields since the left-hand member is estimated below by α , while the right-hand member is estimated above by α conforming to the hypothesis on f 1 and f 2 , respectively.
    The uniqueness of the solution of the minimising problem occurs. □
Remark 1. 
The solution of (11) is characterised as the solution to a partial differential equation described by the stationary equation.
d i v f p ( · ) 2 f + λ f 0 f = 0 , i n Ω .
Indeed, employing the Euler–Lagrange equation, we can noticed that the unique solution of the minimising problem (11) coincides with the problem of finding solutions for the partial differential equation in (24). Therefore, to determine in practice this unique solution of the optimal problem (11), it is equivalent to solving the partial differential equation in (24).
This point is not our aim in this paper. We propose in the next part another approach comprising the discretisation of the minimisation (11).

3.4. The Corresponding Discretised Model by Means of the Graph Theory

After reviewing the continuous case, we now discuss how to translate the framework to the discrete setting of a graph. In contrast to the well-known numerical recalled earlier, it is proposed to apply the theory of weighted graphs applied to regularisation processing [13].
We introduce this section by recalling the main tools on graph theory.
Let d N { 0 } . V = v 1 , , v d designate a finite set of d vertices v i , i = 1 , , d . Every vertex v V consists of a pixel for the image. The finite set V plays the same role as the domain Ω of R N given previously in the general case. Let u and v be two vertices, where ( u , v ) is named the edge. We write u v for two adjacent vertices. The neighbourhood of a vertex u (i.e., the set of vertices adjacent to u) is denoted by N ( u ) . We say that an edge ( u , v ) is incident on vertex u if the edge starts from u. An edge ( u , v ) is a self-loop if u = v .
Let ω : V × V R + be a symmetric function (i.e., for every u , v in V , ω ( u , v ) = ω ( v , u ) ). In [13,14], one can find some examples of the weight functions ω . If ( u , v ) is not an edge, one has ω ( u , v ) = 0 , and thus the set of edges E V × V can be characterised by the weight function ω as E = ( u , v ) V × V | ω ( u , v ) > 0 .
The degree of a vertex u is defined as δ ω ( u ) = v u ω ( u , v ) .
V, E, and ω define a weighted graph G = ( V , E , ω ) . The elements of V are the vertices of the graph, and the elements of E are the edges of the graph. A path is a sequence of vertices ( v 1 , v 2 , , v d ) such that ( v i , v i + 1 ) is an edge for all 1 i d 1 .
The following highlights some properties of graph theory.
A graph is connected when there is a path between any two vertices.
A graph is undirected when the set of edges is symmetric, i.e., for each edge ( u , v ) E , we also have ( v , u ) E , and the weight function ω satisfies the symmetry condition ω ( u , v ) = ω ( v , u ) . In the following, the graphs are always assumed to be connected, undirected, and have no self-loops or multiple edges.
Now, we designate by H ( V ) the space of real-valued functions f defined on the vertices of a weighted graph ( V , E , ω ) , i.e., each function f : V R in H ( V ) is assigned to be a real value f ( u ) to each vertex u V . For a function f H ( V ) , the L p ( x ) ( V ) norm of f is given by
f p ( u ) = inf λ > 0 ; u V f ( u ) λ p ( u ) < + .
The formulation in (25) must be understood as a discretised version of the definition in (9).
Let us note that, for p ( u ) = p to be constant, f p ( u ) simply becomes the classical L p -norm. This means that
f p p = u V f ( u ) p .
The mapping ( f , g ) f , g H ( V ) = u V f ( u ) g ( u ) with f , g H ( V ) is an inner product on H ( V ) . Moreover, the mapping
f f H ( V ) = f , f H ( V )
is a norm.
H ( V ) , f H ( V ) has a Hilbert’s space structure on R .
The notation u v always denotes two adjacent vertices. The local variation defined by
ω f ( v ) = v u ω ( u , v ) ( f ( v ) f ( u ) ) 2
is a semi-norm which measures the regularity of a function around a vertex of the graph.
Translated from the usual notation applied to the classical differential operators in PDEs, the fundamental tools are the weighted partial differentials on graphs. For more detailed information on these operators, one is referred to [1,13,15].
Let G ( V , E , ω ) be a weighted graph and f H ( V ) be a function on the set of vertices V of G . Then, one can define the weighted partial differentials of f at a u V vertex in the direction of a v V vertex as
ω , i f ( u ) = ω u , v i f ( v i ) f ( u ) , i = 1 , , d .
Based on this definition, one can straightforwardly introduce the weighted gradient operator on graphs ω : H ( V ) H ( E ) , which is defined on a u V vertex, i.e.,
ω f ( u ) = ω , i f ( u ) i = 1 , d T = ( ω u , v i f ( v i ) f ( u ) ) i = 1 , d T , u V .
The weighted gradient at a u V vertex must be understood as a function of H ( V ) , and thus the L p ( u ) ( V ) norm of this finite vector represents its local variation and is given as follows
v V 1 p ( v ) v u ω ( u , v ) ( f ( v ) f ( u ) ) 2 p ( v ) 2 .
We notice that the bounded domain Ω presented previously is a subset of R 2 . In the discretised formulation, the smoothing process is reduced to be an optimal problem as follows
m i n f H ( V ) v V 1 p ( v ) w f ( v ) p ( v ) + λ 2 v V f ( v ) f 0 ( v ) 2 2 .
The use of this framework may be considered as a discrete analogue of the continuous regularisation on weighted graphs. One of the advantages is that the regularisation is directly expressed in a discrete setting, and involved processing methods are computed by simple and efficient iterative algorithms without solving any partial differential equations (PDEs). The idea of using functions on weighted graphs in a regularisation process has been proposed in other contexts, such as semi-supervised data learning [16]. However, in our paper, we adopted a digital approach contrary to the numerical approaches used in [13], for instance. From this, the optimal problem in (26) becomes
m i n f H V v V 1 p ( v ) v u ω ( u , v ) ( f ( v ) f ( u ) ) 2 p ( v ) 2 + λ 2 v V f ( v ) f 0 ( v ) 2 .
The models with p ( · ) = 1 or p ( · ) = 2 have been treated (for more details, the reader can consult [4]).
In our work, we treat the general case 1 < p ( x ) < 2 . The next proposition is the following.
Proposition 2. 
The discretised minimising problem in (27) has a unique solution.
Proof. 
By using the elementaries tools of convex analysis applied to the functional
f v V 1 p ( v ) v u ω ( u , v ) ( f ( v ) f ( u ) ) 2 p ( v ) 2 + λ 2 v V f ( v ) f 0 ( v ) 2 ,
p ( · ) and λ are defined as in the previous case. Obviously, we note that the problem in (27) has a unique global minimum. Indeed, we define the functions φ 1 and φ 2 as follows
f φ 1 v V 1 p ( v ) v u ω ( u , v ) ( f ( v ) f ( u ) ) 2 p ( v ) 2 and f φ 2 λ 2 v V f ( v ) f 0 ( v ) 2 .
It is clear that each function is convex in V . Moreover, φ 2 is coercive in V , which means that φ 2 ( f ) tends to + , when f 2 goes to + . This fact is easily obtained using the triangular inequality.
Because of this, it could be argued that the problem described by (27) has a unique solution in V .
In our work, the weight ω considered is
ω ( u , v ) = e x p f ( u ) f ( v ) 2 σ 2
where
σ 2 = 1 V v V f ( v ) f 0 ( v ) 2
and the term V designates the cardinal of vertices of the graph. This choice is favoured because we only take into account the difference between a vertex and its neighbourhood, and we also consider the deviation such that
v V f ( v ) f 0 ( v ) 2 σ 2 .

4. Some Illustrations of the Effects of Parameters p and λ

The process for calculating the p(x)-Laplacian on an image is that of the implementation in Algorithm 1. The effect parameters p and λ of the model on an image are illustrated in Figure 2.
  • In the case 0 < p < 1 , the pictures are partitioned into several homogeneous areas;
  • In the case 1 p 2 , when p approaches 2, there is a smoothing effect, whereas when p converges on 1 there is no smoothing effect;
  • The increase in parameter λ limits the effects of p.
Overall, when acting on the value of these parameters, several smoothing levels are reached which make the pictures striking or blurred.
Algorithm 1 Algorithm regularisation
x and j are positions in IM
I M F is the map result
I M T is a temporary map
Compute gradient map G on image I M
Compute P map values on image I M such as P [ x ] : = 1 + 1 1 + K G [ x ] I M [ x ] 2
ε := 0.01
λ := user value in [ 0 ; 1 ]
For all x in image I M
    I M F [ x ] := I M [ x ]
   Do
       I M T [ x ] := I M F [ x ]
      For j in neighbours of x
           I M F [ x ] : = I M F [ x ] + 1 P [ j ] G [ j ] P [ j ]
           I M F [ x ] : = I M F [ x ] + λ 2 | | I M F [ j ] I M T [ j ] | | 2
   While | I M F [ x ] I M T [ x ] | > ε

4.1. Algorithmic Formulation of the Mathematical Model

4.1.1. Implementation Procedure

As in [4], an image is represented by a graph. In this structure, every vertex contains the coordinates of a pixel, and each edge indicates the difference between two pixels. The mathematical model in (26) is applied to the graph which must satisfy the existence and uniqueness requirements given in Section 3.1 into account.

4.1.2. Towards an Adaptive Model

In the literature, the parameter p is often treated as static in the studies of this model. Otherwise, in our study, we propose to adapt parameter λ in order to control the smoothing effects due to parameter p. The aim is to take account of the details in the image.
λ  model evaluation process to control the p effects: the parameters p and λ play a crucial role in the dynamical smoothing process. Therefore, we calculate the difference in colour between a pixel and its neighbours in the following way. The gradient was chosen in accordance with the objective that had to be achieved.
Implementation of the parameter control process: In a preliminary step, we calculate the gradient of the image in order to detect prominent areas as well as noise. The idea is to vary the parameters in relation to the gradient in order to preserve details (areas of interest) and smooth out other areas. However, since the gradient is a local operator, it is sensitive to noise and also generates very large anomalous values that alter the process. For this reason, it is advisable to check the values obtained in each pixel in order to reduce the effect of the noise.
Efficient interval for the gradient: the three following types of pixels are focused on:
Noise-affected pixels;
Noise-unaffected pixels;
Outline pixels.
We would like to point out that the gradient values should be found by taking into account pixels (i.e., noise-affected and contour pixels which have the least effect on neighbouring pixels) that affect the process as little as possible. Thanks to these values, we will be able to create intervals that will allow us to select relevant values of the gradient. To do this, these intervals, defined by the interquartile deviation of the gradient distribution, are given by the statistical theory of data dispersion (see [17]) in the series according to the following formula
Q 1 k ( Q 3 Q 1 ) , Q 3 + k ( Q 3 Q 1 )
where Q 1 and Q 3 are the first quartile and the third quartile, respectively. The positive constant k chosen as either 1.5 or 3 was discussed by Frigge et al. in [18].
These authors concluded for k that 1.5 is a suitable value for most situations. According to them, there are cases where 3 may be an appropriate compromise.
Thus, the following intervals were chosen:
I 1.5 = Q 1 1.5 ( Q 3 Q 1 ) , Q 3 + 1.5 ( Q 3 Q 1 )
and
I 3 = Q 1 3 ( Q 3 Q 1 ) , Q 3 + 3 ( Q 3 Q 1 )
For the interval I 1.5 , the outer values correspond to major aberrant values (such as noise).
Study of the sturdiness of these criteria: After a series of tests on an image database, we found that the interval I 1.5 is noise-sensitive and allows one to consider areas to be smoothed. Algorithm 2 implements the process, and Figure 3 shows the effects of such a choice on two images.
Pixels that participate in the dynamic processing are shown in red. They will be submitted to variations in p and λ depending on the gradient value. Pixels that will not be modified are shown in green. Their values are located in the highest gradient areas. The λ values force the smoothing process to be neutralised. The adaptive model effects are shown in Figure 4.
We can see that the areas that are considered to be visually homogeneous (Lena’s skin texture, for example) are smoothed in order to further standardise the colours. On the contrary, the contours with a high gradient are preserved.
Algorithm 2 Algorithm regularisation with λ dynamic
x and j are positions in IM
I M F is the map result
I M T is a temporary map
Compute gradient map G on image I M
Compute P map values on image I M such as P [ x ] : = 1 + 1 1 + K G [ x ] I M [ x ] 2
ε := 0.01
Compute quartile Q 1 , Q 2 , Q 3 , Q 4 on the gradient norm
Compute λ map on image with quartile such as:
λ [ x ] = 1 i f g r a d i e n t n o r m a t x n o t i n I 1.5 1 g r a d i e n t n o r m e l s e .
For all x in image I M
    I M T [ x ] := I M [ x ]
   Do
       I M F [ x ] := I M T [ x ]
      For j in neighbours of x
          I M F [ x ] : = I M F [ x ] + 1 P [ j ] G [ j ] P [ j ]
          I M F [ x ] : = I M F [ x ] + λ [ x ] 2 | | I M F [ j ] I M T [ j ] | | 2
   While | I M F [ x ] I M T [ x ] | > ε

5. Results and Discussion

Presentation of the database and discussion of execution times and complexity
We used two images to illustrate the treatments:
Lena image, which is widely used in image processing;
Tiger image, because it contains the most significant data in image processing. The tiger is present within the focal depth, while the rest of the scene, which is less important, should be homogenised.
The execution of the programme mainly depends on the smoothing factor and the number of pixels to be smoothed. The higher the smoothing factor, the more a pixel must be modified. Processing continues until every pixel in the image is stabilised. The complexity is in O ( k N ) , where N is the number of pixels in an image and k is a constant representing the disorder in the original image. By increasing the value of λ in detailed areas, the proposed model reduces the processing time. Figure 5 shows the results of the adaptive model on noise.
We used the adaptive model on 100 images of the Berkeley database set [19]. This database contains colour photographs of various scenes. The results of these studies are presented in the following sections.

5.1. Results with the Peak Signal-to-Noise Ratio (PSNR)

We evaluated the quality of the filtering using the peak signal-to-noise ratio (PSNR) in a first work. This parameter is used to calculate the presence of noise in an image. Its formulation is described in (31):
P S N R = 10 × l o g 10 S N R
S N R = s i g n a l n o i s e = I o r i g i n a l I o r i g i n a l I n o i s y = x y I o r i g i n a l 2 ( x , y ) x y I o r i g i n a l ( x , y ) I n o i s y ( x , y ) 2 .
In our study, there are no noiseless images, and so we go back to the formula in (33) to evaluate the quality of the images from the dynamic processing.
S N R = I f i l t e r e d I f i l t e r e d I o r i g i n a l
The values obtained can be compared with the reference values in Table 1. The values obtained on Lena’s image after varying p and λ are shown in Table 2. These values can be compared with the reference values given in Table 1.
The Lena image with the adaptive model provides a PSNR = 27.66, while the tiger image gives a PSNR = 24.31. In order to validate the proposed filter, we chose to use the adaptive model on 100 images extracted from the Berkeley database [19] (see Figure 6), 14 images extracted from the USC-SIPI Image Database [20] (see Figure 7), and 24 images extracted from Kodak Lossless True Color Image Suite [21] (see Figure 8). Images from these databases are widely known and show different scenes with contrasts.
From the Berkeley database, Table 3 and Table 4 show the DSSIM and PSNR results obtained, and Figure 9 shows two images that obtained the best PSNR following the application of the adaptive model as well as the two images with the worst PSNR.
The values presented in Table 4 are quite homogeneous, as can be seen from the standard deviation, which remains at 2.19, and the average PSNR, which remains at 31.13. This last value represents a good image quality according to this parameter. With regard to the sample used, 26.36 corresponds to a minimum PSNR. It places the worst case towards a good result. Otherwise, for an image quality considered excellent, 36.24 is the best PSNR obtained by the adaptive model.

5.2. Results with the Structural Dissimilarity Measure (DSSIM)

In the first stage, we use structural dissimilarity DSSIM [22] to assess the quality of filtering, and Table 5 shows the results obtained. This measure was developed to evaluate the visual quality of a compressed image with respect to the original image. It uses the SSIM parameter (structural similarity) to measure the similarity of structure between the two images. Its formula is given by (34):
D S S I M ( x , y ) = 1 S S I M ( x , y ) 2
with
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x σ y + c 2 ) ( 2 c o v x y + c 3 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 ) ( σ x σ y + c 3 )
where μ x and μ y are averages in the x and y direction, and σ x 2 and σ y 2 are the variances of x and y, respectively. c 1 , c 2 , c 3 are parameters needed to stabilise the division.
In order to simulate the images from the network, we compressed all images in the dataset. For this study, we chose to use JPEG compression. The images available in the database were compressed in JPEG format with several compression ratios ranging from 10% to 90% quality. We have already shown the impact of dynamic λ with respect to static λ . Table 6 summarises this study.
The algorithmic process gives the best values in the dynamic λ case and where λ is 1.0. In terms of the DSSIM criterion, these two cases give better results. However, the λ value of 1.0 leads to too much smoothing over the whole image, as shown in Figure 10.
In a second step, we compare one side of the original image with the compressed image and the other side with the compressed image after applying our filtering. We used other filtering algorithms in the evaluation process. The method of D. Tschumperle [23] was used, which consists of regularising an image by an anisotropic smoothing process based on PDEs. This model requires parameterisation for its implementation, so we used standard values given in the literature. Gaussian filtering was used, which consists of applying a Gaussian filter to the rows and then to the columns of the image. Peer group filtering was used, which allows an image to be smoothed by replacing a pixel with the weighted sum of its closest neighbours in terms of colour distance. The purpose was to observe whether our proposal leads to improvements in terms of visual rendering, especially in the case of high losses. The results obtained with PSNR and DSSIM are shown in Figure 11 and Figure 12 (for PSNR, a higher value is better, whilst for DSSIM, a lower value is better).
In Figure 11 and Figure 12, the curve representing the compressed image expresses the relevance of the models. In fact, in the case of PSNR, this relevance is provided by a model that gives results above the compressed image and, up to the case of DSSIM, by a model that gives results below the compressed image. However, the use of Gaussian filtering does not satisfy these conditions. In the worst case, the others are generally confused by the compressed image. Finally, we observed, with the help of Table 7, using the standard deviation, the stability of the results on all the databases used in this work. Table 8 shows the average execution times achieved by our model on the image databases used. The algorithm still needs to be optimized to reduce the time required to process an image. The average processing time observed is approximately 10 min per image.
When using the PSNR parameter, the visual rendering is improved for compression with a quality between 10% and 70%. The use of the DSSIM parameter shows that, for a quality of between 10% and 30%, the proposed model improves. In any case, for compression with high information loss, our proposal gives better PSNR and DSSIM values than the just compressed image. As soon as the compression becomes low, we see that it is no longer necessary to use our proposal, since it provides values that are lower than those given by the compressed image.
The work focuses on the proposal of an adaptive control operator for the smoothing strength of the p(x)-Laplacian method. Tests were carried out on the three following image databases: Berkeley, USC-SIPI, and KODAK. We degraded the images of these three databases by applying compression levels with losses. The results in Figure 11 and Figure 12 show a higher efficiency of the p(x)-Laplacian method with our adaptive controller compared to other smoothing methods in the literature with respect to the PSNR and DSSIM evaluation operators. We showed that the p(x)-Laplacian operator with our control operator produces better results than the p(x)-Laplacian with a static λ . Our method produces a λ value for each point of the image adapted to the situation, as shown in the examples in Table 6 with respect to the DSSIM evaluation criterion. Finally, we compared the p(x)-Laplacian with our control operator to other smoothing models in the literature, still based on the same principle of degradation of all images in the datasets. Our proposal gives better results, although for the KODAK image database, we still have roughly equivalent results in terms of the PSNR evaluation criterion.
The contribution of an adaptive controller of the λ parameter of the p(x)-Laplacian operator has not yet been studied in the literature. Our contribution, which aims to control its smoothing strength according to the locations of an image, shows an improvement in the results obtained using this operator. In a final study, we used two noise generation methods from the literature, again based on the image databases Berkeley, KODAK, and USC-SIPI. These methods consist of creating random additive noise to the images based on a probability law for the distribution, such as the Gaussian distribution or the Poisson distribution. The results of applying these methods and then correcting them using the methods presented in this work are shown in Table 9 and Table 10. Overall, we can see that the proposed method ranks second only to the Tchumperle method.

6. Conclusions and Prospects

In the field of image enhancement, smoothing methods are used to correct singularities. Among these methods, we are interested in the p(x)-Laplacian. It is generally used in the literature with a fixed control parameter. This choice does not allow for finer processing over the entire image. We proposed to apply this nonlinear regularisation operator, which we have made adaptive, in order to emphasise important visual information. We therefore proposed an adaptive control criterion that controls smoothing as a function of the image gradient, which we exploited using statistics. In order to evaluate this new model, we used the PSNR and DSSIM parameters. The former indicates a good reconstruction quality when its value is above 24 dB, and the latter assesses the quality of the filtering. From our test database, the results obtained put the average PSNR at 24.80 dB, with the worst result around a reasonable reconstruction. According to the DSSIM parameter, the quality of the image improved, provided that it had undergone many degradations, such as lossy compression.

Author Contributions

Conceptualization, J.-L.H., J.N. and J.V.; Writing—original draft, J.-L.H., J.N., J.V. and I.-P.M.; Visualization, J.-L.H., J.N. and J.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bougleux, S.; Elmoataz, A.; Melkemi, M. Local and nonlocal discrete regularisation on weighted graphs for image and mesh processing. Int. J. Comput. Vis. 2009, 84, 220–236. [Google Scholar] [CrossRef]
  2. Chambrolle, A.; Lions, P. Image recovery via total variation minimization and related problems. Numer. Math. 1997, 76, 167–188. [Google Scholar] [CrossRef]
  3. Rudin, L.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithm. Phys. D 1992, 60, 259–268. [Google Scholar] [CrossRef]
  4. Chen, Y.; Levine, S.; Rao, M. Variable exponent, linear growth functionals in image restoration. SIAM J. Appl. Math. 2006, 66, 1383–1406. [Google Scholar] [CrossRef]
  5. Coren, S.; Girgus, J.S. Principles of perceptual organization and spatial distortion: The gestalt illusions. J. Exp. Psychol. Hum. Percept. Perform. 1980, 6, 404–412. [Google Scholar] [CrossRef] [PubMed]
  6. Ricard, J.; Baskurt, A.; Idrissi, K.; Lavoue, G. Object of interest-based visual navigation, retrieval, and semantic content identification system. Comput. Vis. Image Underst. 2004, 94, 271–294. [Google Scholar]
  7. Bollt, E.M.; Chartrand, R.; Esedoglu, S.; Schultz, P.; Vixie, K.R. Graduated adaptive image denoting local compromise between total variation and isotropic diffusion. Adv. Comput. Math. 2009, 61–85. [Google Scholar] [CrossRef]
  8. Harjuletho, P.; Hästö, P.; Latvala, V. Minimizers of the variable exponents, nonuniformly convex Dirichlet energy. J. Math. Pures Appl. 2008, 89, 174–197. [Google Scholar] [CrossRef]
  9. Musielak, J. Orlicz Spaces and Modular Spaces; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  10. Diening, L.; Harjulehto, P.; Hasta, P.; Ruzicka, M. Lebesgue and Sobolev Spaces with Variable Exponents; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  11. Kovacik, O.; Rakosnik, J. On spaces Lp(x) and Wk,p(x). Czechoslov. Math. J. 1991, 41, 592–618. [Google Scholar]
  12. Brezis, H. Analyse Fonctionnelle Théorie et Applications Masson et Cie 1983. Available online: https://cir.nii.ac.jp/crid/1571135649569314304 (accessed on 10 February 2023).
  13. Elmoataz, A.; Olivier, L.; Bougleux, S. Nonlocal discrete regularisation on weigthed graphs: A framework for image and manifold processing. IEEE Trans. Image Process. 2008, 17, 1047–1060. [Google Scholar] [CrossRef] [PubMed]
  14. Elmoataz, A.; Desquesnes, X.; Lezoray, O. Non-Local Morphological PDEs and p-Laplacian Equation on Graphs With Applications in Image Processing and Machine Learning. IEEE J. Sel. Top. Signal Process. 2012, 6, 764–779. [Google Scholar] [CrossRef]
  15. Ta, V.T.; Elmoataz, A.; Lezoraty, O. Nonlocal PDEs-based morphology on weighted graphs for image and data processing. IEEE Trans. Image Process. 2011, 20, 1504–1516. [Google Scholar] [PubMed]
  16. Zhou, D.; Schölkopf, B. Regularization on Discrete Spaces. In Proceedings of the Pattern Recognition; Kropatsch, W.G., Sablatnig, R., Hanbury, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 361–368. [Google Scholar]
  17. Albarello, L.; Bourgeois, E.; Guyot, J.L. Statistique Descriptive: Un Outil Pour Les Praticiens Chercheurs; De Boeck Supérieur: Louvain-la-Neuve, Belgium, 2010. [Google Scholar]
  18. Frigge, M.; Hoaglin, D.C.; Iglewicz, B. Some implementations of the boxplot. Am. Stat. 1989, 43, 50–54. [Google Scholar]
  19. Martin, D.; Fowlkes, C.; Ta, D.; Malik, J. A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  20. Weber, A.G. The USC-SIPI Image Database Version 5. USC-SIPI Report 1997, 315. Available online: https://cir.nii.ac.jp/crid/1573950399791107200 (accessed on 27 January 2023).
  21. Franzen, R. Kodak Lossless True Color Image Suite. 1999, 4. Available online: http://r0k.us/graphics/kodak (accessed on 27 January 2023).
  22. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  23. Tschumperlé, D. Fast anisotropic smoothing of multi-valued images using curvature-preserving PDE’s. Int. J. Comput. Vis. 2006, 68, 65–82. [Google Scholar] [CrossRef]
Figure 1. Comparison between a degraded and a non-degraded image. Above: BMP, 2160 × 3840, 23.7 MB; below: an image which has been compressed, 2160 × 3840, 4.8 MB.
Figure 1. Comparison between a degraded and a non-degraded image. Above: BMP, 2160 × 3840, 23.7 MB; below: an image which has been compressed, 2160 × 3840, 4.8 MB.
Applsci 13 11600 g001
Figure 2. Effects of parameters p and λ on the Lena image. From left to right, parameter p increases to values { 0.01 ; 1.00 ; 2.00 } . From top to bottom, the lambda increases by taking values { 0.01 ; 0.10 ; 1.00 } .
Figure 2. Effects of parameters p and λ on the Lena image. From left to right, parameter p increases to values { 0.01 ; 1.00 ; 2.00 } . From top to bottom, the lambda increases by taking values { 0.01 ; 0.10 ; 1.00 } .
Applsci 13 11600 g002
Figure 3. Illustration of the interval I 1.5 effects. The (first line) shows the original images, the (second line) shows the result of the edge and noise detection, and the last line shows the result of the proposed classification. More specifically, the pixels that have a gradient value in the interval are shown in red and the pixels with a value outside of this interval are shown in green.
Figure 3. Illustration of the interval I 1.5 effects. The (first line) shows the original images, the (second line) shows the result of the edge and noise detection, and the last line shows the result of the proposed classification. More specifically, the pixels that have a gradient value in the interval are shown in red and the pixels with a value outside of this interval are shown in green.
Applsci 13 11600 g003
Figure 4. Illustration of the adaptive model effects on the Lena image.
Figure 4. Illustration of the adaptive model effects on the Lena image.
Applsci 13 11600 g004
Figure 5. Illustration the effects of the adaptive model. From top to bottom, the original images, the result of applying the Sobel filter, then the result of applying the adaptive algorithm, and finally the result of applying the Sobel filter. The application of the proposed adaptive method reduces the number of noise and contour points for the Lena image from 21,571 pixels to 19,552 pixels (a reduction of 9.35%) and for the Tiger image from 15,600 to 14,374 (a reduction of 7.86%).
Figure 5. Illustration the effects of the adaptive model. From top to bottom, the original images, the result of applying the Sobel filter, then the result of applying the adaptive algorithm, and finally the result of applying the Sobel filter. The application of the proposed adaptive method reduces the number of noise and contour points for the Lena image from 21,571 pixels to 19,552 pixels (a reduction of 9.35%) and for the Tiger image from 15,600 to 14,374 (a reduction of 7.86%).
Applsci 13 11600 g005
Figure 6. Illustration of the images contained in the Berkeley database.
Figure 6. Illustration of the images contained in the Berkeley database.
Applsci 13 11600 g006
Figure 7. Illustration of the images contained in the USC-SIPI databases.
Figure 7. Illustration of the images contained in the USC-SIPI databases.
Applsci 13 11600 g007
Figure 8. Illustration of the images contained in the KODAK database.
Figure 8. Illustration of the images contained in the KODAK database.
Applsci 13 11600 g008
Figure 9. Illustrations of the results obtained by testing 100 images extracted from the Berkeley database with an alteration (a degraded aspect due to, for instance, bad capture or compression) with a ratio of almost 50%. From top to bottom (ad), we note the two best results and the two worst results obtained using the PSNR and DSSIM parameters. From left to right, the original degraded image followed by the result of the application of the Sobel filter, then the result of the application of the adaptive algorithm on the degraded image, and finally, the result of the Sobel filter. In all cases, there is a decreasing number of pixels with a high gradient.
Figure 9. Illustrations of the results obtained by testing 100 images extracted from the Berkeley database with an alteration (a degraded aspect due to, for instance, bad capture or compression) with a ratio of almost 50%. From top to bottom (ad), we note the two best results and the two worst results obtained using the PSNR and DSSIM parameters. From left to right, the original degraded image followed by the result of the application of the Sobel filter, then the result of the application of the adaptive algorithm on the degraded image, and finally, the result of the Sobel filter. In all cases, there is a decreasing number of pixels with a high gradient.
Applsci 13 11600 g009
Figure 10. Illustration of the effects of the use of static λ and dynamic λ on the Lena image. (Left side) λ = 1.0 ; (Right side) λ is dynamic.
Figure 10. Illustration of the effects of the use of static λ and dynamic λ on the Lena image. (Left side) λ = 1.0 ; (Right side) λ is dynamic.
Applsci 13 11600 g010
Figure 11. Results on the evolution of the PSNR according to the quality of compression.
Figure 11. Results on the evolution of the PSNR according to the quality of compression.
Applsci 13 11600 g011
Figure 12. Results on the evolution of the DSSIM according to the quality of the compression.
Figure 12. Results on the evolution of the DSSIM according to the quality of the compression.
Applsci 13 11600 g012
Table 1. Relationship between PSNR and image quality.
Table 1. Relationship between PSNR and image quality.
PSNRImage Quality
30Excellent
24.99Good
20Reasonable
15.05Very noisy
10Unreadable
Table 2. Result of using PSNR on the Lena image (Figure 2) by varying p and λ .
Table 2. Result of using PSNR on the Lena image (Figure 2) by varying p and λ .
PSNRp = 0.01p = 1.00p = 2.00
λ = 0.0128.9829.6126.57
λ = 0.1033.8036.4627.49
λ = 1.0041.4745.8132.83
Table 3. DSSIM parameter values obtained on images of Figure 9. The Sobel columns give the number of points designating the contours as well as the noise.
Table 3. DSSIM parameter values obtained on images of Figure 9. The Sobel columns give the number of points designating the contours as well as the noise.
ImagePSNRDSSIMSobel on OriginalSobel on Adaptive
a36.2475660.00231515,19310,841
b36.0343930.00214919,85017,310
c26.5776960.00227955,11354,321
d26.3640880.00335555,45454,567
Table 4. Statistical results on the PSNR obtained with the adaptive model on the Berkeley database.
Table 4. Statistical results on the PSNR obtained with the adaptive model on the Berkeley database.
AverageMinMaxStandard Deviation
PSNR31.1326.3636.242.19
Table 5. The result of using the DSSIM parameter on the Lena image (Figure 2) by varying p and λ .
Table 5. The result of using the DSSIM parameter on the Lena image (Figure 2) by varying p and λ .
DSSIMp = 0.01p = 1.00p = 2.00
λ = 0.010.070.060.11
λ = 0.100.020.010.09
λ = 1.000.0040.0010.02
Table 6. Illustrations of DSSIM results obtained from 100 images extracted from the Berkeley database. This study was carried out according to the compression ratio in order to simulate degradation, and different values taken for λ were compared to the method used for dynamic λ (bold values represent the best results).
Table 6. Illustrations of DSSIM results obtained from 100 images extracted from the Berkeley database. This study was carried out according to the compression ratio in order to simulate degradation, and different values taken for λ were compared to the method used for dynamic λ (bold values represent the best results).
Compression/ λ 0.10.51.0Proposal Model
10%0.1247493910.1246856240.1265270000.126313566
20%0.0733990660.0672238330.0675830500.067538158
30%0.05902433330.05047845830.0501695250.050176050
40%0.0526764240.04298421660.0423508250.042386891
50%0.0486715330.03829779160.03745219990.0374951749
60%0.0455546830.03463660000.03364357490.033697841
70%0.04233885830.0307230080.029579850.0296397499
80%0.0392994000.02684655830.02555395830.0256166749
90%0.0368226910.0231592160.02154492500.0216251499
Table 7. Illustration of the mean values and standard deviations on DSSIM and PSNR criteria obtained after image processing by our proposed model on the Berkeley (120 images), Kodak (24 images), and USC-SIPI (13 images) databases.
Table 7. Illustration of the mean values and standard deviations on DSSIM and PSNR criteria obtained after image processing by our proposed model on the Berkeley (120 images), Kodak (24 images), and USC-SIPI (13 images) databases.
Berkeley DatabasesKODAK DatabasesUSC-SIPI Databases
DSSIMPSNRDSSIMPSNRDSSIMPSNR
10%0.13 (0.02)26.23 (2.07)0.1 (0.01)28.31 (2.3)0.09 (0.01)30.07 (2.63)
20%0.07 (0.013)28.28 (2.19)0.05 (0.009)30.64 (2.44)0.05 (0.009)32.7 (2.97)
30%0.05 (0.01)29.49 (2.23)0.03 (0.007)32 (2.45)0.04 (0.008)34.14 (3.15)
40%0.04 (0.01)30.37 (2.22)0.03 (0.005)32.9 (2.43)0.03 (0.007)35 (3.19)
50%0.04 (0.009)31.13 (2.19)0.02 (0.005)33.72 (2.39)0.03 (0.007)35.6 (3.18)
60%0.03 (0.009)31.93 (2.14)0.02 (0.004)34.52 (2.33)0.02 (0.007)36.36 (3.16)
70%0.03 (0.008)33.07 (2.03)0.02 (0.004)35.58 (2.22)0.02 (0.007)37.23 (3.08)
80%0.03 (0.008)34.81 (1.84)0.01 (0.003)37.15 (2)0.02 (0.007)38.41 (2.9)
90%0.02 (0.008)38.07 (1.41)0.01 (0.003)39.88 (1.5)0.02 (0.007)40.58 (2.33)
Table 8. Presentation of the average execution times of the adaptive p(x)-Laplacian model on the image databases used in the experiment.
Table 8. Presentation of the average execution times of the adaptive p(x)-Laplacian model on the image databases used in the experiment.
Average Time in Seconds
Berkeley DatabasesKODAK DatabasesUSC-SIPI Databases
10%601.80854.96505.35
20%464.79966.35423.26
30%517.82928.12601.09
40%471.24911.41505.01
50%538.31741.18367.47
60%540.24794.21428.84
70%527.51657.88520.70
80%468.22866.27553.07
90%437.38758.04603.88
Table 9. Results of the evaluation of the application of smoothing methods on the Berkeley, KODAK, USC-SIPI noisy image databases using the Gaussian-distributed additive noise method.
Table 9. Results of the evaluation of the application of smoothing methods on the Berkeley, KODAK, USC-SIPI noisy image databases using the Gaussian-distributed additive noise method.
Gaussian-Distributed Additive Noise
ImagesFiltering MethodDSSIM (sd)PSNR (sd)
BerkeleyProposal (2)0.019 (0.007)42.495 (1.385)
Tchumperle (1)0.016 (0.007)55.810 (2.214)
Gaussian filtering (4)0.069 (0.021)27.028 (2.571)
Peer group filtering (3)0.020 (0.007)37.556 (2.840)
KODAKProposal (2)0.006 (0.003)44.052 (1.085)
Tchumperle (1)0.005 (0.003)54.110 (2.205)
Gaussian filtering (4)0.045 (0.020)29.020 (2.963)
peer Group filtering (3)0.007 (0.003)40.094 (2.988)
USC-SIPIProposal (2)0.014 (0.007)44.559 (1.733)
Tchumperle (1)0.013 (0.007)53.738 (3.184)
Gaussian filtering (4)0.035 (0.019)30.972 (3.031)
Peer group filtering (3)0.015 (0.007)42.382 (3.374)
Table 10. Results of the evaluation of the application of smoothing methods on the Berkeley, KODAK, USC-SIPI noisy image databases using the Poisson-distributed noise generated method.
Table 10. Results of the evaluation of the application of smoothing methods on the Berkeley, KODAK, USC-SIPI noisy image databases using the Poisson-distributed noise generated method.
Poisson-Distributed Noise
ImagesFiltering MethodDSSIM (sd)PSNR (sd)
BerkeleyProposal (2)0.019 (0.007)42.338 (1.364)
Tchumperle (1)0.016 (0.007)51.841 (0.957)
Gaussian filtering (4)0.069 (0.021)27.027 (2.571)
Peer group filtering (3)0.020 (0.007)37.488 (2.792)
KODAKProposal (2)0.007 (0.003)43.908 (1.079)
Tchumperle (1)0.005 (0.003)51.311 (1.205)
Gaussian filtering (4)0.045 (0.020)29.017 (2.963)
Peer group filtering (3)0.007 (0.003)39.948 (2.914)
USC-SIPIProposal (2)0.014 (0.007)44.434 (1.716)
Tchumperle (1)0.013 (0.007)51.266 (1.857)
Gaussian filtering (4)0.036 (0.019)30.971 (3.030)
Peer group filtering (3)0.015 (0.007)42.167 (3.261)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Henry, J.-L.; Nagau, J.; Velin, J.; Moussa, I.-P. Adaptive Smoothing for Visual Improvement of Image Quality via the p(x)-Laplacian Operator Effects of the p(x)-Laplacian Smoothing Operator on Digital Image Restoration: Contribution to an Adaptive Control Criterion. Appl. Sci. 2023, 13, 11600. https://doi.org/10.3390/app132011600

AMA Style

Henry J-L, Nagau J, Velin J, Moussa I-P. Adaptive Smoothing for Visual Improvement of Image Quality via the p(x)-Laplacian Operator Effects of the p(x)-Laplacian Smoothing Operator on Digital Image Restoration: Contribution to an Adaptive Control Criterion. Applied Sciences. 2023; 13(20):11600. https://doi.org/10.3390/app132011600

Chicago/Turabian Style

Henry, Jean-Luc, Jimmy Nagau, Jean Velin, and Issa-Paul Moussa. 2023. "Adaptive Smoothing for Visual Improvement of Image Quality via the p(x)-Laplacian Operator Effects of the p(x)-Laplacian Smoothing Operator on Digital Image Restoration: Contribution to an Adaptive Control Criterion" Applied Sciences 13, no. 20: 11600. https://doi.org/10.3390/app132011600

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop