Next Article in Journal
Generalized Fixed Point Results with Application to Nonlinear Fractional Differential Equations
Next Article in Special Issue
Applications of Microlocal Analysis in Inverse Problems
Previous Article in Journal
A Weak Convergence Self-Adaptive Method for Solving Pseudomonotone Equilibrium Problems in a Real Hilbert Space
Previous Article in Special Issue
Unique Determination of the Shape of a Scattering Screen from a Passive Measurement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Q-Curve and Area Rules for Choosing Heuristic Parameter in Tikhonov Regularization

Institute of Mathematics and Statistics, University of Tartu, 51009 Tartu, Estonia
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(7), 1166; https://doi.org/10.3390/math8071166
Submission received: 8 June 2020 / Revised: 8 July 2020 / Accepted: 12 July 2020 / Published: 16 July 2020
(This article belongs to the Special Issue Inverse and Ill-Posed Problems)

Abstract

:
We consider choice of the regularization parameter in Tikhonov method if the noise level of the data is unknown. One of the best rules for the heuristic parameter choice is the quasi-optimality criterion where the parameter is chosen as the global minimizer of the quasi-optimality function. In some problems this rule fails. We prove that one of the local minimizers of the quasi-optimality function is always a good regularization parameter. For the choice of the proper local minimizer we propose to construct the Q-curve which is the analogue of the L-curve, but on the x-axis we use modified discrepancy instead of discrepancy and on the y-axis the quasi-optimality function instead of the norm of the approximate solution. In the area rule we choose for the regularization parameter such local minimizer of the quasi-optimality function for which the area of the polygon, connecting on Q-curve this minimum point with certain maximum points, is maximal. We also provide a posteriori error estimates of the approximate solution, which allows to check the reliability of the parameter chosen heuristically. Numerical experiments on an extensive set of test problems confirm that the proposed rules give much better results than previous heuristic rules. Results of proposed rules are comparable with results of the discrepancy principle and the monotone error rule, if the last two rules use the exact noise level.

1. Introduction

Let A L ( H , F ) be a linear bounded operator between real Hilbert spaces H, F. We are interested in finding the minimum norm solution u * of the equation
A u = f * , f * R ( A ) ,
where noisy data f F are given instead of the exact data f * . The range R ( A ) may be non-closed and the kernel N ( A ) may be non-trivial, so in general this problem is ill-posed. We consider the solution of the problem A u = f by the Tikhonov method (see [1,2]) where the regularized solutions in cases of exact and inexact data have the corresponding forms
u α + = α I + A * A 1 A * f * , u α = α I + A * A 1 A * f
and α > 0 is the regularization parameter. Using the well-known estimate u α u α + 1 2 α 1 / 2 f f * (see [1,2]) and notations
e ( α ) : = u α u * , e 1 ( α ) : = u α + u * + u α u α + ,
e 2 ( α , f f * ) : = u α + u * + 1 2 α f f * ,
we have the error estimates
e ( α ) e 1 ( α ) e 2 ( α , f f * ) .
We consider choice of the regularization parameter if the noise level for f f * is unknown. The parameter choice rules which do not use the noise level information are called heuristic rules. Well known heuristic rules are the quasi-optimality criterion [3,4,5,6,7,8,9], L-curve rule [10,11], GCV (generalized cross-validation)-rule [12], Hanke–Raus rule [13], Reginska’s rule [14] and its modifications [15]; for other rules see [16,17,18]. Heuristic rules are numerically compared in [4,17,18,19]. The heuristic rules give good results in many problems, but it is not possible to construct heuristic rules guaranteeing convergence u α u * 0 as the noise level goes to zero (see [20]). All heuristic rules may fail in some problems and without additional information about the solution, it is difficult to decide, is the obtained parameter reliable or not.
In the quasi-optimality criterion the parameter α is chosen as the global minimizer of the function ψ Q ( α ) = α d u α d α on certain interval. We propose to choose the parameter from the set L min of local minimizers of this function from certain set Ω of parameters.
We will call the parameter α R in arbitrary rule R as pseudooptimal, if
u α R u * c min α > 0 e 1 ( α )
with relatively small constant c and we show that at least one parameter from set L min has this property. For the choice of proper parameter from the set L min some algorithms were proposed in [21], in the current work we propose other algorithms. We propose to construct the Q-curve which is the analogue of the L-curve [10], but on the x-axis we use modified discrepancy instead of discrepancy and on the y-axis function ψ Q ( α ) instead of u α . For finding proper local minimizer of the function ψ Q ( α ) , we propose the area rules on the Q-curve. The idea of proposed rules is that we form, for every minimizer of the function ψ Q ( α ) , a certain function which approximates the error of the approximate solution and has one minimizer; we choose for the regularization parameter such local minimizer of ψ Q ( α ) , for which the area of the polygon, connecting this minimum point with certain maximum points, is maximal.
The plan of this paper is as follows. In Section 2 we consider known rules for choice of the regularization parameter, both in case of known and unknown noise level. In Section 3, we prove that the set L min contains at least one pseudooptimal parameter. In Section 4, information about used test problems (mainly from [11,22], but also from [11,23,24,25,26]) and numerical experiments is given. In Section 5, we consider the Q-curve and area rule, in Section 6 further developments of the area rule. These algorithms are also illustrated by the results of numerical experiments.

2. Rules for the Choice of the Regularization Parameter

2.1. Parameter Choice in the Case of Known Noise Level

In the case of known noise level δ , f f * δ we use one of the so-called δ -rules, where certain functional d ( α ) and constant b b 0 ( b 0 depends on d ( α ) ) are chosen and such regularization parameter α ( δ ) is chosen which satisfies d ( α ) = b δ .
(1) Discrepancy principle (DP) [2,27]:
d D ( α ) : = A u α f = b δ , b 1 .
(2) Modified discrepancy principle (Raus–Gfrerer rule) [28,29]:
d MD ( α ) : = B α A u α f = b δ , B α : = α 1 / 2 α I + A A * 1 / 2 , b 1 .
(3) Monotone error rule (ME-rule) [30,31]:
d ME ( α ) : = B α A u α f 2 B α 2 A u α f = δ .
The name of this rule is justified by the fact that the chosen parameter α ME satisfies
d d α u α u * > 0 α > α ME .
Therefore α ME α opt : = argmin u α u * .
(4) Monotone error rule with post-estimation (MEe-rule) [4,18,32,33,34,35]. The inequality α ME α opt suggests to use somewhat smaller parameter than α ME . Extensive numerical experiments suggest to compute α ME and to use the post-estimated parameter α MEe : = 0.4 α ME . Then typically u α MEe u * / u α ME u * ( 0.7 , 0.9 ) . If the exact noise level is known, this MEe-rule gives typically the best results from all δ -rules.
(5) Rule R1 [36]. Let b > 2 3 3 . Choose α ( δ ) as the smallest solution of the equation
d R 1 ( α ( δ ) ) : = α 1 / 2 A * B α 2 A u α f = b δ .
Note that this equation can be rewritten using the 2-iterated Tikhonov approximation u 2 , α :
B α 2 A u α f = A u 2 , α f , u 2 , α = α I + A * A 1 ( α u α + A * f ) .
The last four rules are weakly quasioptimal rules (see [37]) for the Tikhonov method: if f f * δ , then u α ( δ ) u * C ( b ) inf α > 0 e 2 ( α , δ ) (see (3)). The rules for the parameter choice in case of approximately given noise level are proposed and analyzed in [18,32,33,34].

2.2. Parameter Choice in the Case of Unknown Noise Level

A classical heuristic rule is the quasi-optimality criterion. In the Tikhonov method, it chooses α = α Q or α = α QD as the global minimizer of corresponding functions
ψ Q ( α ) = α d u α d α = α 1 A * B α 2 A u α f = α A * α I + A A * 2 f = u 2 , α u α ,
ψ QD ( α ) = 1 q 1 u α u q α , 0 < q < 1 .
The Hanke–Raus rule finds parameter α = α H R as the global minimizer of the function
ψ HR ( α ) = α 1 / 2 B α A u α f .
In practice, often the L-curve is used. The L-curve is the log-log-plot of u α versus A u α f . The points A u α f , u α have often a shape similar to the letter L and the parameter α L which corresponds to the corner point is often a good parameter. In the literature several concrete rules for choice of the corner point are proposed. In [14], a parameter is chosen as the global minimizer of the function
ψ RE ( α ) = A u α f u α τ , τ 1
(below we use this rule with τ = 1 ). Another rule for choice of the corner point is the maximum curvature method ([38,39]), where such parameter α is chosen for which the curvature of the L-curve as the function
ψ MC ( α ) = 2 ρ ^ ξ ^ ρ ^ ξ ^ ( ( ρ ^ ) 2 + ( ξ ^ ) 2 ) 3 / 2
is maximal. Here ρ ^ , ξ ^ , ρ ^ , ξ ^ are first and second order derivatives of functions log d D ( α ) and log u α .
We propose also a new heuristic rule, where the global minimizer of the function
ψ WQ ( α ) = d MD ( α ) ψ Q ( α )
is chosen for the parameter. We call this rule as the weighted quasioptimality criterion.
In the following we will find the regularization parameter from the set of parameters
Ω = α j : α j = q α j 1 , j = 1 , 2 , , N , 0 < q < 1 ,
where α 0 , q , α N are given. If in the discretized problem the minimal eigenvalue λ min of the matrix A T A is larger than α N , the heuristic rules above often choose parameter α N , which is generally not a good parameter. The works [6,7,8] propose to search the global minimum of the function ψ Q ( α ) in the interval [ max α N , λ min , α 0 ] .
Definition 1.
We say that the discretized problem A u = f does not need regularization if
e 1 ( λ min ) 2 min α Ω , α λ min e 1 ( α ) .
If the discretized problem does not need regularization then α = 0 or α = α N is the proper parameter while for α λ m i n we have
u α u * e 1 ( α ) = u α + u * + ( α I + A * A ) 1 A * ( f f * ) u λ m i n + u * + 2 ( λ m i n I + A * A ) 1 A * ( f f * ) 2 e 1 ( λ m i n ) 4 min α Ω , α λ m i n e 1 ( α ) .
Searching the parameter from the interval [ max α N , λ min , α 0 ] means the a priori assumption that the discretized problem needs regularization. Note that if λ min > α N , then in general it is not possible to decide (without additional information about solution or about noise of the data) whether the discretized problem needs regularization or not.

3. Local Minimum Points of the Function ψ Q ( α )

In the following we investigate the function ψ Q ( α ) in (5) and show that at least one local minimizer of this function is the pseudooptimal parameter. We need some preliminary results.
Lemma 1.
The functions ψ Q ( α ) , ψ QD ( α ) satisfy for each α > 0 the estimates
ψ Q ( α ) e 1 ( α ) ,
ψ QD ( α ) q 1 e 1 ( α ) ,
ψ Q ( α ) ψ QD ( α ) q 1 ψ Q ( q α ) .
Proof. 
Using relations f = A u * + f f * ,
u α u q α = q 1 α α I + A * A 1 q α I + A * A 1 A * f ,
A * A α I + A * A 1 1 , α α I + A * A 1 1 ,
we have
ψ Q ( α ) = α A * α I + A A * 2 f = α ( α I + A * A ) 2 A * f α A * A α I + A * A 2 u * + α α I + A * A 2 A * ( f f * ) α ( α I + A * A ) 1 u * + ( α I + A * A ) 1 A * ( f f * ) = e 1 ( α ) , ψ QD ( α ) α A * A ( q α I + A * A ) 1 ) ( α I + A * A ) 1 ) u * + α ( q α I + A * A ) 1 ) ( α I + A * A ) 1 ) A * ( f f * ) q 1 e 1 ( α ) , ψ Q ( α ) = α α I + A * A 2 A * f α α I + A * A 1 q α I + A * A 1 A * f = ψ QD ( α ) α q α I + A * A 2 A * f = q 1 ψ Q ( q α ) .
 ☐
Remark 1.
Note that lim α ψ Q ( α ) = 0 , but lim α e 1 ( α ) = u * . Therefore in the case of too large α 0 this α 0 may be a global (or local) minimizer of the function ψ Q ( α ) . The scaling argument suggests: by multiplying the equation A * A u = A * f by some constant, it is also necessary to multiply α by this constant in the Tikhonov method. Therefore the parameter α 0 should be proportional to A * A . We recommend to take α 0 = c A * A , c 1 or to minimize the function ψ ¯ Q ( α ) : = ( 1 + α / A * A ) ψ Q ( α ) instead of ψ Q ( α ) . Due to the equality lim α 0 ( 1 + α / A * A ) = 1 , the function ψ ¯ Q ( α ) approximately satisfies (9) for small α.
In the following we define the local minimum points of the function ψ Q ( α ) on the set Ω (see (8)).
Definition 2.
We say that the parameter α k , 0 k N 1 is the local minimum point of the sequence ψ Q ( α k ) , if ψ Q ( α k ) < ψ Q ( α k + 1 ) and in case k > 0 there exists index j 1 such that ψ Q ( α k ) = ψ Q ( α k 1 ) = = ψ Q ( α k j + 1 ) < ψ Q ( α k j ) . The parameter α N is the local minimum point if there exists index j 1 so that
ψ Q ( α N ) = ψ Q ( α N 1 ) = = ψ Q ( α N j + 1 ) < ψ Q ( α N j ) .
Denote the local minimum points by m k , k = 1 , , K (K is the number of minimum points) and corresponding set by L min = m k : m 1 > m 2 > > m K .
Definition 3.
The parameter α k , 0 < k < N is the local maximum point of the sequence ψ Q ( α k ) if ψ Q ( α k ) > ψ Q ( α k + 1 ) and there exists index j 1 so that
ψ Q ( α k ) = ψ Q ( α k 1 ) = = ψ Q ( α k j + 1 ) > ψ Q ( α k j ) .
We denote by M k the local maximum point between the local minimum points m k + 1 and m k , 1 k K 1 . Denote M 0 = α 0 , M K = α N . Then by the construction
M K m K < M K 1 < < m 2 < M 1 < m 1 M 0 .
Theorem 1.
The following estimates hold for the local minimizers of the function ψ Q ( α ) .
1 
For arbitrary α 0 , α N we have
min α L min u α u * q 1 C min α N α α 0 e 1 ( α ) ,
C : = 1 + max 1 k K max α j Ω , M k α j M k 1 T m k , α j 1 + c q ln α 0 α N , T ( α , β ) : = u α u β ψ Q ( β ) .
2 
If α 0 = A * A , α N = α 0 f f * 2 u * 2 , then
min α L min u α u * q 1 ( 1 + 2 max { 1 , c q ln f f * 2 A u * } ) min α > 0 e 2 ( α , f f * ) ,
where c q : = q 1 1 / ln q 1 1   i f   q 1 .
Moreover, if u * = A p v , v ρ , p > 0 , where | A | : = ( A * A ) 1 / 2 , then
min α L min u α u * c p , q ρ 1 p + 1 ln f f * f f * p p + 1 , 0 < p 2 .
Proof. 
For arbitrary parameters α 0 , β 0 the inequalities
u α u * u α u β + u β u * T ( α , β ) ψ Q ( β ) + e 1 ( β )
and (9) lead to the estimate
u α u * 1 + T ( α , β ) e 1 ( β ) .
It is easy to see that
min α j Ω e 1 ( α j ) q 1 min α N α α 0 e 1 ( α ) ,
while in case q α α α we have e 1 α q 1 e 1 α .
Let α j * = α 0 q j * be the global minimizer of the function e 1 ( α ) on the set of the parameters Ω . Then, α j * [ M k , M k 1 ] for some k , 1 k K and this k defines index m with m k = α m . From (14) we get the estimate
u m k u * 1 + T ( m k , α j * ) e 1 ( α j * ) 1 + min M k α j M k 1 T ( m k , α j ) min α j Ω e 1 ( α j ) ,
which together with (15) gives also the estimate (11).
Now we show that C 1 + c q ln α 0 α N . If m k α j M k 1 , then using Lemma 1 and equality (6) we get
u α m u α j j i m 1 u i u i + 1 q 1 ( 1 q ) j i m 1 ψ Q ( α i + 1 ) .
Due to the inequalities ψ Q ( α i + 1 ) ψ Q ( α j ) for all i , j i m 1 we have
T ( m k , α j ) = u α m u α j ψ Q ( α j ) q 1 ( 1 q ) j i m 1 ψ Q ( α i + 1 ) ψ Q ( α j ) ( q 1 1 ) ( m j ) ( q 1 1 ) N = ( q 1 1 ) ln q 1 ln α 0 α N = c q ln α 0 α N .
If M k α j m k , then analogous estimation of T ( m k , α j ) gives the same result.
Now we prove the estimate (12). For the global minimum point α * of the function e 2 ( α , f f * ) the inequality α * α N holds, while for α < α N we have
e 2 ( α * ) u * = f f * / ( 2 α N ) f f * / ( 2 α ) < e 2 ( α ) .
In the case α * α 0 we get similarly as in the proof of the estimate (11) that
min α L min u α u * q 1 ( 1 + c q ln α 0 α N ) min α > 0 e 2 ( α , f f * ) ;
due to ln α 0 α N = ln f f * / f * , the estimate (12) holds. Consider the case α * > α 0 . Then,
e 2 ( α * , f f * ) u α 0 + u * α 0 α 0 + A * A u * = u * 2
and for each local minimum point m k , α N m k α 0 the inequalities
u m k u * e 2 ( m k , f f * ) u α 0 + u * + 0.5 α N 1 / 2 f f * = u α 0 + u * + u * 3 u α 0 + u * 3 e 2 ( α * , f f * )
hold. Therefore the inequality (12) holds also in this case.
For source-like solution u * = A p v , v ρ , p > 0 the error estimate
min α N α α 0 e 1 ( α ) c p ρ 1 / ( p + 1 ) f f * p / ( p + 1 ) , 0 < p 2
is well-known (see [1,2]) and the estimate (13) follows immediately from (12). ☐
Remark 2.
Theorem 3 holds also in the case if the equation A u = f * has only the quasisolution, i.e., in the case f * R ( A ) , Q f * R ( A ) , where Q is the orthoprojector F R ( A ) ¯ .
Remark 3.
The inequality (11) holds also in the case if the noise of the function f is not finite but min α N α α 0 e 1 ( α ) is finite (this holds if A * ( f f * ) is finite).
Remark 4.
Use of the inequality (10) enables to prove the analogue of Theorem 3 for set L min of local minimizers of the function ψ QD ( α ) : then the inequality (11) holds, where T ( α , β ) = q 1 u α u β ψ Q D ( β ) .
In choice of the regularization parameter we may exclude from the observation some local minimizers. It is natural to assume that α N is so small that
d MD ( α N ) ( 1 + ϵ ) f f *
with small ϵ > 0 . Then the following theorem holds.
Theorem 2.
Let (16) hold. Let m k 0 be some local minimizer in L min . Then
min α L min , α m k 0 u α u * max { q 1 C 1 min α 0 e 1 ( α ) , C 2 ( b , ϵ ) min α 0 e 2 ( α , f f * ) } ,
where b = d MD ( m k 0 ) / d MD ( α N ) 1 , C 2 ( b , ϵ ) : = b ( 1 + ϵ ) + 2 and
C 1 : = 1 + max 1 k k 0 max α j Ω , M k α j M k 1 T m k , α j 1 + c q ln α 0 m k 0 .
Proof. 
Let α i * , i = 1 , 2 be global minimizers of the functions e 1 ( α ) and e 2 ( α , f f * ) , respectively. We consider separately 3 cases. If m k 0 α 1 * we get similarly to the proof of Theorem 3 the estimate
min α L min , α m k 0 u α u * q 1 C 1 min α N α α 0 e 1 ( α ) .
If α 1 * m k 0 < α 2 * , we estimate
u m k 0 u * u α 2 * + u * + f f * 2 α 1 * min α 0 e 2 ( α , f f * ) + min α 0 e 1 ( α ) .
If α 1 * m k 0 and α 2 * m k 0 , we have
B m k 0 A u m k 0 f b d MD ( α N ) b ( 1 + ϵ ) f f *
and now we can prove analogically to the proof of the weak quasioptimality of the modified discrepancy principle ([37]) that under assumption α 2 * m k 0 the error estimate
u m k 0 u * C 2 ( b , ϵ ) min α 0 e 2 ( α , f f * )
holds. Now the assertion 1 of Theorem 7 follows from the inequalities (17)–(19). ☐

4. On Test Problems and Numerical Experiments

We made numerical experiments for the local minimizers of the function ψ Q ( α ) using three sets of test problems. The first set contains 10 well-known test problems from Regularization Toolbox [11] and the following 6 Fredholm integral equations of the first kind (discretized by the midpoint quadrature formula)
a b K ( t , s ) u ( s ) d s = f ( t ) , c t d .
  • Groetsch1 [24]: K ( t , s ) = t exp ( t 2 / ( 4 s ) ) 2 π s 3 / 2 , 0 s , t 100 , u ( s ) = 40 + 5 cos ( ( 100 s ) / 5 ) + 2.5 cos ( 2 ( 100 s ) / 2.5 ) + 1.25 cos ( 4 ( 100 s ) / 2 ) ;
  • Groetsch2 [24]: K ( t , s ) = 1 k 100 sin ( k t ) sin ( k s ) k , 0 s , t π , u ( s ) = s ( π s ) ;
  • Indram [25]: K ( t , s ) = e s t , 0 s , t 1 , u ( s ) = s , f ( t ) = 1 ( t + 1 ) e t t 2 ;
  • Ursell [11]: K ( t , s ) = 1 1 + s + t , 0 s , t 1 , u ( s ) = s ( 1 s ) , f ( t ) = 3 + 2 t 2 + ( 2 + 3 t + t 2 ) log 1 + t 2 + t ;
  • Waswaz [26]: K ( t , s ) = cos ( t s ) , 0 s , t π , u ( s ) = cos ( s ) , f ( t ) = π 2 cos ( t ) ;
  • Baker [23]: K ( t , s ) = e s t , 0 s , t 1 , u ( s ) = e s , f ( t ) = e t + 1 1 t + 1 .
The second set of test problems are well-known problems from [22]: Gauss, Hilbert, Lotkin, Moler, Pascal, Prolate. As in [22], we combined these six n × n matrices with 6 solution vectors x i = 1 , x i = i / n , x i = ( ( i [ n / 2 ] ) / [ n / 2 ] ) 2 , x i = sin ( 2 π ( i 1 ) / n ) , x i = i / n + 1 / 4 sin ( 2 π ( i 1 ) / n ) , x i = 0 if i [ n / 2 ] and x i = 1 if i > [ n / 2 ] . For getting the third set of test problems we combined the matrices of the first set of test problems with 6 solutions of the second set of test problems.
Numerical experiments showed that performance of different rules depends essentially on eigenvalues of the matrix A T A . We characterize these eigenvalues via three indicators: the value of minimal eigenvalue λ min , by the value N 1 , showing the number of eigenvalues less than α N and by the value Λ , characterizing the density of location of eigenvalues on the interval [ max ( α N , λ min ) , 1 ] . More precisely, let the eigenvalues of the matrix A T A be λ 1 λ 2 λ n = λ min . Then the value of Λ is found by the formula Λ = max λ k > max ( α N , λ n ) λ k / λ k + 1 . We characterize the smoothness of the solution by the value
p 1 = log min α e 2 ( α , f f * ) log u * log f f * log f * ,
where f f * = 10 6 . Table 1 contains the results of characteristics of the matrix A T A in case n = 100 , α N = 10 18 .
In all tests the discretization parameters n { 60 , 80 , 100 , 120 , 140 , 160 , 180 } were used. We present the results of numerical experiments in tables for n = 100 . Since the performance of rules generally depends on the smoothness p of the exact solution in (1), we complemented the standard solutions u * of (now discrete) test problems with smoothened solutions | A | p u * , p = 2 computing the right-hand side as A ( | A | p u * ) . Results for p = 2 are given in Table 2, in all other tables and figures p = 0 . After discretization, all problems were scaled (normalized) in such a way that the norms of the operator and the right-hand side were 1. All norms here and in the text below are Euclidean norms. On the base of exact data f * we formed the noisy data f, where f f * has values 10 1 , 10 2 , , 10 6 , noise f f * has normal distribution and the components of the noise were uncorrelated. We generated 20 noise vectors and used these vectors in all problems. We search the regularization parameter from the set Ω , where α 0 = 1 , q = 0.95 and N is chosen so that α N 10 18 > α N + 1 . To guarantee that calculation errors do not influence essentially the numerical results, calculations were performed on geometrical sequence of decreasing α -s and finished for largest α with d MD ( q α ) > d MD ( α ) , while theoretically the function d MD ( α ) is monotonically increasing. Actually this precautionary measure was needed only in problem Groetsch2, calculations on α > α N were finished only in this problem. Since in model equations the exact solution is known, it is possible to find the regularization parameter α * , which gives the smallest error on the set Ω . For every rule R the error ratio
E = u α R u * u α * u * = u α R u * min α Ω u α u *
describes the performance of the rule R on this particular problem. To compare the rules or to present their properties, the following tables show averages and maximums of these error ratios over various parameters of the data set (problems, noise levels δ ). We say that the heuristic rule fails if the error ratio E > 100 . In addition to the error ratio E we present in some cases also error ratios
E 1 = u α R u * min α Ω e 1 ( α ) , E 2 = u α R u * min α Ω e 2 ( α ) .
The results of numerical experiments for local minimizers α L min of the function ψ Q ( α ) are given in Table 3. For comparison, the results of δ -rules with δ = f f * are presented in the columns 2–4. Columns 5 and 6 contain respectively the averages and maximums of error ratios E for the best local minimizer α L min . The results show that for many problems the Tikhonov approximation with the best local minimizer α L min is even more accurate than with the δ -rules parameters α ME , α MEe or α DP . Table 1 and Table 3 show also that for rules ME and MEe the average error ratio E may be relatively large for problems where Λ is large and most of eigenvalues are smaller than α N , while in this case min α Ω e ( α ) may be essentially smaller than min α Ω e 2 ( α , f * f ) . In these problems the discrepancy principle gives better parameter than ME and MEe rules. The average error in ME rule was largest in the problem Waswaz2, but error ratio E2 there is still under 1. This is due to the fact that for the tasks where the size of a Λ (defined on previous page) is large, the minimal error may be significantly smaller than min e 2 ( α ) .
Columns 7 and 8 contain the averages and maximums of cardinalities | L min | of sets L min (number of elements of these sets). Note that the number of local minimizers depends on the parameter q (for smaller q the number of local minimizers is smaller) and on the length of minimization interval determined by the parameters α N , α 0 . The number of local minimizers is smaller also for larger noise level. Columns 9 and 10 contain the averages and maximums of values of constant C in the a posteriori error estimate (11). The value of C and error estimate (11) allow to assert, that in our test problems the choice of α as the best local minimizer in L min guarantees that the error of the Tikhonov approximation has the same order as min α N α α 0 e 1 ( α ) . Note that over all test problems the maximum of error ratio E 1 for the best local minimizer in L min and for the discrepancy principle were 1.93 and 9.90, respectively. This confirms the result of Theorem 3 that at least one minimizer of the function ψ Q ( α ) is a good regularization parameter.

5. Q-Curve and Triangle Area Rule for Choosing Heuristic Regularization Parameter

We showed in the previous section that at least one local minimizer of the function ψ Q ( α ) is pseudooptimal parameter and we may omit small local minimizers α , for which d MD ( α ) is only slightly larger than d MD ( α N ) . We propose to construct for the parameter choice the Q-curve. The Q-curve figure uses the log-log scale with functions d MD ( α ) and ψ Q ( α ) on the x-axis and y-axis, respectively. The Q-curve can be considered as the analogue of the L-curve, where functions A u α f and u α = α 1 A * ( A u α f ) are replaced by functions B α ( A u α f ) and α 1 A * B α 2 ( A u α f ) (see (4)), respectively. We denote d ˜ MD ( α ) : = log 10 d MD ( α ) , ψ ˜ Q ( α ) : = log 10 ψ Q ( α ) . For many problems the curve ( d ˜ MD ( α ) , ψ ˜ Q ( α ) ) (or a part of this) has the form of letter L or V and we choose the minimizer at the “corner” point of L or V. We use the common logarithm instead of the natural logarithm, while then the Q-curve allows easier estimation of the supposed value of the noise level. In Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 n = 100 is used, in Figure 9 and Figure 10 n = 60 . In Figure 1, Figure 2, Figure 3 and Figure 4 the L-curves and Q-curves are compared for two problems, the global minimizer α o p t of the function e 1 ( α ) is also presented. Note that in problem Baart λ min < α N and in problem Deriv2 λ min > α N .
In most cases one can see on the Q-curve only one clear corner area with one local minimizer and then we take this local minimizer as corner point. If the corner area contains several local minimizers, we recommend to choose such local minimizer, for which the sum of coordinates of the corresponding point on the Q-curve is minimal. If Q-curve has several corner areas we recommend to use the very right of them. Actually, it is useful to present in the parameter choice besides the figures for every local minimizer m k of the function ψ Q ( α ) also the coordinates of the point ( d ˜ MD ( m k ) , ψ ˜ Q ( m k ) ) and the sums of the coordinates.
For finding the proper local minimizer of the function ψ Q ( α ) we present now a rule which works well for all test problems from set 1. The idea of the rule is to search the proper local minimizer m k constructing certain triangles on the Q-curve and finding which of them has the maximal area. For every parameter α corresponds a point P ( α ) on the Q-curve with the corresponding coordinates d ˜ MD ( α ) , ψ ˜ Q ( α ) . For every local minimizer m k of the function ψ Q ( α ) corresponds a triangle T ( k , r ( k ) , l ( k ) ) with the vertices P ( m k ) , P ( M r ( k ) ) and P ( M l ( k ) ) on the Q-curve, where indices r ( k ) and l ( k ) correspond to the largest local maximums of the function ψ Q ( α ) on two sides of the local minimum m k :
ψ Q ( M r ( k ) ) = max j < k ψ Q ( M j ) , ψ Q ( M l ( k ) ) = max j k ψ Q ( M j ) .
Triangle area rule (TA-rule). We choose for the regularization parameter such local minimizer m k of the function ψ Q ( α ) for which the area of the triangle T ( k , r ( k ) , l ( k ) ) is the largest.
Figure 5, Figure 6 and Figure 7 show examples of Q-curves and the triangle T ( k , r ( k ) , l ( k ) ) with the largest area, the TA-rule chooses for the regularization parameter a corresponding minimizer. In some problems the function ψ Q ( α ) may be monotonically increasing as for the problem Groetsch2 (Figure 8), then function ψ Q ( α ) has only one local minimizer α N . Then vertices P ( M l ( k ) ) and P ( m k ) coincide and the area of the corresponding triangle is zero. Then this is the only triangle, the TA-rule chooses for the regularization parameter α N . The results of the numerical experiments for test set 1 ( n = 100 ) for the TA-rule and some other rules (see Section 2.2) are given in Table 4 and Table 5. These results show that the TA-rule works well in all these test problems, the accuracy is comparable with δ -rules (see Table 3), but previous heuristic rules fail in some problems. Note that average of the error ratio increases for decreasing noise level. For example, for f f * { 10 1 , 10 2 , 10 5 , 10 6 } corresponding error ratios E were 1.47, 1.49, 1.78 and 2.08 respectively.
Let us comment on other heuristic rules. The accuracy of the quasi-optimality criterion is for many problems the same as for the TA-rule, but this rule fails in problem Heat. A characteristic feature of the problem Heat is that location of the eigenvalues in the interval [ α N , 1 ] is sparse and only some eigenvalues are smaller than α N (see Table 1). The weighted quasioptimality criterion behaves in a similar way as the quasioptimality criterion, but is more accurate in problems where λ min α N ; if λ min > α N , the quasioptimality criterion is more accurate. The rule of Hanke–Raus may fail in test problems with large Λ and for other problems the error of the approximate solution is in most problems approximately two times larger than for parameter chosen by the quasi-optimality principle. The problem in this rule is that it chooses too large parameter compared with the optimal parameter. However, HR-rule is stable in the sense that the largest error ratio E2 is relatively small in all considered test problems. Reginska’s rule may fail in many problems but it has the advantage that it works better than other previous rules if the noise level is large. The Reginska’s rule did not fail in case f f * 10 3 and has average of error ratios of all problems E = 2.24 and E = 2.80 in cases f f * = 10 1 and f f * = 10 2 , respectively. Advantage of the maximum curvature rule is the small percentage of failures compared with other previous rules.
Distribution of error ratios E in Table 5 show also that in test problems set 1 the TA-rule is the most accurate rule from all considered rules.
Note that the figure of the Q-curve enables to estimate the reliability of the chosen parameter. If the Q-curve has only one corner, then the chosen parameter is quasioptimal with small constant C, if λ min < α N , but in case λ min α N it is quasioptimal under assumption that the problem needs regularization.

6. Further Developments of the Area Rule

The TA-rule may fail for problems which do not need regularization, if function ψ Q ( α ) is not monotonically increasing. In this case, the TA rule selects the parameter α λ min , but the parameter α < λ min would be better. For example, the TA-rule fails for matrix Moler in some cases. Let us consider now the question, in which cases the regularization parameter α N is good. If function ψ Q ( α ) is monotonically increasing then function ψ Q ( α ) has only one local minimizer m 1 = M 1 = α N and then for the parameter α N we have the error estimate
u α N u * q 1 ( 1 + T ( α N ) ) min α N α α 0 e 1 ( α ) ,
where the value of T ( α N ) = max α j Ω , α N α j α 0 T α N , α j c q ln ( α 0 α N ) (see Theorem 3) can be computed a posteriori and this value is the smaller the faster the function ψ Q ( α ) increases. We can take α N for the regularization parameter also in the case if the condition
ψ Q ( α ) ψ Q ( α ) c 0 α , α Ω , α N α < α α 0
holds while one can show similarly to the proof of Theorem 3 that T ( α N ) c 0 c q ln ( α 0 α n ) and the error of the regularized solution is small. For the problems which do not need regularization we can improve the performance of the TA-rule searching the proper local minimizer smaller or equal than α HQ : = max { α HR , α Q } , where α HR , α Q are the global minimizers of functions ψ HR ( α ) and ψ Q ( α ) , respectively on the interval [ max α N , λ min , α 0 ] .
These ideas enable to formulate the following upgraded version of the TA-rule.
Triangle area rule 2 (TA-2-rule). We fix a constant c 0 , 1 c 0 2 . If condition (20) holds, we choose the parameter α N . Otherwise choose for the regularization parameter such local minimizer m k α HQ of the function ψ Q ( α ) for which the area of triangle T ( k , r ( k ) , l ( k ) ) is largest.
Results of numerical experiments for the rule TA-2 with the discretization parameter n = 100 and problem sets 1–3 are given in Table 6 and Table 7 (columns 2 and 3). The results show that the rule TA-2 works well in all considered testsets 1–3. However, the rule TA-2 may fail in some other problems which do not need regularization. Such example is the problem with matrix moler and solution x i = sin ( 12 π ( i 1 ) / n ) , where the rule TA-2 fails if the noise level is below 10 4 ; but in this case all other considered heuristic rules fail too.
The rules TA and TA-2 fail in problem Heat in some cases for the discretization parameter n = 60 . Figure 9 and Figure 10 show the form of the Q-curve in problem Heat with n = 60 . The function ψ Q ( α ) has two local minimizers with corresponding points P ( m 1 ) and P ( m 2 ) on the Q-curve and 3 local maximum points P ( M k ) , k = 0 , 1 , 2 . On Figure 9 the Rule TA-2 chooses the local minimizer corresponding to the point P ( m 2 ) , but then the error ratios are large: E = 20.3 , E 2 = 18.34 .
In the following we consider methods which work well also in this problem. Let g [ α 1 , α 2 ] ( α ) , α [ α 1 , α 2 ] be the parametric representation of the straight line segment connecting points P ( α 1 ) ja P ( α 2 ) , thus
g [ α 1 , α 2 ] ( α ) = ψ ˜ Q ( α 1 ) + β ( d ˜ MD ( α ) d ˜ MD ( α 1 ) ) , β = ψ ˜ Q ( α 2 ) ψ ˜ Q ( α 1 ) d ˜ MD ( α 2 ) d ˜ MD ( α 1 ) .
Let g [ α 1 , α 2 , , α k ] ( α ) , k > 2 be the parametric representation of the broken line connecting points P ( α 1 ) , P ( α 2 ) , , P ( α k ) , thus
g [ α 1 , α 2 , , α k ] ( α ) = g [ α j , α j + 1 ] ( α ) , α j α α j + 1 , 1 j k 1 .
In the triangle rule certain points P ( M l ( k ) ) , P ( m k ) , P ( M r ( k ) ) are connected by the broken line t 1 ( α ) = g [ M l ( k ) , m k , M r ( k ) ] ( α ) which approximates the error function e ˜ 1 ( α ) = log 10 ( e 1 ( α ) ) well, if m k is the “right” local minimizer. By construction of the function t 1 ( α ) we use only 3 points on the Q-curve. We will get a more stable rule if the form of the Q-curve has more influence to the construction of approximates to the error function e ˜ 1 ( α ) . Let { i ( 1 ) , i ( 2 ) , , i ( n 1 ) } and { j ( 1 ) , j ( 2 ) , , j ( n 2 ) } be the largest sets of the indices, satisfying the inequalities
k i ( 1 ) < i ( 2 ) < < i ( n 1 ) K , ψ Q ( M i ( 1 ) ) ψ Q ( M i ( 2 ) ) ψ Q ( M i ( n 1 ) ) ,
k > j ( 1 ) > j ( 2 ) > > j ( n 2 ) 0 , ψ Q ( M j ( 1 ) ) ψ Q ( M j ( 2 ) ) ψ Q ( M j ( n 2 ) ) .
It is easy to see that i ( n 1 ) = l ( k ) and j ( n 2 ) = r ( k ) . For approximating the error function e ˜ 1 ( α ) we propose to connect points P ( M i ( n 1 ) ) , , P ( M i ( 1 ) ) , P ( m k ) , P ( M j ( 1 ) ) , , P ( M j ( n 2 ) ) by the broken line t 2 ( α ) = g [ M i ( n 1 ) , , M i ( 1 ) , m k , M j ( 1 ) , , M j ( n 2 ) ] ( α ) and to find for every m k the area S 2 ( k ) of the polygon surrounded by the lines T 2 ( α ) = max { t 2 ( α ) , g [ M i ( n 1 ) , M j ( n 2 ) ] ( α ) } and t 2 ( α ) . The second possibility is to approximate the error function e ˜ 1 ( α ) by the curve t 3 ( α ) = max { t 2 ( α ) , ψ ˜ Q ( α ) } and to find S 3 ( k ) as the area of the polygon surrounded by the broken lines t 3 ( α ) and curve T 3 ( α ) = max { T 2 ( α ) , ψ ˜ Q ( α ) } . Note that functions t i ( α ) , i = 1 , 2 , 3 are monotonically increasing if α > m k , and monotonically decreasing if α < m k .
Area rules 2 and 3. We fix a constant c 0 , 1 c 0 2 . First we choose the local minimizer m k α H Q , for which the area S i ( k ) , i { 2 , 3 } is largest. We take for the regularization parameter the smallest m k 0 m k , satisfying condition (compare with (20))
ψ Q ( α ) ψ Q ( α ) c 0 α , α Ω , m k 0 α < α m k .
Let us consider Figure 9. The reason for the failure of the triangle rule is, that for the local minimizer m ( 2 ) the broken line g [ M ( 0 ) , m ( 2 ) , M ( 2 ) ] ( α ) do not approximate well the function e ˜ 1 ( α ) , while point M ( 1 ) is located above the interval [ m ( 2 ) , M ( 0 ) ] . Here the function e ˜ 1 ( α ) is better approximated by the broken line g [ M ( 0 ) , M ( 1 ) , m ( 2 ) , M ( 2 ) ] ( α ) , see Figure 10. For the local minimizer m ( 1 ) , we approximate function e ˜ 1 ( α ) by the broken line g [ M ( 0 ) , m ( 1 ) , M ( 1 ) , M ( 2 ) ] ( α ) and due to the inequality S 2 ( 1 ) > S 2 ( 2 ) the area rule 2 chooses m 1 for the regularization parameter, then E = 1.05 .
The area rules 2 and 3 work in problem Heat well for every n and all α N = 10 k , 12 k 24 , but in some other problems the accuracy of the area rules 2 and 3 (see columns 4–7 of Table 6 and Table 7) is slightly worse than for the rule TA-2. The advantage of the area rule 3, as compared to the area rule 2, is to be highlighted in problem Heat if all noise of the right hand side is placed on one eigenelement (then we use the condition m k α H Q only in case λ min > α N ). Then the area rule 3 did not fail if n 80 and α N 10 20 . So we can say, the more precisely we take into account the form of the Q-curve in construction of the approximating function for the error function e ˜ 1 ( α ) , the more stable is the rule.
Based on the above rules it is possible to formulate a combined rule, which chooses the parameter according to the rule TA-2 or area rule 3 in dependence of certain condition.
Area rule 4 (Combined area rule). Fix a constant c 0 , 1 c 0 2 , b 0 . Let the local minimizer m k be chosen by the rule TA-2. If
max m k α M r ( k ) ψ ˜ Q ( α ) g [ m k , M r ( k ) ] ( α ) b ,
we take m k for the regularization parameter, otherwise we choose the regularization parameter by the area rule 3.
Note that the combined rule coincides with the rule TA-2, if b = 0 and with the area rule 3, if b = . Experiments of combined rule with c 0 = 2 , b = 1 (columns 8 and 9 in Table 6 and Table 7) show that the accuracy of this rule is almost the same as in the triangle rule, but unlike the TA-2 rule, it works well also in problem Heat for all n and α N . Although, in some cases, in test set 3 the error ratio E > 100 for rule 4, the high qualification of the rule is characterized by the fact, that over all problems sets 1–3 the largest error ratio E1 was 16.91 (5.06 for set 1) and the largest error ratio E2 was 4.67 (2.62 for set 1). Numerical experiments show that it is reasonable to use parameter b ( 0.8 , 1.2 ] . We studied the behavior of the area rules for different α N = 10 k , 12 k 24 . The results were similar to the results of Table 6 and Table 7, but for smaller α N the error ratios were 2–3% smaller than for α N = 10 18 and for larger α N the error ratios were about 5% larger than in Table 6 and Table 7.
The Table 2 gives results of the numerical experiments in the case of smooth solution, p = 2 . We see that combined rule worked well also in this case, no failure.
Remark 5.
It is possible to modify the Q-curve. We may use the function ψ QD ( α ) instead of function ψ Q ( α ) and find proper local minimizer of the function ψ QD ( α ) . Unlike the quasi-optimality criterion the use of function ψ QD ( α ) in the Q-curve and in the area rule does not increase the amount of calculations, while approximation u 2 , α is needed also in computation of d MD ( α ) . We can use in these rules the function d ME ( α ) instead of d MD ( α ) , it increases the accuracy in some problems, but the average accuracy of the rules is almost the same. In case of nonsmooth solutions we can modify the Q-curve method and area rule, using the function d D ( α ) instead of d MD ( α ) . In this case, we get even better results for p = 0 but for p = 2 , the error ratio E is on average 2 times higher.
Note that if the solution is smooth, then L-curve rule and Reginska’s rule often fail, but replacing in these rules the function A u α f by the function B α ( A u α f ) gives often better results.
Remark 6.
If the solution is smooth, then using α-s from (8) much better approximate solution than single Tikhonov approximation may be get using the linear combinations of Tikhonov approximations, see [40].
In the case of a heuristic parameter choice, it is also possible to use the a posteriori estimates of the approximate solution, which, in many tasks, allows to confirm the reliability of the parameter choice. Let α H be the regularization parameter from some heuristic rule and α * be the local minimizer of the function e 1 ( α ) on the set Ω . Then in the case α * α H the error estimate
u α H u * 1 + T ( α H , α * ) e 1 ( α * ) q 1 1 + T 1 ( α H ) min α e 1 ( α )
holds where T 1 ( α H ) = max α α H , α Ω T ( α H , α ) . Using the last estimate, we can prove similarly to the Theorem 7 that if α N is so small that d MD ( α N ) ( 1 + ϵ ) f f * , then
u α H u * max { q 1 ( 1 + T 1 ( α H ) ) min α 0 e 1 ( α ) , C 2 ( b , ϵ ) min α 0 e 2 ( α , f f * ) } ,
where b = d MD ( α H ) / d MD ( α N ) . If values T 1 ( α H ) and b, what we find a posteriori, are small (for example b 2 and T 1 ( α H ) 9 ), then this estimate allows to argue that the error of the approximate solution for this parameter is not much larger than the minimal error. The conditions b 2 , T 1 ( α H ) 9 were satisfied in set 1 of test problems in the combined rule for 73% of cases and inequalities b 2 , T 1 ( α H ) 4 for 61% of cases. The reason of failure of heuristic rule is typically that the chosen parameter is too small. To check this, we can use the error estimate (21). If T 1 ( α H ) is relatively small (for example T 1 ( α H ) 9 ), then the estimate (21) allows to argue that the regularization parameter is not chosen too small. In set 1 of test problems the conditions T 1 ( α H ) 9 and T 1 ( α H ) 4 were satisfied in 97% and in 82% of cases, respectively.

7. Conclusions

We finish the paper with the following conclusion. For the heuristic choice of the regularization parameter we recommend to choose the parameter from the set of local minimizers of the function ψ Q ( α ) or the function ψ QD ( α ) . For choice of the parameter from the local minimizers we proposed the Q-curve method and different area rules. The proposed rules gave much better results than previous heuristic rules on extensive set of test problems. Area rules fail in very few cases in comparison with previous rules, and the accuracy of these rules is comparable even with the δ -rules if the exact noise level is known. In addition, we also provided a posteriori error estimates of the approximate solution, which allows to check the reliability of parameter chosen heuristically.

Author Contributions

Writing—original draft preparation, T.R.; writing—review and editing, U.H. Both authors have read and agreed to the published version of the manuscript.

Funding

The authors are supported by “Personal research funding: Team grant” project PRG864 of the Estonian Ministry of Education and Research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Engl, H.W.; Hanke, M.; Neubauer, A. Regularization of Inverse Problems, Volume 375 of Mathematics and Its Applications; Kluwer: Dordrecht, The Netherlands, 1996. [Google Scholar]
  2. Vainikko, G.M.; Veretennikov, A.Y. Iteration Procedures in Ill- Posed Problems; Nauka: Moscow, Russia, 1986. (In Russian) [Google Scholar]
  3. Bauer, F.; Kindermann, S. The quasi-optimality criterion for classical inverse problems. Inverse Probl. 2008, 24, 035002. [Google Scholar] [CrossRef]
  4. Hämarik, U.; Palm, R.; Raus, T. On minimization strategies for choice of the regularization parameter in ill-posed problems. Numer. Funct. Anal. Optim. 2009, 30, 924–950. [Google Scholar] [CrossRef]
  5. Kindermann, S. Convergence analysis of minimization-based noise level-free parameter choice rules for linear ill-posed problems. Electron. Trans. Numer. Anal. 2011, 38, 233–257. [Google Scholar]
  6. Kindermann, S. Discretization independent convergence rates for noise level-free parameter choice rules for the regularization of ill-conditioned problems. Electron. Trans. Numer. Anal. 2013, 40, 58–81. [Google Scholar]
  7. Kindermann, S.; Neubauer, A. On the convergence of the quasioptimality criterion for (iterated) Tikhonov regularization. Inverse Probl. Imaging 2008, 2, 291–299. [Google Scholar] [CrossRef]
  8. Neubauer, A. The convergence of a new heuristic parameter selection criterion for general regularization methods. Inverse Probl. 2008, 24, 055005. [Google Scholar] [CrossRef]
  9. Tikhonov, A.N.; Glasko, V.B.; Kriksin, Y. On the question of quasioptimal choice of a regularized approximation. Sov. Math. Dokl. 1979, 20, 1036–1040. [Google Scholar]
  10. Hansen, P.C. Analysis of discrete ill-posed problems by means of the L-curve. SIAM Rev. 1992, 34, 561–580. [Google Scholar] [CrossRef]
  11. Hansen, P.C. Regularization tools: A Matlab package for analysis and solution of discrete ill-posed problems. Numer. Algorithms 1994, 6, 1–35. [Google Scholar] [CrossRef] [Green Version]
  12. Golub, G.H.; Heath, M.; Wahba, G. Generalized cross-validation as a method for choosing a good ridge parameter. Technometrics 1979, 21, 215–223. [Google Scholar] [CrossRef]
  13. Hanke, M.; Raus, T. A general heuristic for choosing the regularization parameter in ill-posed problems. SIAM J. Sci. Comput. 1996, 17, 956–972. [Google Scholar] [CrossRef]
  14. Reginska, T.A. A regularization parameter in discrete ill-posed problems. SIAM J. Sci. Comput. 1986, 17, 740–749. [Google Scholar] [CrossRef]
  15. Gockenbach, M.S.; Gorgin, E. On the convergence of a heuristic parameter choice rule for Tikhonov regularization. SIAM J. Sci. Comput. 2018, 40, A2694–A2719. [Google Scholar] [CrossRef] [Green Version]
  16. Hämarik, U.; Kangro, U.; Kindermann, S.; Raik, K. Semi-heuristic parameter choice rules for Tikhonov regularisation with operator perturbations. J. Inverse Ill Posed Probl. 2019, 27, 117–131. [Google Scholar] [CrossRef]
  17. Hochstenbach, M.E.; Reichel, L.; Rodriguez, G. Regularization parameter determination for discrete ill-posed problems. J. Comput. Appl. Math. 2015, 273, 132–149. [Google Scholar] [CrossRef]
  18. Palm, R. Numerical Comparison of Regularization Algorithms for Solving ill-Posed Problems. Ph.D. Thesis, University of Tartu, Tartu, Estonia, 2010. [Google Scholar]
  19. Bauer, F.; Lukas, M.A. Comparing parameter choice methods for regularization of ill-posed problems. Math. Comput. Simul. 2011, 81, 1795–1841. [Google Scholar] [CrossRef] [Green Version]
  20. Bakushinskii, A.B. Remarks on choosing a regularization parameter using the quasi-optimality and ratio criterion. Comp. Math. Math. Phys. 1984, 24, 181–182. [Google Scholar] [CrossRef]
  21. Raus, T.; Hämarik, U. Heuristic parameter choice in Tikhonov method from minimizers of the quasi-optimality function. In New Trends in Parameter Identification for Mathematical Models; Hofmann, B., Leitao, A., Zubelli, J., Eds.; Birkhäuser: Basel, Switzerland, 2018; pp. 227–244. [Google Scholar]
  22. Brezinski, B.; Rodriguez, G.; Seatzu, S. Error estimates for linear systems with applications to regularization. Numer. Algor. 2008, 49, 85–104. [Google Scholar] [CrossRef]
  23. Baker, C.T.H.; Fox, L.; Mayers, D.F.; Wright, K. Numerical solution of Fredholm integral equations of the first kind. Comput. J. 1964, 7, 141–148. [Google Scholar] [CrossRef] [Green Version]
  24. Groetsch, C.W. Integral equations of the first kind, integral equations, and regularization: A crash course. J. Phys. Conf. Ser. 2007, 73, 012001. [Google Scholar] [CrossRef] [Green Version]
  25. Indratno, S.W.; Ramm, A.G. An iterative method for solving Fredholm integral equations of the first kind. Int. J. Comput. Sci. And Mathematics 2009, 2, 354–379. [Google Scholar] [CrossRef]
  26. Wazwaz, A.-M. The regularization method for Fredholm integral equations of the first kind. Comput. Math. Appl. 2011, 61, 2981–2986. [Google Scholar] [CrossRef] [Green Version]
  27. Morozov, V.A. On the solution of functional equations by the method of regularization. Soviet Math. Dokl. 1966, 7, 414–417. [Google Scholar]
  28. Gfrerer, H. An a posteriori parameter choice for ordinary and iterated Tikhonov regularization of ill-posed problems leading to optimal convergence rates. Math. Comp. 1987, 49, 507–522. [Google Scholar] [CrossRef]
  29. Raus, T. On the discrepancy principle for solution of ill-posed problems with non-selfadjoint operators. Acta Comment Univ. Tartu. 1985, 715, 12–20. (In Russian) [Google Scholar]
  30. Hämarik, U.; Kangro, U.; Palm, R.; Raus, T.; Tautenhahn, U. Monotonicity of error of regularized solution and its use for parameter choice. Inverse Probl. Sci. Eng. 2014, 22, 10–30. [Google Scholar] [CrossRef]
  31. Tautenhahn, U.; Hämarik, U. The use of monotonicity for choosing the regularization parameter in ill-posed problems. Inverse Probl. 1999, 15, 1487–1505. [Google Scholar] [CrossRef]
  32. Hämarik, U.; Palm, R.; Raus, T. Comparison of parameter choices in regularization algorithms in case of different information about noise level. Calcolo 2011, 48, 47–59. [Google Scholar] [CrossRef]
  33. Hämarik, U.; Palm, R.; Raus, T. A family of rules for parameter choice in Tikhonov regularization of ill-posed problems with inexact noise level. J. Comp. Appl. Math. 2012, 36, 221–233. [Google Scholar] [CrossRef] [Green Version]
  34. Raus, T.; Hämarik, U. New rule for choice of the regularization parameter in (iterated) Tikhonov method. Math. Model. Anal. 2009, 14, 187–198. [Google Scholar] [CrossRef]
  35. Raus, T.; Hämarik, U. On numerical realization of quasioptimal parameter choices in (iterated) Tikhonov and Lavrentiev regularization. Math. Model. Anal. 2009, 14, 99–108. [Google Scholar] [CrossRef] [Green Version]
  36. Raus, T. About regularization parameter choice in case of approximately given error bounds of data. Acta Comment Univ. Tartu. 1992, 937, 77–89. [Google Scholar]
  37. Raus, T.; Hämarik, U. On the quasioptimal regularization parameter choices for solving ill-posed problems. J. Inverse Ill Posed Probl. 2007, 15, 419–439. [Google Scholar] [CrossRef]
  38. Calvetti, D.; Reichel, L.; Shuib, A. L-curve and curvature bounds for Tikhonov regularization. Numer. Algorithms 2004, 35, 301–314. [Google Scholar] [CrossRef]
  39. Hansen, P.C. Rank-Deficient and Discrete Ill-Posed Problems; SIAM: Philadelphia, PA, USA, 1998. [Google Scholar]
  40. Hämarik, U.; Palm, R.; Raus, T. Extrapolation of Tikhonov regularization method. Math. Model. Anal. 2010, 15, 55–68. [Google Scholar] [CrossRef] [Green Version]
Figure 1. L-curve for Baart.
Figure 1. L-curve for Baart.
Mathematics 08 01166 g001
Figure 2. Q-curve for Baart.
Figure 2. Q-curve for Baart.
Mathematics 08 01166 g002
Figure 3. L-curve for Deriv2.
Figure 3. L-curve for Deriv2.
Mathematics 08 01166 g003
Figure 4. Q-curve for Deriv2.
Figure 4. Q-curve for Deriv2.
Mathematics 08 01166 g004
Figure 5. Q-curve in Heat.
Figure 5. Q-curve in Heat.
Mathematics 08 01166 g005
Figure 6. Q-curve in Spikes.
Figure 6. Q-curve in Spikes.
Mathematics 08 01166 g006
Figure 7. Q-curve in Foxgood.
Figure 7. Q-curve in Foxgood.
Mathematics 08 01166 g007
Figure 8. Q-curve in Groetsch2.
Figure 8. Q-curve in Groetsch2.
Mathematics 08 01166 g008
Figure 9. Q-curve, problem Heat, n = 60 .
Figure 9. Q-curve, problem Heat, n = 60 .
Mathematics 08 01166 g009
Figure 10. Q-curve, problem Heat, n = 60 .
Figure 10. Q-curve, problem Heat, n = 60 .
Mathematics 08 01166 g010
Table 1. Characteristics of matrix A T A and the solution u * .
Table 1. Characteristics of matrix A T A and the solution u * .
Problem λ min N 1 Λ p1Problem λ min N 1 Λ p1
Baart5.2 × 10 35 921665.70.197Spikes1.3 × 10 33 891529.30.005
Deriv26.7 × 10 9 016.00.286Wing2.9 × 10 37 949219.10.057
Foxgood9.0 × 10 33 85210.10.426Baker1.0 × 10 33 949153.10.498
Gravity1.6 × 10 33 684.10.403Ursell6.9 × 10 34 943090.20.143
Heat5.5 × 10 33 32.4 × 10 20 0.341Indramm2.7 × 10 33 949154.60.395
Ilaplace3.8 × 10 33 7916.10.211Waswaz22.0 × 10 34 981.7 × 10 30 0.654
Phillips1.4 × 10 13 09.40.471Groetsch15.8 × 10 33 7811.20.176
Shaw2.3 × 10 34 85289.70.244Groetsch21.0 × 10 4 04.00.652
Table 2. Results of the numerical experiments, p = 2 .
Table 2. Results of the numerical experiments, p = 2 .
ProblemMEMEeDPBest of L min | L min | Combined Area Rule
Aver EAver EAver EAver EAverAver EMax E
Baart1.861.192.931.184.741.6014.57
Deriv21.091.193.651.032.001.041.17
Foxgood1.561.133.581.142.081.223.58
Gravity1.331.052.651.091.721.143.18
Heat1.131.122.551.052.101.051.14
Ilaplace1.471.062.781.112.731.133.51
Phillips1.261.063.351.042.101.041.20
Shaw1.371.062.581.113.721.298.96
Spikes1.851.122.101.194.781.315.75
Wing1.671.142.471.224.531.756.63
Baker2.111.292.961.214.381.7711.33
Ursell1.861.194.101.164.821.6718.08
Indramm1.691.142.871.284.531.916.42
Waswaz2127.249.81.202.441.002.439.01
Groetsch11.401.062.361.112.141.144.56
Groetsch21.021.231.711.141.671.553.97
Set 19.374.182.741.223.061.4418.08
Set 22.101.262.911.192.831.3729.03
Set 36.863.212.681.183.121.4252.98
Table 3. Results for the set L min .
Table 3. Results for the set L min .
ProblemMEMEeDPBest of L min | L min | Apost. C
Aver EAver EAver EAver EMax EAverMaxAverMax
Baart1.431.321.371.232.516.9183.193.72
Deriv21.291.071.211.081.341.7123.544.49
Foxgood1.981.421.341.476.193.6363.724.16
Gravity1.401.131.161.131.831.6433.714.15
Heat1.191.031.051.122.363.1953.924.50
Ilaplace1.331.211.261.202.562.6454.846.60
Phillips1.271.021.021.061.722.1433.994.66
Shaw1.371.241.281.192.154.6873.484.43
Spikes1.011.001.011.001.028.83103.273.70
Wing1.161.131.151.091.385.2063.073.72
Baker3.912.382.092.3116.175.3863.143.72
Ursell2.141.972.031.694.445.5363.073.43
Indramm5.203.263.373.3825.675.6463.083.71
Waswaz2127.249.91.202.449.031.0012.002.00
Groetsch11.121.071.081.061.513.9974.235.20
Groetsch21.021.221.671.131.691.6725.6213.72
Set 19.624.461.461.4825.673.99103.6713.72
Set 21.571.321.361.205.334.40103.505.43
Set 37.193.451.471.4861.023.64103.739.12
Table 4. Averages of error ratios E and failure % (in parenthesis) for heuristic rules.
Table 4. Averages of error ratios E and failure % (in parenthesis) for heuristic rules.
ProblemTA RuleQuasiopt.WQHRReginskaMCurv
Mean EMax EMean EMean EMean EMean EMean E
Baart1.5114.571.541.432.581.324.75
Deriv21.181.272.012.262.283.67(9.2%)
Foxgood1.563.391.571.578.36(10.8%)5.95
Gravity1.142.271.131.132.66(0.8%)2.04
Heat1.261.34(65.8%)(66.8%)1.64(4.2%)4.11
Ilaplace1.242.341.241.221.941.662.99
Phillips1.071.201.09(3.3%)2.27(44.2%)1.34
Shaw1.428.961.431.412.341.804.64
Spikes1.015.751.011.011.031.011.05
Wing1.396.631.401.301.511.181.57
Baker3.3011.333.303.30(0.8%)(21.7%)7.78
Ursell2.8731.063.542.354.711.867.54
Indramm3.749.074.434.16(2.5%)(9.2%)(15.8%)
Waswaz22.439.012.432.43(65.8%)2.33(3.3%)
Groetsch11.144.561.141.121.611.261.52
Groetsch21.131.741.272.731.665.491.81
Total1.7131.06>100>10050.543.88.17
Failure %0% 4.11%4.38%4.32%5.68%1.77%
Max E22.61 >100>1002.63>10024.5
Table 5. Distribution of error ratios E in different rules.
Table 5. Distribution of error ratios E in different rules.
DecileTA RuleQuasiopt.WQHRReginskaMCurvMEMEeDP
101.001.001.001.081.001.061.011.001.00
201.011.011.011.361.041.271.031.001.01
301.021.031.031.561.121.481.091.011.02
401.041.061.061.821.271.831.161.031.04
501.091.131.122.121.662.311.221.081.08
601.181.291.292.432.423.051.331.161.16
701.351.571.593.194.194.511.521.291.30
801.712.172.295.949.937.032.021.501.52
902.276.456.1819.3543.9112.954.452.882.11
Table 6. Averages and maximums of error ratios E in case of area rules, problem set 1.
Table 6. Averages and maximums of error ratios E in case of area rules, problem set 1.
ProblemTA-2 RuleArea Rule 2Area Rule 3Combined Area Rule
Aver EMax EAver EMax EAver EMax EAver EMax E
Baart1.515.181.582.911.592.911.535.18
Deriv21.121.421.121.421.121.421.121.42
Foxgood1.576.691.536.191.536.191.576.69
Gravity1.174.121.216.101.216.101.174.12
Heat1.122.361.122.361.122.361.122.36
Ilaplace1.242.681.222.681.222.681.242.68
Phillips1.071.721.061.721.061.721.071.72
Shaw1.423.721.473.641.473.641.423.72
Spikes1.011.051.011.021.011.021.011.05
Wing1.391.861.441.861.441.861.391.86
Baker3.3045.292.6722.672.6722.672.9133.12
Ursell2.8716.784.5527.924.5527.923.1216.78
Indramm3.7425.679.5083.2010.7683.203.8725.67
Waswaz22.439.012.439.012.439.012.439.01
Groetsch11.142.121.152.121.152.121.142.12
Groetsch21.523.841.523.841.523.841.523.84
Total1.7345.292.1683.202.2283.201.7333.12
Table 7. Averages and maximums of error ratios E in proposed rules, problem sets 2 and 3.
Table 7. Averages and maximums of error ratios E in proposed rules, problem sets 2 and 3.
ProblemTA-2 RuleArea Rule 2Area Rule 3Combined Area Rule
Aver EMax EAver EMax EAver EMax EAver EMax E
Gauss1.245.051.266.561.266.561.245.05
Hilbert1.467.251.8321.221.8121.221.467.25
Lotkin1.4711.171.9118.661.8811.171.4711.17
Moler1.517.351.437.351.437.351.517.35
Prolate1.5715.961.8220.641.7715.961.5815.96
Pascal1.041.131.061.181.061.181.051.18
Set 21.3815.961.5521.221.5321.221.3915.96
Set 31.85136.62.81188.12.77188.12.02153.5

Share and Cite

MDPI and ACS Style

Raus, T.; Hämarik, U. Q-Curve and Area Rules for Choosing Heuristic Parameter in Tikhonov Regularization. Mathematics 2020, 8, 1166. https://doi.org/10.3390/math8071166

AMA Style

Raus T, Hämarik U. Q-Curve and Area Rules for Choosing Heuristic Parameter in Tikhonov Regularization. Mathematics. 2020; 8(7):1166. https://doi.org/10.3390/math8071166

Chicago/Turabian Style

Raus, Toomas, and Uno Hämarik. 2020. "Q-Curve and Area Rules for Choosing Heuristic Parameter in Tikhonov Regularization" Mathematics 8, no. 7: 1166. https://doi.org/10.3390/math8071166

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop