Abstract
In this paper, we propose a suitable combination of two different Nyström methods, both using the zeros of the same sequence of Jacobi polynomials, in order to approximate the solution of Fredholm integral equations on . The proposed procedure is cheaper than the Nyström scheme based on using only one of the described methods . Moreover, we can successfully manage functions with possible algebraic singularities at the endpoints and kernels with different pathologies. The error of the method is comparable with that of the best polynomial approximation in suitable spaces of functions, equipped with the weighted uniform norm. The convergence and the stability of the method are proved, and some numerical tests that confirm the theoretical estimates are given.
1. Introduction
Let the following be a Fredholm Integral Equation (FIE) of the second kind:
where is a Jacobi weight, g and k are known functions defined in and , respectively, is a non zero real parameter and f is the unknown function we want to look for. The kernel function k is also allowed to be weakly singular along the diagonal , or it could show some other pathologies such as high oscillating behaviour or a “nearly singular” factor. The nature of the kernel, with the presence of the Jacobi weight inside the integral, implies that the solution f can have a singular behaviour at the endpoints of the definition interval (see for instance [,]); therefore, the natural choice is to study Equation (1) in suitable spaces of weighted functions.
A high number of papers on the numerical methods for FIEs is disposable in the literature, and in the last two decades a deep attention was devoted, in the case under consideration, to the so-called “global approximation methods”. They are essentially based on polynomial approximation and use zeros of orthogonal polynomials (see for instance [,] and the references therein). There are also examples of global approximation methods based on equispaced points [], which are especially convenient when the data are available in discrete form but are limited to the unweighted case (see [,]). Global methods, more or less, behave as the best polynomial approximation of the solution in suitable spaces of weighted functions; consequently, this approximation strategy provides a powerful performance in the case of very smooth functions. On the other hand, these methods can converge slowly if the functions are not smooth or if the kernel has pathologies as described above.
Recently in [], a new method based on the collocation approach using the so-called Extended Interpolation was proposed in order to reduce the computational effort in the cases where the solution is not so smooth []. Moreover the method delays the computation of high degree polynomial zeros that becomes progressively unstable as the degree increases.
Following a similar idea, we propose here a Mixed Nyström scheme based on product quadrature rules of the “extended” type, i.e., based on the zeros of the polynomial , where denotes the orthonormal sequence with respect to a suitable fixed Jacobi weight w []. The advantages of using a Nyström scheme with respect to a collocation one include several benefits. Indeed, first of all, we will use here only one sequence of orthonormal polynomials, while in the collocation method in [] two different sequences ( and , where ) were required to obtain optimal Lebesgue constants of the interpolating operators. Secondly, due to its nature, the Nyström strategy with a fixed kernel provides a faster convergence with respect to the collocation approach if the right-hand side g in (1) is not so smooth. Third, a Nyström method based on a product rule allows treating kernel functions having different pathologies.
The idea of the proposed scheme is the following. Consider two sequences of Nyström interpolants and : The first is based on the product rule using the zeros of , and the second one is based on the extended product rule using the zeros of . Each step of the procedure consists in solving, for a fixed m, the first Nyström method and in using the coefficients defining the corresponding Nyström interpolant in order to “reduce” about one half of the computation of the coefficients of the Extended Nyström interpolant . In other words, we will assume that the two interpolants “coincide” on the zeros of . This assumption results in solving only a linear system of order instead of one of dimension in order to obtain an approximating function that is comparable with from the convergence point of view.
The outline of the paper is the following. Section 2 contains preliminary notations and a collection of tools needed to introduce the main results stated in Section 3. Here, we present an extended Nyström method, and the combined algorithm that allows us to solve Equation (1) faster based on this. In Section 4, we provide some computational details for the effective construction of the linear systems. Section 5 concerns the numerical tests, while Section 6 contains the proofs.
2. Notation and Preliminary Results
Throughout the paper, we use in order to denote a positive constant, which may have different values at different occurrences, and we write to mean that is independent of .
2.1. Function Spaces
Let u be the Jacobi weight defined as follows:
We denote by the Banach space of the locally continuous functions f on such that the following limit conditions are satisfied:
is equipped with the following norm:
The limit conditions (2) are necessary in order to assure that the following is the case (see for instance []):
where, denoted by the space of all algebraic polynomials having degree at most m, it is
the error of best polynomial approximation of .
For smoother functions, we consider the following Sobolev-type subspaces of of order , defined as the following:
where denotes the space of all the functions that are absolutely continuous on every closed subset of , and . is equipped with the following norm:
Finally, by , we denote the set of all measurable function f defined in such that the following is the case:
For any bivariate function , we will write (or ) in order to regard k as the univariate function in the only variable x (or y).
2.2. Solvability of the Equation (1) in
Let us set the following:
In order to provide the sufficient conditions assuring the compactness of the operator , we need to recall the following definition. For any and with , in [], it was defined the following modulus of smoothness:
where
and the following is the case:
For any , the modulus is estimated by means of the following inequality (see for instance [], p. 314):
We are now able to state a theorem that guarantees the solvability of the Equation (1) in the space and for which its proof is given in Section 6.
Theorem 1.
Under the following assumptions, with and ,
the operator is compact. Therefore, if , for any , Equation (1) admits a unique solution in .
2.3. Product Integration Rules
Denoted by , the system of the orthonormal polynomials with respect to the Jacobi weight , the polynomial is so defined:
Let be the zeros of and let the following:
be the Christoffel numbers with respect to w.
For the following integral:
consider the following product integration rule:
where
According to a consolidated terminology, we will refer to the product integration rule in (5) as Ordinary Product Rule only to distinguish it from the extended product integration rule introduced below. Moreover, we recall that are known as Modified Moments [] (see, e.g., []).
With respect to the stability and the convergence of the previous rule, the following result, useful for our aim, can be deduced by ([], p. 348) (see also []).
Theorem 2.
Under the following assumptions:
for any we obtain the following bounds:
and
with .
In addition to the previous well-known product rule, we recall the following Extended Product Rule (see []) based on the zeros of . Denoted by the zeros of , the extended formula is as follows: -4.6cm0cm
where
and
are known as the Generalized Modified Moments (GMMs).
With respect to the stability and convergence of the extended quadrature rule (8), we recall the following.
Theorem 3
([], Theorem 3.2). Under the following assumptions:
for any , we obtain the following bounds:
and
with .
2.4. A Nyström Method
In order to approximate the solution of (1), we recall the following weighted Nyström method based on the product quadrature rule (5). Introducing the sequence where the following is the case:
we proceed to solve in the following finite dimensional equation in the unknown :
By multiplying both sides of the previous equation by the weight function u and collocating at the nodes , we reach the following linear system of order m in the unknowns :
where is defined in (6).
Details on the matrix will be given in Section 4.
Once the solution of the system (16) has been determined, the Ordinary Nyström interpolant takes the form:
About the convergence of this method, the following theorem can be obtained by using weighted arguments in []:
Theorem 4.
Under the assumptions of Theorems 1 and 2, for any , the finite dimensional Equation (14) admits a unique solution such that we have the following:
with and .
In what follows, we will refer to this Nyström method as the Ordinary Nyström Method (ONM).
3. Main Results
3.1. The Extended Nyström Method
Now, we introduce a Nyström method based on the extended product quadrature rule (8), calling it the Extended Nyström Method (ENM). Proceeding in analogy to the ONM, we begin by constructing the following sequence :
Moreover, we solve in the following extended finite dimensional equation in the unknown :
By multiplying both sides of (19) by the weight function u and collocating at the quadrature nodes , we obtain the following linear system:
where the following is the case:
The coefficients and are defined in (9)–(10). The linear system of order can be rewritten in the following more convenient block-matrix form:
with and . Details on the effective construction of the system will be provided in Section 4.
Denoted by , the vector solution of (21), the extended Nyström interpolant takes the following form:
With respect to the convergence, we are able to prove the following:
Theorem 5.
Under the assumption of Theorems 1 and 3, for any , the finite dimensional Equation (19) admits a unique solution such that the following error estimate holds:
3.2. The Mixed Nyström Method
We observe that, under suitable assumptions, both sequences and uniformly converge to the solution f of (1). Thus, it makes sense to consider a mixed scheme that combines the two methods previously introduced.
Therefore, the Mixed Nyström Method (MNM) consists of two steps:
- For a given m, solve the linear system of order m:
- By assuming in the linear system (21), we obtain the following:Denoted by its solution, we construct the interpolant :
- Restart the procedure determining and and so on.
The mixed sequence of Nyström interpolants is obtained by iterating a couple of steps of the types (24)–(26), allowing us to obtain the following mixed sequence of Nyström interpolants:
Denoting by the infinity norm of the matrix , the uniform convergence of to the solution of (1) is stated in the following:
Theorem 6.
Under the assumptions of Theorems 1, 4, 5 and supposing that the matrix in (25) is invertible, with , for any , the sequence uniformly converges to , and the following error estimate holds:
Remark 2.
By implementing the Mixed Nyström Method, we gain different advantages, specifically the reduction in the sizes of the involved linear systems.
More precisely, at each step of the mixed scheme, setting , we solve two systems of order m and . By doing so, the obtained error is comparable with that performed by solving the two systems of order m and by the Ordinary Nyström Method.
Therefore, the computational cost of the global procedure is strongly reduced. Indeed, if we compute the solution of the linear systems by Gaussian Elimination, we save 77.8% off long operations and 33.2% off function evaluations.
Furthermore, the difficulties in the evaluation of the modified moments for “large” degree are delayed, as well as the instability in the construction of Jacobi polynomial zeros of high degrees by the Golub–Welsh algorithm.
4. Computational Details
Given two integers , in this section, we use the short notation to denote the set Denoting by the identity matrix of order m, the matrix of the linear system (16) takes the following form:
with
It is well known that the system (16) and the finite dimensional Equation (14) are equivalent (see for instance ([], Theorem 12.7, p. 202)).
About the block-matrix of the system (21), according to the previously introduced matrices, we have the following:
where
and the matrices , , and are as follows:
Remark 3.
The entries of the matrices in (29) require the computation of the GMMs. As usual, the ordinary Modified Moments (MMs), which depend on the specific kernel we considered, are often derived by suitable recurrence relations (see, e.g., []). In [] a general scheme for deriving GMMs starting from MMs was proposed. Alternatively, for very smooth kernels, Gaussian rules can be also used. In any case, the global algorithm can be organized in such a manner that the matrices in (29), requiring the most expensive computation effort, can be performed once for a given couple .
5. Numerical Experiments
Now we propose some tests showing the numerical results that were obtained by approximating the solution of equations of the type (1) by the mixed sequence in (27). We will compare the results with those attained by the corresponding ordinary sequence used in the standard method and in Example 2 also with those achieved by the mixed collocation method proposed in []. Indeed, in this test, the kernel is moderately smooth, and the convergence conditions of both methods are satisfied.
We have selected g possessing different regularities and kernels k presenting some kinds of drawback such as a contemporary high oscillating behaviour with a “near” singular fixed point or being weakly singular. In each test, we will report either the common weight w used in the construction of the quadrature formulae and the weight u defining the space to which f belongs.
For effective comparison between the ordinary and the extended sequences on the same number of nodes, we have considered the following sequences:
Since the solution f is unknown, we retain the exact values attained by the approximating function , for sufficiently large N. We remark that turns out to be a suitable choice for the functions under consideration. In the tables, for increasing n, we will report the weighted maximum error attained by and at the set , of equally spaced points of by setting
In each table, first and third columns contain the size of the ordinary system (o.l.s.) and its condition number related to the ordinary sequence. In the fourth and sixth columns, the sizes of the couple of linear systems of the mixed scheme (m.l.s.) and the condition number of the “reduced” system (25) are reported. All the condition numbers have been computed with respect to the infinity norm.
Finally, in order to contain possible moderate loss of accuracy in computing GMMs, we have carried out their construction by using the software Wolfram Mathematica 12.1 in quadruple precision. All the other computations have been performed in double-machine precision 2.220446049250313e-16.
Example 1.
Let us consider the following equation:
In this case and according to Theorem 6, which holds since all the assumptions are satisfied, the errors are , and the numerical results reported in Table 1 are even better. All the linear systems are well conditioned, the ordinary condition numbers being slightly smaller than the mixed ones. The weighted absolute errors by ONM and MNM are displayed in Figure 1.

Table 1.
Example 1.

Figure 1.
Example 1. (a) Errors by Ordinary Nyström Method. (b) Errors by Mixed Nyström Method.
Example 2.
Let us consider the following equation:
In Table 2 and Table 3 we report the results achieved by the mixed and ordinary Nyström methods and those obtained by the mixed and ordinary collocation methods in []. Indeed, the assumptions assuring stability and convergence for all the methods are satisfied; hence, the comparison makes sense.

Table 2.
Example 2: Ordinary and Mixed Nyström methods.

Table 3.
Example 2: Ordinary and Mixed Collocation methods [].
We denote by and the weighted maximum error attained by the Ordinary Collocation Method (OCM) and the Mixed Collocation Method (MCM) in [] at the set , of equally spaced points of , respectively.
The results show that both the Nyström methods behave better than the collocation ones, and this is quite common in cases such as the one under consideration. Indeed, even if the solution (since ), the rate of convergence of the collocation approach depends on both the approximations of the integral operator and the right-hand side. On the contrary, the order of convergence of the Nyström method depends essentially on the smoothness of the kernel. This is one of the reasons why in these cases the Nyström approach produces better results than the collocation one, as also announced in the Introduction.
Example 3.
Let us consider the following equation:
In this test, the kernel presents a fast oscillating behaviour. Its graphic is reported in Figure 2. Hence, the product formula allows us to overcome the drawbacks deriving from the use of the Gauss–Jacobi rule. About the rate of convergence, since , we expect that the errors are . We have reported the values attained by ONM and MNM in three different points of the interval (Table 4) and the maximum errors on the entire interval (Table 5). In all the cases the theoretical estimates are attained. Moreover, in both methods the condition numbers of the linear systems are comparable.

Figure 2.
Example 3: graphic of k(x) = cos(250x).

Table 4.
Example 3: numerical values of the weighted solution attained by ONM and MNM.

Table 5.
Example 3.
Example 4.
Let us consider the following equation:
In this case, ρ and u satisfy the assumptions for the convergence of the Nyström methods, while they do not satisfy those of the collocation methods []. We recall that for the convergence of both the collocation methods smoother kernels and more restrictive assumptions on the weights are required. About the rate of convergence, since , we expect that the errors are . Moreover, we have chosen this test to propose a comparison with the Nyström method obtained by approximating the coefficients in (5) by Gaussian rules. We will refer to this procedure as the Ordinary Nyström method by Gaussian rule (shortly, ONG). We point out that the nature of the kernel k makes this comparison possible, since the computation of the coefficients by the Gauss–Jacobi rule can be performed. Thus, in Table 6, in addition to the results by the mixed and ordinary Nyström methods, in the last two columns, we will set the maximum weighted errors attained by the ONG at the same set of nodes , and the condition numbers of the corresponding linear systems. They will be shortly denoted as and , respectively. The results by ONM and MNM are slightly better than the expected accuracy, and the condition numbers of the mixed linear systems are a little bit lower than their ordinary counterparts. With respect to the ONG method, as we can observe, the errors result as stagnant.

Table 6.
Example 4.
Example 5.
Let us consider the following equation:
This test deals with a kernel that is a product of a periodic function having high frequency and multiplied by a “nearly” singular function. Such kernels, treated in the bidimensional case in [], appear for instance in the solution of problems of propagation in uniform wave-guides with non-perfect conductors []. The graphic of the kernel is given in Figure 3. With respect to the results of the equation, since , a very fast convergence is expected, and this is confirmed also by the numerical results reported in Table 7. The mixed condition numbers are significantly smaller than the ordinary ones. The graphic of the weighted solution is provided in Figure 4.

Figure 3.
Example 5: graphic of .

Table 7.
Example 5.

Figure 4.
Example 5: graphic of the weighted solution .
6. Proofs
In order to prove Theorem 1, we recall the following well-known inequality ([], p. 171).
Proposition 1
(Weak Jackson Inequality). Let and with . Then, the following inequality holds:
where and .
Proof of Theorem 1.
Observing that the following is the case:
the boundedness of the operator K is proved by using the first assumption of (4).
A well-known result (see ([], 2.5.1, p. 44)) states that the bounded operator K is compact if and only if . Then, by using the weak Jackson inequality and (4), we obtain the following bound:
Thus, the theorem follows. □
To prove Theorem 5, we recall the following well-known result (see, for instance ([], Theorem 4.1.1, p. 106)).
Theorem 7.
Let be a Banach space. Assume to be a bounded compact operator and , , to be a sequence of given bounded operators with for all . Consider the operator equations of the following:
where I is the identity operator in and . If , i.e., the sequence is collectively compact, then , for all sufficiently large m, exists and is uniformly bounded with the following.
Moreover, denoted by and , the unique solutions of (32) and (33). respectively, the following results.
Proof of Theorem 5.
The invertibility of follows under the assumption of Theorem 1, while the uniformly boundedness of and the following limit condition
hold under the assumptions of Theorem 3.
It remains to be proven that the sequence is collectively compact. This is equivalent (see ([], p. 114) and ([], p. 44)) to prove that . Hence, noting that the following is the case:
the thesis is easily proved using (34) and the uniform boundedness of the operator K. □
Proof of Theorem 6.
In order to prove (28), we first define the following sequence:
obtained by composing two sub-sequences of those defined by the Nyström methods ONM in (17) and ENM in (22) . As proved in Theorems 4 and 5, under the assumptions of Theorems 1, 4 and 5, these sequences are both convergent to the unique solution of the integral Equation (1). Therefore, all the sub-sequences are convergent to the same limit function f, and the speed of convergence is the same.
Consequently, we can conclude that with , the following is the case:
Hence, what remain to be conducted in order to obtain (28) is to estimate the following distance:
From the definition of the two polynomial sequences, we can write the following:
We immediately recognize that, by the definition of and , using estimates (18) and (23), we obtain the following.
Consequently, the following is the case:
where , , denotes the infinity norm in .
Now, we remark that by (21) and (25) and under the assumption that is invertible, the following identity holds true:
Therefore, we have the following:
If we denote by the matrix of coefficients in (21), we note that is a submatrix of it. By using standard arguments (see for instance []), it is possible to show that and the operator norm at the right-hand side is uniformly bounded with respect to m since the sequence is bounded in virtue of (12). Therefore, since we are assuming that , we can conclude that the following is the case:
7. Conclusions
In this research, we have proposed a global Nyström method involving ordinary and extended product integration rules, both based on Jacobi zeros. For the nature of the method, we can handle FIE with kernels presenting some kind of pathological behaviours since the coefficients of the rules are exactly computed via recurrence relations. The method employs two different discrete sequences, namely the ordinary and the extended sequences, that are suitably mixed to strongly reduce the computational effort required by the ordinary Nyström method. Advantages are achieved with respect to the mixed collocation method in [] from different points of view that can be summarised as follows: we can treat FIEs as having less regular kernels and under wider assumptions in order to obtain a better rate of convergence. Such improvements have been shown by means of some numerical tests. In particular, Example 2 evidences how the mixed Nyström method provides a better performance than the mixed collocation one in []. Furthermore, Example 4 shows how the assumptions of the mixed Nyström method are wider than those of the above mentioned mixed collocation one. Both methods allow us to reduce the sizes of the involved linear systems but require the computation of Modified and Generalized Modified Moments. In any case, once the kernel k and the order m are given, the algorithm can be organized pre-computing the matrix of the system. Moreover, once Modified Moments are given, Generalized Modified Moments can be always deduced by a suitable recurrence relation (see, e.g., []). Therefore, the global process has a general applicability and only requires the assumptions of convergence to be satisfied. With respect to the Modified Moments, they can be computed through recurrence relations (see, e.g., []). However, when these relations are unstable, Modified Moments can be accurately computed by suitable numerical methods. For instance, in the case of high oscillating or nearly singular kernels, this approach has been successfully tried by implementing “dilation” techniques [,]. The major cost represents a well known limit of the classical Nyström methods based on product integration rules. They are more expensive since the coefficients of the rule possessing many and different pathological kernels have to be “exactly” computed. On the other hand, this major effort is amply repaid by the better performance with respect to other cheaper procedures. Finally, establishing that the convergence conditions are also necessary is still an open problem. This will be a subject for further investigations.
Author Contributions
All authors equally contributed to the paper. Conceptualization, D.M., D.O. and M.G.R.; methodology, D.M., D.O. and M.G.R.; software, D.M., D.O. and M.G.R.; validation, D.M., D.O. and M.G.R.; analysis, D.M., D.O. and M.G.R.; investigation, D.M., D.O. and M.G.R.; resources, D.M., D.O. and M.G.R.; data curation, D.M., D.O. and M.G.R.; writing—original draft preparation, writing—review and editing, D.M., D.O. and M.G.R.; visualization, D.M., D.O. and M.G.R.; supervision D.M., D.O. and M.G.R. All authors have read and agreed to the published version of the manuscript.
Funding
This research was partially supported by University of Basilicata (local funds) and by GNCS Project 2020 “Approssimazione multivariata ed equazioni funzionali per la modellistica numerica”.
Acknowledgments
The authors thank the anonymous referees for their suggestions and remarks, which allowed to improve the paper. The research has been accomplished within “Research ITalian network on Approximation” (RITA). All the authors are members of the INdAM-GNCS Research Group. The second and third authors are members of the TAA-UMI Research Group.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Fermo, L.; Russo, M.G. Numerical methods for Fredholm integral equations with singular right-hand sides. Adv. Comput. Math. 2010, 33, 305–330. [Google Scholar] [CrossRef]
- Vainikko, G.; Pedas, A. The properties of solutions of weakly singular integral equations. J. Aust. Math. Soc. 1981, 22, 419–430. [Google Scholar] [CrossRef]
- De Bonis, M.C.; Mastroianni, G. Projection methods and condition numbers in uniform norm for Fredholm and Cauchy singular integral equations. SIAM J. Numer. Anal. 2006, 44, 1351–1374. [Google Scholar] [CrossRef]
- Occorsio, D.; Russo, M.G. A mixed collocation scheme for solving second kind Fredholm integral equations in [−1,1]. Electron. Trans. Numer. Anal. 2021, 54, 443–459. [Google Scholar] [CrossRef]
- Occorsio, D.; Russo, M.G. Nyström Methods for Fredholm Integral Equations Using Equispaced Points. Filomat 2014, 28, 49–63. [Google Scholar] [CrossRef]
- Occorsio, D.; Russo, M.G.; Themistoclakis, W. Some numerical applications of generalized Bernstein Operators. Constr. Math. Anal. 2021, 4, 186–214. [Google Scholar]
- Criscuolo, G.; Mastroianni, G.; Occorsio, D. Convergence of extended Lagrange interpolation. Math. Comput. 1990, 55, 197–212. [Google Scholar] [CrossRef]
- Occorsio, D.; Russo, M.G. A mixed scheme of product integration rules in (−1,1). Appl. Numer. Math. 2020, 149, 113–123. [Google Scholar] [CrossRef]
- Mastroianni, G.; Milovanović, G.V. Interpolation Processes—Basic Theory and Applications; Springer Monographs in Mathematics; Springer: Berlin, Germany, 2008. [Google Scholar]
- Ditzian, Z.; Totik, V. Moduli of Smoothness; Springer Series in Computational Mathematics; Springer: New York, NY, USA, 1987; Volume 9. [Google Scholar]
- Mastroianni, G.; Russo, M.G.; Themistoclakis, W. Numerical Methods for Cauchy Singular Integral Equations in Spaces of Weighted Continuous Functions. Oper. Theory Adv. Appl. 2005, 160, 311–336. [Google Scholar]
- Gautschi, W. On the Construction of Gaussian Quadrature Rules from Modified Moments. Math. Comput. 1970, 24, 245–260. [Google Scholar]
- Piessens, R.; Branders, M. Numerical Solution of Integral equations of Mathematical Physics using Chebyshev Polynomials. J. Comput. Phys. 1976, 21, 178–196. [Google Scholar] [CrossRef]
- Nevai, P. Mean Convergence of Lagrange Interpolation III. Trans. Am. Math. Soc. 1984, 282, 669–698. [Google Scholar] [CrossRef]
- Atkinson, K. The Numerical Solution of Integral Equations of the Second Kind; Cambridge Monographs on Applied and Computational Mathematics; Cambridge University Press: Cambridge, UK, 1997. [Google Scholar]
- Kress, R. Linear Integral Equations, 2nd ed.; Applied Mathematical Sciences; Springer: New York, NY, USA, 1999; Volume 82. [Google Scholar]
- Occorsio, D.; Serafini, G. Cubature formulae for nearly singular and highly oscillating integrals. Calcolo 2018, 55, 33. [Google Scholar] [CrossRef]
- Dobbelaere, D.; Rogier, H.; Zutter, D.D. Accurate 2.5-D boundary element method for conductive media. Radio Sci. 2014, 49, 389–399. [Google Scholar] [CrossRef][Green Version]
- Timan, A.F. Theory of Approximation of Functions of a Real Variable; Dover Publications: New York, NY, USA, 1994. [Google Scholar]
- De Bonis, M.C.; Pastore, P. A quadrature formula for integrals of highly oscillatory functions. Rend. Circ. Mat. Palermo 2010, 2 (Suppl. 82), 279–303. [Google Scholar]
- Fermo, L.; Russo, M.G.; Serafini, G. Numerical treatment of the generalized Love integral equation. Numer. Algorithms 2021, 86, 1769–1789. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).