1. Introduction
When studying various applied problems related to the properties of media with a period structure, it is necessary to study differential equations with rapidly oscillating irregularities. Equations of this type are often found, for example, in electrical tric systems under the influence of high-frequency external forces. The presence of such forces creates serious problems for the numerical integration of the corresponding differential equations. Therefore, asymptotic methods are usually applied to such equations, the most famous of which are the Feshchenko–Shkil–Nikolenko splitting method [
1,
2] and the Lomov’s regularization method [
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14]. The splitting method is especially effective when applied to equations with a rapidly oscillating inhomogeneity, and in the case of an inhomogeneity containing both rapidly and slow components, the Lomov’s regularization method turned out to be the most effective. However, both of these methods were developed mainly for singularly perturbed equations that do not contain an integral operator. The transition from differential equations to integro-differential equations requires a significant restructuring of the algorithm of the regularization method. The integral term generates new types of singularities in solutions that differ from the previously known ones, which complicates the development of the algorithm for the regularization method. As far as we know, the splitting method has not been applied to integro-differential equations.
In the present study, the Lomov’s regularization method is generalized to previously not studied classes of problems such as
where
is an unknown vector function,
is
–matrix,
are known vector functions,
is the frequency of a rapidly oscillating inhomogeneity,
is a small parameter. Problem (1) will be considered under the following conditions:
- (1)
- (2)
The spectrum of the matrix and the frequency satisfy for all the requirements:
- (a)
- (b)
Thus, problem (1) is considered (in the terminology of [
3]) in the nonresonant case
The resonant case
it is supposed to be considered in our subsequent work.
Note that in [
15], the scalar case of a singularly perturbed integro-differential equation with a rapidly oscillating factor for an unknown function was considered. In this paper, in contrast to [
15], we investigate a more general case of a system of integro-differential equations, and the rapidly oscillating factor is a coefficient on the right-hand side, for one of its inhomogeneities.
2. Regularization of the Problem (1)
We denote
and introduce the regularizing variables
and instead of problem (1) consider the problem
for the function
which is denoted
It is clear that, if
is the solution of the problem (2), then the vector function
is an exact solution of the problem (1); therefore, the problem (2) is extended with respect to the problem (1). However, it cannot be considered completely regularized, since the integral term
has not been regularized in it. To regularize it, we introduce a class
that is asymptotically invariant with respect to the operator
(see [
3]; p. 62).
Let us first consider the space
U of vector functions
represented by the sums
Note that in (3) elements of the space U depend on bounded in the constant which does not affect the development of the algorithm described below; therefore, henceforth, in the record of the element (3) of this space, for the sake of brevity, we omit the dependence on .
Let us show that the class
is asymptotically invariant with respect to the operator
The image of the operator
J on the element (3) of the space
U has the form:
Integrating by parts, we will have
Continuing this process further, we obtain the expansion
where the operators are introduced:
Hence, the image of the operator
J on an element (3) of the space
U can be represented as a series
It is easy to show (see, for example, [
16], pp. 291–294) that this series converges asymptotically for
(uniformly in
). This means that the class
is asymptotically invariant (for
) with respect to the operator
Let us introduce operators
acting on each element
of the form (3) according to the law:
Now let
be an arbitrary continuous for
function, having an asymptotic expansion
converging when
(uniformly in
). Then the image
of this function expands into an asymptotic series
This equality is the basis for introducing an extension of the operator
J on series of the form (5):
Although the operator
is defined formally, its usefulness is obvious, since in practice the
N-th approximation of the asymptotic solution of problem (2) is usually constructed, in which only
N-th partial sums of the series (5) will participate, which have not formal but true meaning. Now we can write down the problem that is completely regularized with respect to the original problem (1):
3. Iterative Problems and Their Solvability in the Space U
Substituting the series (5) into (6) and equating the coefficients at the same degrees of
we obtain the following iterative problems:
Each of the iterative problems (7d) can be written as
where
is the known function of the space
and the operator
has the form (see (4a)):
We introduce the scalar (for each
) product in the space
where
denotes the usual scalar product in a complex space
We prove the following statement.
Theorem 1. Let conditions (1), (2a) hold and the right-hand side of the system (8) belongs to the space Then, for the solvability of system (8) in U it is necessary and sufficient that the identitiesare carried out (here are the eigenvectors of the conjugate matrix corresponding to the eigenvalues ) Proof. We will define the solution of the system (8) as an element (3) of the space
U:
Substituting (10) into the system (8), we will have
Equating here separately the free term and coefficients at the same exponents, we obtain the following equations:
Due to the smoothness of the kernel and heterogeneity , the integral Equation (11a) has a unique solution As is not a point in the spectrum of the matrix (see condition 2a)), then the Equation (11b) has a unique solution in the space
Systems (11b) for , are solvable in the space if and only if the identities hold. It is easy to see that these identities coincide with the identities (9). Thus, the conditions (9) are necessary and sufficient for the solvability of the system (8) in the space U. The theorem is proved. □
Remark 1. If the identities (9) hold, then under the conditions (1) and (2a) the system (8) has the following solution in the space where are eigenvectors of the matrix corresponding to the eigenvalues are arbitrary scalar functions is the solution of the integral system (11a). 4. Unique Solvability of the General Iterative Problem in the Space U Remainder Theorem
As seen from (12), the solution of the system (8) is determined ambiguously. However, if its solution satisfies to the additional conditions
where
is a known function of the space
is a constant number of the complex space
then system (8) will be uniquely solvable in the space
More precisely, the following result holds.
Theorem 2. Let conditions (1), (2a) be satisfied, the right-hand side of the system (8) belongs to the space U and satisfies the orthogonality condition (9). Then system (8) under additional conditions (13a) is uniquely solvable in
Proof. Under condition (9), system (8) has a solution (12) in the space
where the functions
are still arbitrary. Subordinating (12) to the first condition (13a), i.e.,
, we obtain the equation
where
Multiplying scalarly the equality (13b) by
and taking into account the biorthogonality of the systems
and
we find the values
Let us now subordinate solution (12) to the second condition (13a). The right-hand side of this equation has the form
Now multiplying (15) scalarly by
we obtain equations
Adding the initial conditions (14) to them, we can uniquely find the functions
where
and hence, we define the solution (12) of the system (8) in the space in a unique way. The theorem is proved.
Applying Theorems 1 and 2 to iterative problems (7d), we uniquely find their solutions in the space
U and construct series (5). Let
is the restriction of the
N-th partial sum of series (5) for
The same as in [
3], it is easy to prove the following statement. □
Lemma 1. Let conditions (1) and (2) be satisfied. Then the partial sum satisfies problem (1) up to i.e.,where for all and for all Consider now the following problem:
Let us show that this problem is solvable in the space
(i.e., it has a solution for any right-hand side
) and that in this case there is an estimate
Theorem 3. Let conditions (1) and (2) be satisfied. Then, for sufficiently small , problem (17) for any right-hand side has a unique solution in the space and estimate (18) holds, where is a constant independent of
Proof. Introduce an additional unknown function
Differentiating it with respect to
we will have
From this and (17) it follows that the vector function
satisfies the following system:
Denote by
the normal fundamental matrix of the homogeneous system
i.e., the matrix satisfying the equation
Since the matrix
is a matrix of simple structure and its spectrum
lies in the half-plane
then the Cauchy matrix
is uniformly bounded, i.e.,
where the constant
does not depend on
(see, for example, [
3], pp. 119–120). We now write down an integral system equivalent to system (19a):
Since for each
there exists the solution
of the system (19a) in the space
then substituting it into (19b), we obtain the identity. Let us move on to the norms:
where
We get the inequality
Applying the Gronwall–Bellman lemma to this inequality, we have
where
The theorem is proved. □
Theorem 4. Let conditions (1) and (2) be satisfied. Then for any where is small enough, the problem (1) has a unique solution ; in this case, the estimateholds true, where the constant does not depend on Proof. By the lemma, the partial sum
satisfies the problem (16), so the remainder
satisfies the following problem:
where
By Theorem 3, we have the estimate
for all
and all
, which means that the partial sum
satisfies the inequality
Using the inequality
valid for any numbers
a and
we will have
whence we derive the estimate
where the constant
does not depend on
. □
5. Construction of the Solution of the First Iteration Problem in the Space
Using Theorem 1, we will try to find a solution to the first iterative problem (7d). Since the right-hand side
of the system (7a), satisfies the condition (9), this system has (according to (12)) a solution in the space
U in the form
where
are arbitrary functions,
is the solution of the integral system
Subordinating (20) to the initial condition
we will have
Now, scalarly multiplying this equality by
we obtain values
For a complete calculation of the function
, we pass to the next iterative problem (7b). Substituting the solution (20) of the system (7a) into it, we arrive at the following equation:
Subordinating the right-hand side of this system to the solvability conditions (9), we obtain the system of ordinary differential equations
Adding the initial condition (21) to them, we find
and, therefore, the solution (15) of the problem will be found uniquely in the space
In this case, the leading term of the asymptotics has the following form:
where
is the solution of the integral system
Example 1. Let us consider the following problem:where Let’s try to construct its main term of the asymptotic solution. In this system, the matrix is constant and has the following spectrum:
and a rapidly oscillating multiplier with the inhomogeneity
corresponds to the spectral value
In this case, the eigenvectors of constant matrices
A and
will have the form
Regularizing problem (23) using the functions
we get the following extended problem:
where
is extension of the integral operator
J on series of the form
with coefficients
from the space
U of vector functions
This extension has the form
where the operators
are calculated by the formulas:
(operators
at
we do not write out, because we do not need them when constructing the leading term of the asymptotics). Defining the solution of the problem (24) in the form of series (25), we obtain the following iterative problems:
The solution of the first iterative problem (26) will be as follows:
where
are arbitrary functions,
is the solution of the integral system
Subordinate (28) to the initial condition
. Taking into account the form of vector functions (29) and
we obtain the equation
Multiplying this equality scalarly by
and
, we find the values
For a complete calculation of the functions
, we pass to the next iterative problem (27). Taking into account that, under the conditions of solvability (9) of problem (27) only exponentials
and
are involved, we keep in its right-hand side only terms depending on these exponentials:
and subordinate the resulting sum to the orthogonality conditions (9). We will have
or (in more detail, taking into account the eigenvectors
and the kernel
):
Adding to these equations the initial conditions
found earlier, we uniquely find the functions
and hence, we will uniquely construct solution (28) of the first iterative problem (26). Making a narrowing in it at
we obtain the leading term of the asymptotic solution of the problem (23):
where
is the solution of the integral system (29). It is seen from (30) that, at the exact solution
of the problem (23) does not tend to the solution
of the integral system (29) at
, but performs quick oscillations near it.