1. Introduction
When studying various applied problems related to the properties of media with a period structure, it is necessary to study differential equations with rapidly oscillating irregularities. Equations of this type are often found, for example, in electrical tric systems under the influence of high-frequency external forces. The presence of such forces creates serious problems for the numerical integration of the corresponding differential equations. Therefore, asymptotic methods are usually applied to such equations, the most famous of which are the Feshchenko–Shkil–Nikolenko splitting method [
1,
2] and the Lomov’s regularization method [
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14]. The splitting method is especially effective when applied to equations with a rapidly oscillating inhomogeneity, and in the case of an inhomogeneity containing both rapidly and slow components, the Lomov’s regularization method turned out to be the most effective. However, both of these methods were developed mainly for singularly perturbed equations that do not contain an integral operator. The transition from differential equations to integro-differential equations requires a significant restructuring of the algorithm of the regularization method. The integral term generates new types of singularities in solutions that differ from the previously known ones, which complicates the development of the algorithm for the regularization method. As far as we know, the splitting method has not been applied to integro-differential equations.
In the present study, the Lomov’s regularization method is generalized to previously not studied classes of problems such as
      
      where 
 is an unknown vector function, 
 is 
–matrix, 
 are known vector functions, 
 is the frequency of a rapidly oscillating inhomogeneity, 
 is a small parameter. Problem (1) will be considered under the following conditions:
- (1)
 - (2)
 The spectrum  of the matrix  and the frequency  satisfy for all  the requirements:
- (a)
 - (b)
 
Thus, problem (1) is considered (in the terminology of [
3]) in the nonresonant case 
 The resonant case 
  it is supposed to be considered in our subsequent work.
Note that in [
15], the scalar case of a singularly perturbed integro-differential equation with a rapidly oscillating factor for an unknown function was considered. In this paper, in contrast to [
15], we investigate a more general case of a system of integro-differential equations, and the rapidly oscillating factor is a coefficient on the right-hand side, for one of its inhomogeneities.
  2. Regularization of the Problem (1)
We denote 
 and introduce the regularizing variables
      
      and instead of problem (1) consider the problem
      
      for the function 
 which is denoted 
It is clear that, if 
 is the solution of the problem (2), then the vector function 
 is an exact solution of the problem (1); therefore, the problem (2) is extended with respect to the problem (1). However, it cannot be considered completely regularized, since the integral term
      
      has not been regularized in it. To regularize it, we introduce a class 
 that is asymptotically invariant with respect to the operator 
 (see [
3]; p. 62).
Let us first consider the space 
U of vector functions 
 represented by the sums
      
Note that in (3) elements of the space U depend on bounded in  the constant  which does not affect the development of the algorithm described below; therefore, henceforth, in the record of the element (3) of this space, for the sake of brevity, we omit the dependence on .
Let us show that the class 
 is asymptotically invariant with respect to the operator 
 The image of the operator 
J on the element (3) of the space 
U has the form:
Integrating by parts, we will have
      
Continuing this process further, we obtain the expansion
      
      where the operators are introduced:
Hence, the image of the operator 
J on an element (3) of the space 
U can be represented as a series
      
It is easy to show (see, for example, [
16], pp. 291–294) that this series converges asymptotically for 
 (uniformly in 
). This means that the class 
 is asymptotically invariant (for 
) with respect to the operator 
Let us introduce operators 
 acting on each element 
 of the form (3) according to the law: 
Now let 
 be an arbitrary continuous for 
 function, having an asymptotic expansion
      
      converging when 
 (uniformly in 
). Then the image 
 of this function expands into an asymptotic series
      
This equality is the basis for introducing an extension of the operator 
J on series of the form (5):
Although the operator 
 is defined formally, its usefulness is obvious, since in practice the 
N-th approximation of the asymptotic solution of problem (2) is usually constructed, in which only 
N-th partial sums of the series (5) will participate, which have not formal but true meaning. Now we can write down the problem that is completely regularized with respect to the original problem (1):
  3. Iterative Problems and Their Solvability in the Space U
Substituting the series (5) into (6) and equating the coefficients at the same degrees of 
 we obtain the following iterative problems: 
Each of the iterative problems (7d) can be written as
where 
 is the known function of the space 
 and the operator 
 has the form (see (4a)):
We introduce the scalar (for each 
) product in the space 
where 
 denotes the usual scalar product in a complex space 
 We prove the following statement.
Theorem 1. Let conditions (1), (2a) hold and the right-hand side  of the system (8) belongs to the space  Then, for the solvability of system (8) in U it is necessary and sufficient that the identitiesare carried out (here  are the eigenvectors of the conjugate matrix  corresponding to the eigenvalues )  Proof.  We will define the solution of the system (8) as an element (3) of the space 
U:
Substituting (10) into the system (8), we will have
Equating here separately the free term and coefficients at the same exponents, we obtain the following equations:
Due to the smoothness of the kernel  and heterogeneity , the integral Equation (11a) has a unique solution  As  is not a point in the spectrum of the matrix  (see condition 2a)), then the Equation (11b) has a unique solution  in the space 
Systems (11b) for , are solvable in the space  if and only if the identities  hold. It is easy to see that these identities coincide with the identities (9). Thus, the conditions (9) are necessary and sufficient for the solvability of the system (8) in the space U. The theorem is proved.    □
 Remark 1. If the identities (9) hold, then under the conditions (1) and (2a) the system (8) has the following solution in the space where  are eigenvectors of the matrix  corresponding to the eigenvalues  are arbitrary scalar functions  is the solution of the integral system (11a).    4. Unique Solvability of the General Iterative Problem in the Space U Remainder Theorem
As seen from (12), the solution of the system (8) is determined ambiguously. However, if its solution satisfies to the additional conditions
      
      where 
 is a known function of the space 
   is a constant number of the complex space 
 then system (8) will be uniquely solvable in the space 
 More precisely, the following result holds.
Theorem 2. Let conditions (1), (2a) be satisfied, the right-hand side  of the system (8) belongs to the space U and satisfies the orthogonality condition (9). Then system (8) under additional conditions (13a) is uniquely solvable in 
 Proof.  Under condition (9), system (8) has a solution (12) in the space 
 where the functions 
 are still arbitrary. Subordinating (12) to the first condition (13a), i.e., 
, we obtain the equation
        
        where
        
Multiplying scalarly the equality (13b) by 
 and taking into account the biorthogonality of the systems 
 and 
 we find the values
        
Let us now subordinate solution (12) to the second condition (13a). The right-hand side of this equation has the form
        
Now multiplying (15) scalarly by 
 we obtain equations
        
Adding the initial conditions (14) to them, we can uniquely find the functions 
        where 
 and hence, we define the solution (12) of the system (8) in the space in a unique way. The theorem is proved.
Applying Theorems 1 and 2 to iterative problems (7d), we uniquely find their solutions in the space 
U and construct series (5). Let 
 is the restriction of the 
N-th partial sum of series (5) for 
 The same as in [
3], it is easy to prove the following statement.    □
 Lemma 1. Let conditions (1) and (2) be satisfied. Then the partial sum  satisfies problem (1) up to  i.e.,where  for all  and for all   Consider now the following problem: 
Let us show that this problem is solvable in the space 
 (i.e., it has a solution for any right-hand side 
) and that in this case there is an estimate
      
Theorem 3. Let conditions (1) and (2) be satisfied. Then, for sufficiently small , problem (17) for any right-hand side  has a unique solution  in the space  and estimate (18) holds, where  is a constant independent of 
 Proof.  Introduce an additional unknown function
        
Differentiating it with respect to 
 we will have
        
From this and (17) it follows that the vector function 
 satisfies the following system:
        
Denote by 
 the normal fundamental matrix of the homogeneous system 
 i.e., the matrix satisfying the equation
        
Since the matrix 
 is a matrix of simple structure and its spectrum 
 lies in the half-plane 
 then the Cauchy matrix 
 is uniformly bounded, i.e.,
        
        where the constant 
 does not depend on 
 (see, for example, [
3], pp. 119–120). We now write down an integral system equivalent to system (19a):
        
Since for each 
 there exists the solution 
 of the system (19a) in the space 
 then substituting it into (19b), we obtain the identity. Let us move on to the norms:
        
        where 
 We get the inequality
        
Applying the Gronwall–Bellman lemma to this inequality, we have
        
        where 
 The theorem is proved.    □
 Theorem 4. Let conditions (1) and (2) be satisfied. Then for any  where  is small enough, the problem (1) has a unique solution ; in this case, the estimateholds true, where the constant  does not depend on   Proof.  By the lemma, the partial sum 
 satisfies the problem (16), so the remainder 
 satisfies the following problem:
        
        where 
 By Theorem 3, we have the estimate
        
        for all 
 and all 
, which means that the partial sum 
 satisfies the inequality
        
Using the inequality 
 valid for any numbers 
a and 
 we will have
        
        whence we derive the estimate
        
        where the constant 
 does not depend on 
.    □
   5. Construction of the Solution of the First Iteration Problem in the Space 
Using Theorem 1, we will try to find a solution to the first iterative problem (7d). Since the right-hand side 
 of the system (7a), satisfies the condition (9), this system has (according to (12)) a solution in the space 
U in the form
      
      where 
 are arbitrary functions, 
 is the solution of the integral system 
Subordinating (20) to the initial condition 
 we will have
      
Now, scalarly multiplying this equality by 
 we obtain values
      
For a complete calculation of the function 
, we pass to the next iterative problem (7b). Substituting the solution (20) of the system (7a) into it, we arrive at the following equation:
Subordinating the right-hand side of this system to the solvability conditions (9), we obtain the system of ordinary differential equations
      
Adding the initial condition (21) to them, we find
      
      and, therefore, the solution (15) of the problem will be found uniquely in the space 
 In this case, the leading term of the asymptotics has the following form: 
      where 
 is the solution of the integral system
      
Example 1. Let us consider the following problem:where  Let’s try to construct its main term of the asymptotic solution. In this system, the matrix is constant and has the following spectrum: 
 and a rapidly oscillating multiplier with the inhomogeneity 
 corresponds to the spectral value 
 In this case, the eigenvectors of constant matrices 
A and 
 will have the form
      
Regularizing problem (23) using the functions
      
      we get the following extended problem: 
      where 
 is extension of the integral operator 
J on series of the form
      
      with coefficients 
 from the space 
U of vector functions
      
This extension has the form 
 where the operators 
 are calculated by the formulas:
	  (operators 
 at 
 we do not write out, because we do not need them when constructing the leading term of the asymptotics). Defining the solution of the problem (24) in the form of series (25), we obtain the following iterative problems: 
The solution of the first iterative problem (26) will be as follows: 
      where 
 are arbitrary functions, 
 is the solution of the integral system
      
Subordinate (28) to the initial condition 
. Taking into account the form of vector functions (29) and 
 we obtain the equation 
 Multiplying this equality scalarly by 
 and 
, we find the values 
 For a complete calculation of the functions 
, we pass to the next iterative problem (27). Taking into account that, under the conditions of solvability (9) of problem (27) only exponentials 
 and 
 are involved, we keep in its right-hand side only terms depending on these exponentials:
      and subordinate the resulting sum to the orthogonality conditions (9). We will have
      
      or (in more detail, taking into account the eigenvectors 
 and the kernel 
):
Adding to these equations the initial conditions 
 found earlier, we uniquely find the functions 
 and hence, we will uniquely construct solution (28) of the first iterative problem (26). Making a narrowing in it at 
 we obtain the leading term of the asymptotic solution of the problem (23):
      where 
 is the solution of the integral system (29). It is seen from (30) that, at the exact solution 
 of the problem (23) does not tend to the solution 
 of the integral system (29) at 
, but performs quick oscillations near it.