Parameterization of the Stochastic Model for Evaluating Variable Small Data in the Shannon Entropy Basis

The article analytically summarizes the idea of applying Shannon’s principle of entropy maximization to sets that represent the results of observations of the “input” and “output” entities of the stochastic model for evaluating variable small data. To formalize this idea, a sequential transition from the likelihood function to the likelihood functional and the Shannon entropy functional is analytically described. Shannon’s entropy characterizes the uncertainty caused not only by the probabilistic nature of the parameters of the stochastic data evaluation model but also by interferences that distort the results of the measurements of the values of these parameters. Accordingly, based on the Shannon entropy, it is possible to determine the best estimates of the values of these parameters for maximally uncertain (per entropy unit) distortions that cause measurement variability. This postulate is organically transferred to the statement that the estimates of the density of the probability distribution of the parameters of the stochastic model of small data obtained as a result of Shannon entropy maximization will also take into account the fact of the variability of the process of their measurements. In the article, this principle is developed into the information technology of the parametric and non-parametric evaluation on the basis of Shannon entropy of small data measured under the influence of interferences. The article analytically formalizes three key elements: -instances of the class of parameterized stochastic models for evaluating variable small data; -methods of estimating the probability density function of their parameters, represented by normalized or interval probabilities; -approaches to generating an ensemble of random vectors of initial parameters.


Introduction
One of the most relevant problems of modern science is the extraction of useful information from available data. In various fields of science, methodologies aimed at solving this problem are being developed. Each such methodology is based on a certain hypothesis about the properties of the data and the real or hypothetical source of their origin. In the context of the data evaluation problem, two fundamental hypotheses can be distinguished [1][2][3][4][5]. The first hypothesis focuses on directly measurable, deterministic parameters to identify potential functional dependencies between them. All data that cannot be attributed to one or more defined parameters are considered influences in this hypothesis and are rejected. Naturally, such an approach is adequate and productive only if the information is extracted from data obtained from a known, sufficiently investigated the source of origin. The second hypothesis focuses on the analysis of the data as such and is focused on identifying patterns in them, the presence of which can be assessed using a certain defined metric. This can be, for example, a measure of data sufficiency, a property of a sample from the general population, the normality of probability distribution densities, etc. It is practically impossible to guarantee the characteristics of these properties for specific data. However, the improbable becomes common if we analyze not data, but Big Data. This trend is the basis for the progress of such methodologies as mathematical statistics [2,[6][7][8], machine learning [9][10][11][12], econometrics [13][14][15][16], financial mathematics [17][18][19] and control theory [20][21][22][23].
In recent decades, the first two of the methodologies just mentioned have been heard. Machine learning is based on the axiomatic perception of probability spaces, as outlined in the paradigm of the theory of statistical learning developed in the 1960s [24][25][26]. There are several dominant categories of machine learning, but the most common is tutored learning [9,10,27,28]. In this category, researchers work with symmetric finite datasets, summarized in the "input" and "output" entities. The purpose of data analysis is to identify the functional dependence between these entities. The set of admissible types of functions forms the hypothesis space of this category of machine learning. The machine learning algorithm consistently evaluates the expected risks of describing the dependence of the existing "input" and "output" entities by each type of function from the hypothesis space. The evaluation is carried out by calculating a single loss function for the entire research. The expected risk is understood as the product of the sum of the estimates and the probability distribution of the data. If the compatible mapping probability distribution is known, then finding the best hypothesis is a trivial task. In the general case, the distribution is unknown, so the machine learning algorithm chooses the most appropriate hypothesis according to a certain rule and proves this thesis by calculating the empirical risk. In addition to the computational complexity, the disadvantage of machine learning is the tendency of the algorithms of this methodology to minimize the loss function by overfitting the potentially best hypothesis to the available data (so-called overtraining [9,27,29]). A typical way to detect (but not prevent) overtraining is to test the best hypothesis on data that the algorithm has not yet worked on (the control sample). Methods of mathematical statistics are not subject to retraining, because they do not assess empirical risk as such.
A typical example of a problem, in the process of solving which the characteristic features of mathematical statistics and machine learning are manifested, is linear regression [7][8][9][10][11]. In the classic formulation of this problem, we need to find the regression coefficients that minimize the root mean square error between the reference entity "output" and its pattern as generated by the model. Such a problem can be solved in a closed form. The theory of statistical learning states that, if we choose the root mean square error as the loss function and carry out empirical risk optimization, then the obtained result will coincide with the one that we will obtain by applying traditional linear regression analysis. However, the maximum likelihood method [2,6,7,30] characteristic for mathematical statistics will demonstrate a similar result in this situation. By the way, the methods of mathematical statistics do not operate with the concepts of initial and test samples, but use metrics to evaluate the results of the model. In our example, the statistical approach allows us to reach the optimal solution because the solution itself exists in a closed form. The maximum likelihood method does not test alternative hypotheses and does not converge to the optimal solution, unlike a machine learning algorithm. However, if the piecewise linear loss function is used for the machine learning algorithm in the same problem, the final result does not coincide with the maximum likelihood method. The machine learning algorithm allows us to expand the space of relevant hypotheses with an a priori considered loss function. The process of their evaluation is carried out automatically. The maximum likelihood method can estimate the accuracy of the original model but does not allow us to automatically change its appearance. Therefore, the methods of machine learning and mathematical statistics work in different ways, while producing similar results. If the task of the researcher is to accurately predict the cost of housing, then machine learning tools are exactly what is needed. If a scientist is investigating the relationships between parameters or making scientifically based conclusions about the data, then a statistical model cannot be dispensed with.
Finally, machine learning experts say, "There are no such things as unsolvable problems, either data or computing power is scarce". Indeed, everyone has heard about Big Data analysis [10][11][12]31]. Now, however, the issue of analyzing so-called "small data" is becoming increasingly common [32,33]. Classical machine learning approaches are helpless in such a situation. This circumstance prompted the authors to write this article.
Taking into account the strengths and weaknesses of the mentioned methods, we will formulate the necessary attributes of scientific research.
The object of the research is the process of the parameterization of the stochastic model for evaluating variable small data for machine learning purposes.
The research subject is probability theory and mathematical statistics, evaluation theory, information theory, mathematical programming methods and experiment planning theory.
The research aims to formalize the process of finding the best estimates of the probability density functions for the characteristic parameters of instances of the class of stochastic models for evaluating variable small data.
The research objectives are: (1) To formalize the process of calculating the variable entropy estimation of the probability density functions of the characteristic parameters of the stochastic variable small data estimation model, represented by normalized probabilities; (2) To formalize the process of calculating the variable entropy estimation of the probability density functions of the characteristic parameters of the stochastic variable small data estimation model, represented by interval probabilities; (3) To justify the adequacy of the proposed mathematical apparatus and demonstrate its functionality with an example.
The main contribution of the research is that the article analytically summarizes the idea of applying the Shannon entropy maximization principle to sets that represent the results of observations of the "input" and "output" entities of the stochastic model for evaluating variable small data. To formalize this idea, a sequential transition from the likelihood function to the likelihood functional and the Shannon entropy functional is analytically described. Shannon's entropy characterizes the uncertainty caused not only by the probabilistic nature of the parameters of the stochastic data evaluation model but also by influences that distort the results of the measurements of the values of these parameters. Accordingly, based on the Shannon entropy, it is possible to determine the best estimates of the values of these parameters for maximally uncertain (per entropy unit) influences that cause measurement variability. This postulate is organically transferred to the statement that the estimates of the probability distribution density of the parameters of the stochastic model of small data obtained as a result of Shannon entropy maximization will also take into account the fact of the variability of the process of their measurements. In the article, this principle is developed into the information technology of parametric and non-parametric evaluation on the basis of Shannon entropy of small data measured under the influence of interferences.
The highlights of the research are: (1) Instances of the class of parameterized stochastic models for evaluating variable small data; (2) Methods of estimating the probability density function of their parameters, represented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial parameters; (4) A technique for statistical processing of such an ensemble using the Monte Carlo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of physical, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
Let there be a stochastic parameterized research object represented by the results of measurements, in which the matrix of values of the input parameters X with the dimension [o × n] (entity "input") is matched by a vector of values of the output parameter y with the dimension [o × 1] (entity "output"), where o is the number of censored observations, and n is the number of input characteristic parameters of the research object.
The process of measuring the values of matrix X and vector y is characterized by errors, which are represented by the symmetrical matrix N = ν ji (variability of the measurement process), i = 1, n, j = 1, o, and vector υ = (υ i ), where ν ji , υ i are independent stochastic values, ∀i, j. The value of these stochastic quantities belongs to the intervals N ji = ν − ji , ν + ji and (1) Instances of the class of parameterized stochastic models for evaluating variable small data; (2) Methods of estimating the probability density function of their parameters, represented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial parameters; (4) A technique for statistical processing of such an ensemble using the Monte Carlo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of physical, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
Let there be a stochastic parameterized research object represented by the results of measurements, in which the matrix of values of the input parameters X with the dimen- where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters Let us assume that the parameters of the stochastic model and the variability of the measurements are continuous stochastic quantities, the values of which belong to the corresponding intervals of the tuple ,, = υ − j , υ + j , respectively. The stochastic model of the X, y data evaluation is represented by an expression where F is a defined o-dimensional vector function, α is a random n-dimensional vector formed by independent stochastic parameters (1) Instances of the class of parameterized stochastic models for evaluating vari small data; (2) Methods of estimating the probability density function of their parameters, resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial paramete (4) A technique for statistical processing of such an ensemble using the Monte C method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of ph cal, medical, economic, biological and other sources of origin is the functional purpos evaluation theory as a branch of mathematical statistics. To solve the problem of eva tion, parametric and non-parametric approaches are used. In recent decades, the latter noticeably dominated the former, which has become possible thanks to the "reactive" gress in the field of machine learning and artificial intelligence. At the same time, the fo of researchers' interest is shifting from the study of the processes represented by Big D to that of those processes about which the amount of data small, and the data itself tains errors. Such a preamble encourages the perception of the parameters of the s data evaluation model as stochastic quantities. Accordingly, we will call such a mod stochastic model for small data evaluation. The characteristics of such a model are probability density functions of the stochastic parameters. The primary task in identify a stochastic estimation model for specific small data is to estimate the parameters of t probability density functions. If this step is passed, then the identified stochastic eva tion model can be taken as a basis for forming moment models of small data, genera an ensemble of random vectors of the initial parameters and carrying out the statis processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to desired numerical characteristics. The formalization of the way to solve the primary p lem formulated above has scientific potential and applied value. , A i (hereinafter-the "genuine" version of the stochastic Model (1) or GνSM).
In this case, the probability density functions of the stochastic parameters of GνSM (variability of measurements P(α), input W(N) and output Q(υ) parameters) (Independent Stochastic Parameters of the Small Data Estimation Model) are described by the expressions: where α i ∈ A i , ν ji ∈ N ji and υ j ∈ measurements, in which the matrix of values of the input parameters X with the dimension   оn  (entity "input") is matched by a vector of values of the output parameter у with the dimension   1 о (entity "output"), where о is the number of censored observations, and n is the number of input characteristic parameters of the research object.
where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters і  , Let us assume that the parameters of the stochastic model and the variability of the measurements are continuous stochastic quantities, the values of which belong to the corresponding intervals of the tuple , respectively. Formulating Expressions (2)-(4), the authors implied a priori that the measurement results were obtained in accordance with the provisions of the experiment planning theory. The corresponding variables are statistically independent.
The stochastic Model (1) generates an ensemble of random vectors s, which can be compared with the vector υ obtained as a result of measurements. To carry out such an estimation of the probability density Functions (2)-(4), we will use k moments of the stochastic components of the vector s: where (Numerical characteristics for estimating these stochastic parameters) cal, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
Let there be a stochastic where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters і  , Let us assume that the parameters of the stochastic model and the variability of the measurements are continuous stochastic quantities, the values of which belong to the corresponding intervals of the tuple ,, Next, we will use moments of the first order (k = 1). In accordance: (3) Approaches to generating an ensemble of random vectors of initial parameters; (4) A technique for statistical processing of such an ensemble using the Monte Carlo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of physical, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
Let there be a stochastic where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters Let us assume that the parameters of the stochastic model and the variability of the measurements are continuous stochastic quantities, the values of which belong to the corresponding intervals of the tuple ,, Another version of the implementation of the Model (1) will be one in which the parameters of the stochastic model and the variability of the measurements are continuous stochastic values, the belonging of which to the corresponding interval of the tuple N ji , (1) Instances of the class of parameterized stochastic models for evaluating variable small data; (2) Methods of estimating the probability density function of their parameters, represented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial parameters; (4) A technique for statistical processing of such an ensemble using the Monte Carlo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of physical, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
Let there be a stochastic parameterized research object represented by the results of measurements, in which the matrix of values of the input parameters X with the dimen- where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters Let us assume that the parameters of the stochastic model and the variability of the measurements are continuous stochastic quantities, the values of which belong to the corresponding intervals of the tuple ,, (hereinafter-the "genuine" version of the stochastic model (1) or GvSM ).
, A i will be characterized by a certain probability (hereinafter-the "quasi" version of the stochastic Model (1) or QνSM). In this case: (1) the parameters α i take values in A i intervals with probabilities p i ∈ [0, 1], i = 1, n; (2) the parameters ν ji take values in intervals N ji with probabilities w ji ∈ [0, 1], j = 1, o, i = 1, n; (3) the parameters υ j take values in intervals (1) Instances of the class of parameterized stochastic models for evaluati small data; (2) Methods of estimating the probability density function of their param resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial pa (4) A technique for statistical processing of such an ensemble using the M method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomen cal, medical, economic, biological and other sources of origin is the functional evaluation theory as a branch of mathematical statistics. To solve the problem tion, parametric and non-parametric approaches are used. In recent decades, th noticeably dominated the former, which has become possible thanks to the "re gress in the field of machine learning and artificial intelligence. At the same tim of researchers' interest is shifting from the study of the processes represented to that of those processes about which the amount of data small, and the dat tains errors. Such a preamble encourages the perception of the parameters o data evaluation model as stochastic quantities. Accordingly, we will call such stochastic model for small data evaluation. The characteristics of such a mo probability density functions of the stochastic parameters. The primary task in a stochastic estimation model for specific small data is to estimate the paramet probability density functions. If this step is passed, then the identified stocha tion model can be taken as a basis for forming moment models of small data, an ensemble of random vectors of the initial parameters and carrying out th processing of such an ensemble using the Monte Carlo method [6][7][8] to bri desired numerical characteristics. The formalization of the way to solve the pri lem formulated above has scientific potential and applied value.
Let there be a stochastic Let us assume that the parameters of the stochastic model and the varia measurements are continuous stochastic quantities, the values of which belong responding intervals of the tuple ,, The available a priori information is summarized by the vector At the same time, Expressions (2)-(4) retain their legitimacy. We generalize the initial numerical characteristics of QνSM in the form of a vector of quasi-momentums of the first order: where The analytical expression for the first-order quasi-momentum of the stochastic vector s can be obtained by substituting numerical Characteristics (6) into Expression (1): In the context of the proposed statement of the research, we specify its aim and objectivities.
The research aims to formalize the process of finding the best estimates of the probability density functions for the p, q parameters of GνSM and QνSM represented by Expressions (5) and (7), respectively.
The objectives of the research are: (1) To formalize the process of calculating the variable entropy estimation of the probability density functions of characteristic parameters of GνSM represented by normalized probabilities; (2) To formalize the process of calculating the variable entropy estimation of the probability density functions of characteristic parameters of QνSM represented by interval probabilities; (3) To justify the adequacy of the proposed mathematical apparatus and demonstrate its functionality with an example.

Parameterization of the Stochastic Model for Evaluating Variable Small Data in the Shannon Entropy Basis
Let us formulate the corresponding probability functionals for the available information about the values of the input and output parameters of the stochastic Model (1).
Taking into account the independence of the parameters of the "input" and "output" entities in the stochastic Model (1) and the variability of their measurement procedure, we determine the compatible probability density function Φ(α, ν, υ) and the corresponding logarithmic likelihood ratio φ(α, ν, υ) as Based on Expressions (8) and (9), we formulate the likelihood functional L(P(α), W(ν), Q(υ)): (1) Instances of the class of parameterized stochastic models for evaluating variable small data; (2) Methods of estimating the probability density function of their parameters, represented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial parameters; (4) A technique for statistical processing of such an ensemble using the Monte Carlo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of physical, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
Let there be a stochastic where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters і  , Let us assume that the parameters of the stochastic model and the variability of the measurements are continuous stochastic quantities, the values of which belong to the corresponding intervals of the tuple ,, (hereinafter-the "genuine" version of the stochastic model (1) or (1) Instances of the class of parameterized stochastic models for evaluating small data; (2) Methods of estimating the probability density function of their paramet resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial param (4) A technique for statistical processing of such an ensemble using the Mon method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena o cal, medical, economic, biological and other sources of origin is the functional pu evaluation theory as a branch of mathematical statistics. To solve the problem o tion, parametric and non-parametric approaches are used. In recent decades, the l noticeably dominated the former, which has become possible thanks to the "react gress in the field of machine learning and artificial intelligence. At the same time, of researchers' interest is shifting from the study of the processes represented by to that of those processes about which the amount of data small, and the data it tains errors. Such a preamble encourages the perception of the parameters of t data evaluation model as stochastic quantities. Accordingly, we will call such a stochastic model for small data evaluation. The characteristics of such a mode probability density functions of the stochastic parameters. The primary task in ide a stochastic estimation model for specific small data is to estimate the parameters probability density functions. If this step is passed, then the identified stochastic tion model can be taken as a basis for forming moment models of small data, ge an ensemble of random vectors of the initial parameters and carrying out the s processing of such an ensemble using the Monte Carlo method [6][7][8] to bring desired numerical characteristics. The formalization of the way to solve the prima lem formulated above has scientific potential and applied value.
Let there be a stochastic .
Expression (10) presented in the −L(P(α), W(ν), Q(υ)) format is the Shannon entropy functional [34,35]. According to its purpose, such a functional is a measure for evaluating the degree of variability of the elements of a tuple P(α), W(ν), Q(υ) . This fact determines the perspective of using such a functional for evaluating Functions (2)-(4). In the context of this motivation, let us transform Expression (10) into the form (1) Instances of the class of parameterized stochastic models for evaluating variable small data; (2) Methods of estimating the probability density function of their parameters, represented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial parameters; (4) A technique for statistical processing of such an ensemble using the Monte Carlo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena of physical, medical, economic, biological and other sources of origin is the functional purpose of evaluation theory as a branch of mathematical statistics. To solve the problem of evaluation, parametric and non-parametric approaches are used. In recent decades, the latter has noticeably dominated the former, which has become possible thanks to the "reactive" progress in the field of machine learning and artificial intelligence. At the same time, the focus of researchers' interest is shifting from the study of the processes represented by Big Data to that of those processes about which the amount of data small, and the data itself contains errors. Such a preamble encourages the perception of the parameters of the small data evaluation model as stochastic quantities. Accordingly, we will call such a model a stochastic model for small data evaluation. The characteristics of such a model are the probability density functions of the stochastic parameters. The primary task in identifying a stochastic estimation model for specific small data is to estimate the parameters of these probability density functions. If this step is passed, then the identified stochastic evaluation model can be taken as a basis for forming moment models of small data, generating an ensemble of random vectors of the initial parameters and carrying out the statistical processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it to the desired numerical characteristics. The formalization of the way to solve the primary problem formulated above has scientific potential and applied value.
where F is a defined о -dimensional vector function,  is a random n -dimensional vector formed by independent stochastic parameters The Functional (11) is defined for estimating the probability density functions of stochastic parameters of GνSM. For QνSM, based on Expression (10), we obtain: Based on Definition (11), we formulate the problem of finding the optimal estimate of the probability density functions of stochastic parameters of GνSM, taking into account the fact of their variability, i.e., E s .
We define the objective function of such an optimization problem as: We define the restrictions of the E s optimization problem as that is, the probability distribution density of the variability of measurements P(α) ∈ P, input W(N) ∈ W and output Q(υ) ∈ Q parameters of GνSM must belong to the space E defined by Expression (13), and that is, the elements of the vector with the results of measurements y are equal to the elements of the kth moment of the vector s raised to the k −1 th power. By analogy with the formulation of the optimization Problem (12)- (14), we formulate the problem of finding the optimal estimate of the probability density functions of stochastic parameters of QνSM, taking into account the fact of their variability, i.e., E s .
We define the objective function of such an optimization problem as: Recall that the complex parameter s generalizes a tuple of interval controlled parameters P(α), W(ν), Q(υ) (see Expressions (10) and (5)), and the complex parameter s focuses on the variability of measuring these characteristic parameters (see Expression (7)).
Considerations regarding the formulation of restrictions for finding the extremum of the objective Function (15) are identical to those embodied in Restrictions (13) and (14). At the same time, Restriction (13) fully satisfies the statement of the Problem (15), while Restriction (14) can be written in terms of the definition of QνSM: Let us pay attention to the situation when the measurement errors υ(t) and the values of the vector of the initial parameters of the stochastic model s(t) are characterized by non-linearity of the rth degree: where α = (α i ) is a vector of parameters, the independent stochastic elements of which take values from the ranges A i = α − i , α + i with the probability distribution densities p i (α i ), i = 1, n. The measurement of the components of the entities "input" and "output" of the investigated process takes place at moments t j , j = 1, o. The entity "input" is represented by a set of r-matrices, r = 1, R, of the form and the entity "output" is represented by stochastic elements of the vector s = s t j , where the independent elements of the vector of the variability of measurements of the entity "output" υ take values in intervals gress in the field of machine learning and artificial intelligence. At the same time, the of researchers' interest is shifting from the study of the processes represented by Big to that of those processes about which the amount of data small, and the data itsel tains errors. Such a preamble encourages the perception of the parameters of the data evaluation model as stochastic quantities. Accordingly, we will call such a mo stochastic model for small data evaluation. The characteristics of such a model a probability density functions of the stochastic parameters. The primary task in ident a stochastic estimation model for specific small data is to estimate the parameters of probability density functions. If this step is passed, then the identified stochastic ev tion model can be taken as a basis for forming moment models of small data, gene an ensemble of random vectors of the initial parameters and carrying out the stat processing of such an ensemble using the Monte Carlo method [6][7][8] to bring it desired numerical characteristics. The formalization of the way to solve the primary lem formulated above has scientific potential and applied value.
Let there be a stochastic Let us identify and investigate the variable entropy estimate of the probability density functions P(α) = (p i (α i )), i = 1, n, and Q(υ) = q j υ j , j = 1, o.
We present the objective function of the optimization Problem (12)- (14) in the form We present the system of Restrictions (13) and (14) in the form (2) Methods of estimating the probability density function of their paramet resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial param (4) A technique for statistical processing of such an ensemble using the Mon method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena o cal, medical, economic, biological and other sources of origin is the functional pu evaluation theory as a branch of mathematical statistics. To solve the problem of tion, parametric and non-parametric approaches are used. In recent decades, the la noticeably dominated the former, which has become possible thanks to the "reacti gress in the field of machine learning and artificial intelligence. At the same time, t of researchers' interest is shifting from the study of the processes represented by B to that of those processes about which the amount of data small, and the data its tains errors. Such a preamble encourages the perception of the parameters of th data evaluation model as stochastic quantities. Accordingly, we will call such a stochastic model for small data evaluation. The characteristics of such a model probability density functions of the stochastic parameters. The primary task in ide a stochastic estimation model for specific small data is to estimate the parameters probability density functions. If this step is passed, then the identified stochastic tion model can be taken as a basis for forming moment models of small data, gen an ensemble of random vectors of the initial parameters and carrying out the st processing of such an ensemble using the Monte Carlo method [6][7][8] to bring desired numerical characteristics. The formalization of the way to solve the prima lem formulated above has scientific potential and applied value.
Let there be a stochastic q j υ j dυ j = 0, (1) Instances of the class of parameterized stochastic mod small data; (2) Methods of estimating the probability density functio resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vec (4) A technique for statistical processing of such an ensem method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signa cal, medical, economic, biological and other sources of origin i evaluation theory as a branch of mathematical statistics. To so tion, parametric and non-parametric approaches are used. In re noticeably dominated the former, which has become possible th gress in the field of machine learning and artificial intelligence. of researchers' interest is shifting from the study of the process to that of those processes about which the amount of data sm tains errors. Such a preamble encourages the perception of th data evaluation model as stochastic quantities. Accordingly, w stochastic model for small data evaluation. The characteristic probability density functions of the stochastic parameters. The a stochastic estimation model for specific small data is to estim probability density functions. If this step is passed, then the id tion model can be taken as a basis for forming moment model an ensemble of random vectors of the initial parameters and processing of such an ensemble using the Monte Carlo meth desired numerical characteristics. The formalization of the way lem formulated above has scientific potential and applied valu Let there be a stochastic parameterized research object re measurements, in which the matrix of values of the input param sion s of the probability density functions P(α) and Q(υ) belong to continuously differentiable functions, respectively: where a i , b i , c j , d j are fixed coefficients, i = 1, n, j = 1, o. The conclusion generalized by Expressions (17) and (18) can be interpreted as follows: (1) For a linear stochastic model of estimation of variable small data: entropy estimates E (1) s are always exponential functions. The results of measuring the entities "input" and "output" of the investigated process determine the form, and not the type, of the s -functions depends on the organization of the measurement process of these "input" and "output" entities. Therefore, it remains to formalize the variable entropy estimates E (1) s E (1) s of the probability density functions p and q of the parameters of GνSM(QνSM), respectively. Let us investigate the linear GνSM without taking into account the variability of the measurement of the "input" entity: where Ξ(α − , υ − ) = Xα − + υ − . We define the a priori probabilities by the elements of the tuple p 0 , q 0 . Let us present the objective function of the optimization Problem (15) and (16) in the form and the system of restrictions we present in the form In terms of the Lagrange function, we present the solution of the mathematical programming Problem (20) and (21) as where β µ are fixed coefficients and ψ = (ψ 1 , . . . , ψ o ) is a set of Lagrange multipliers.
Entropy estimates E (1) Now let us investigate how the formulation and solution of the optimization Problem (20) and (21) will change if interval restrictions 0 ≤ p i ≤ 1, ∀i ∈ 1, n , 0 ≤ q j ≤ 1, ∀j ∈ 1, o , are respectively imposed on the values of the elements of the stochastic vectors p, q .
Under such conditions, the variable entropy estimate E where p 0 (24) is supplemented by the adapted balance Equation (21):

The objective Function
Applying the method of Lagrange multipliers [7,8,36], the extreme entropy estimates s for the optimization Problem (24) and (25) will be obtained as a result of solving the system of equations where i = 1, n, j = 1, o. The starting point for calculating the variable entropy estimate E s of the probability density functions of the parameters p and q of QνSM, both in the Interpretation (20) and (21), and in the Interpretation (24) and (25), is the calculation of the Lagrange multipliers ψ as a result of solving the systems of equations represented by Expressions (23) and (26), respectively. This process can be arranged, for example, according to the multiplicative algorithm [36]: where ϕ i = exp −ψ j are exponential Lagrange multipliers, ϕ 0 j > 0, j = 1, o.

Experiments
Let us demonstrate the functionality of the mathematical apparatus proposed in Section 2 using the example of calculating the variable entropy estimate of the probability density functions of the characteristic parameters of the linear stochastic small data estimation model with the dimension of the entities "input" × "output" of [5] × [2]. The matrix of the measurements of the "input" entity looks like this: The vector of the measurements of the "output" entity, taking into account variability, looks like this: y = 21.091 32.814 .
Quasi-moments of the first order are described by the expressions: 10], i = 1, 5; υ 1 = −1 + 2q 1 , υ 2 = −2 + 4q 2 , υ 1 ∈ Entropy 2022, 24, x FOR PEER REVIEW (1) Instances of the class of parameterized stochastic models for evaluating small data; (2) Methods of estimating the probability density function of their parame resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vectors of initial para (4) A technique for statistical processing of such an ensemble using the Mo method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric signals or phenomena cal, medical, economic, biological and other sources of origin is the functional pu evaluation theory as a branch of mathematical statistics. To solve the problem o tion, parametric and non-parametric approaches are used. In recent decades, the noticeably dominated the former, which has become possible thanks to the "react gress in the field of machine learning and artificial intelligence. At the same time, of researchers' interest is shifting from the study of the processes represented by to that of those processes about which the amount of data small, and the data i tains errors. Such a preamble encourages the perception of the parameters of t data evaluation model as stochastic quantities. Accordingly, we will call such a stochastic model for small data evaluation. The characteristics of such a mode probability density functions of the stochastic parameters. The primary task in id a stochastic estimation model for specific small data is to estimate the parameter probability density functions. If this step is passed, then the identified stochasti tion model can be taken as a basis for forming moment models of small data, ge an ensemble of random vectors of the initial parameters and carrying out the s processing of such an ensemble using the Monte Carlo method [6][7][8] to bring desired numerical characteristics. The formalization of the way to solve the prima lem formulated above has scientific potential and applied value. (1) Instances of the class of parameterized stochastic mo small data; (2) Methods of estimating the probability density functi resented by normalized or interval probabilities; (3) Approaches to generating an ensemble of random vec (4) A technique for statistical processing of such an ensem method to bring it to the desired numerical characteristics.

Statement of the Research
Evaluation based on data that represent parametric sign cal, medical, economic, biological and other sources of origin evaluation theory as a branch of mathematical statistics. To s tion, parametric and non-parametric approaches are used. In r noticeably dominated the former, which has become possible t gress in the field of machine learning and artificial intelligence of researchers' interest is shifting from the study of the proces to that of those processes about which the amount of data sm tains errors. Such a preamble encourages the perception of t data evaluation model as stochastic quantities. Accordingly, stochastic model for small data evaluation. The characteristi probability density functions of the stochastic parameters. The a stochastic estimation model for specific small data is to estim probability density functions. If this step is passed, then the i tion model can be taken as a basis for forming moment mode an ensemble of random vectors of the initial parameters and processing of such an ensemble using the Monte Carlo met desired numerical characteristics. The formalization of the wa lem formulated above has scientific potential and applied val Let there be a stochastic parameterized research object r measurements, in which the matrix of values of the input para sion   The deviations from the values specified in the vector α 0 caused by the variability of the measurements are characterized by an error ε = α 0 − α / α 0 + α .
Summarizing the given initial information in the format of Expression (19), we obtain: and p 0 C , q 0 E imply uneven distributions of the characteristic parameters and influences, while the latter represents the variant combined according to the a priori probabilities of the corresponding entities.
We obtain optimization problem Statements (20) and (24) for the initial parameters presented above.
The formulation of the optimization Problem (20) and (21) for the above-mentioned initial data has the form: The formulation of the optimization Problem (24) and (25) for the above-mentioned initial data has the form: 0.747p 1 + 0.873p 2 + 1.366p 3 + 0.834p 4 + 0.622p 5 + 0.249q 1 = 1, 1.065p 1 + 0.982p 2 + 0.67p 3 + 0.721p 4 + 0.449p 5 + 0.312q 1 = 1; Such optimization problems can be solved by methods of non-linear mathematical programming [36]. In particular, for the above optimization problems, the extremum point is analytically identified as p * i = 0.36p 0 i , q * j = 0.36q 0 j , i = 1, 5, j = 1, 2. So, for our example, the entropy H( s) reaches its maximum at the point (p * , q * ), where p * = f i, p 0 i , q * = f j, q 0 j , i = 1, 5, j = 1, 2. Let us examine these dependencies, taking into account that we previously defined schemes for a priori values: For clarity, we present the dependences p * = f i, p 0 {A,B,C} and q * = f j, q 0 {D,E} in the form of diagrams (Figures 1 and 2, respectively).  , D E q f j q * = in the form of diagrams ( Figure 1 and Figure 2, respectively).  More detailed information on the values of the characteristic parameters of the investigated linear stochastic model of the small data evaluation presented in Section 3 can be seen in Figures 3 and 4 Comparing the symmetrical values visualized in Figures 3 and 4, it can be concluded that the parameter estimates calculated for interval probabilities (i.e., for Q vSM ) are characterized by a larger value of the conditional maximum entropy than that inhered for GvSM (i.e., for the normalized probabilities). The theoretical justification of this empirical fact is presented in Section 4.
Information about the state of the linear stochastic models, summarized by Expressions (27) and (28), is supplemented by such calculated data as:  These figures visualize the values at the extremum point (p * , q * ) of E (1) -estimates of such characteristic parameters as p ∪ i , i = 1, 5; q ∪ j , j = 1, 2, and H * (s) (calculated by Expression (20) adapted to form (27)) and H * ( s) (calculated by Expression (24) adapted to form (28)). At the same time, the schemes of the initial values of the vectors p 0 = p 0 i , i = 1, 5, and q 0 = q 0 j , j = 1, 2, are taken into account. Comparing the symmetrical values visualized in Figures 3 and 4, it can be concluded that the parameter estimates calculated for interval probabilities (i.e., for QνSM) are characterized by a larger value of the conditional maximum entropy than that inhered for GνSM (i.e., for the normalized probabilities). The theoretical justification of this empirical fact is presented in Section 4.
Information about the state of the linear stochastic models, summarized by Expressions (27) and (28), is supplemented by such calculated data as: (1) the value at the point of extremum (p * , q * ) of the quasi-moments of the characteristic parameters of GνSM and QνSM (α * i , i = 1, 5), (2) estimates of the variability of the above-mentioned parameters caused by interferences (υ * j , j = 1, 2), (3) the errors ε and ε, which characterize the deviation of the measured parameters α, υ from the reference α 0 , υ 0 for GνSM and QνSM, respectively.
These data are visualized in Figures 5 and 6.    From the information shown in Figures 5 and 6 (in addition to the information presented in Figures 3 and 4), it can be concluded that the reference parameters and a priori probabilities are correlated. That is, the closer the values in the scheme of a priori probabilities are to the values of the reference parameters, the smaller the value of the error ε . This interpretation, in particular, explains the superiority of the scheme ( )

Discussion
Let us begin the analysis of the results presented in Section 3 of the applied use of the mathematical apparatus proposed in Section 2 with the fact that the estimates of the parameters , p q obtained as a result of solving optimization Problems (27) (derived from Problem (20), (21) and (28)) (derived from Problem (24) and (25)), turn out to be different in terms of the value of the generalized entropy (Expressions (20) and (24), respectively). We will explain this fact on the theoretical basis of the models presented in Section 2.
To simplify the formulations, we will introduce several renovations. Let us redefine . Accordingly, 1 е * will be the optimal estimate of the parameters From the information shown in Figures 5 and 6 (in addition to the information presented in Figures 3 and 4), it can be concluded that the reference parameters and a priori probabilities are correlated. That is, the closer the values in the scheme of a priori probabilities are to the values of the reference parameters, the smaller the value of the error ε. This interpretation, in particular, explains the superiority of the scheme (p B , q D ) over the scheme (p C , q D ), because ε BD < ε CD .

Discussion
Let us begin the analysis of the results presented in Section 3 of the applied use of the mathematical apparatus proposed in Section 2 with the fact that the estimates of the parameters p, q obtained as a result of solving optimization Problems (27) (derived from Problem (20), (21) and (28)) (derived from Problem (24) and (25)), turn out to be different in terms of the value of the generalized entropy (Expressions (20) and (24), respectively). We will explain this fact on the theoretical basis of the models presented in Section 2.
To simplify the formulations, we will introduce several renovations. Let us redefine entropy H(s) as H(e) = H(s), where e = p, q . Accordingly, e * 1 will be the optimal estimate of the parameters p * , q * represented by normalized probabilities (H(s) variant) and e * 2 will be the optimal estimate of the parameters p * , q * represented by interval probabilities (H( s) variant). Let us denoteê = argmaxH(e) and define the sets Summarizing what has been entered, we formulate the following: ifê ∈ R (n+o) + \E then H e * 1 < H(e * 2 ). The equality H e * 1 = H(e * 2 ) holds when e * 1 =ê. Let us explain our conclusions. The analysis of the function described by Expression (20) shows that it is a concave function with a single maximum at the pointê. The value of entropy H(e) depends on the distance of a point e from the extreme pointê. In this context, we denote as ∆ ê, e * 1 the distance between the extreme pointê and the point e * 1 , the coordinates of which we obtain as a result of solving optimization Problem (20) and (21). Accordingly, the parameter ∆(ê, e * 2 ) characterizes the distance between the extreme pointê and the point e * 2 , the coordinates of which we obtain as a result of solving optimization Problem (24) and (25). Since Function (20) is strictly concave, based on the Relation (29) we can conclude that ∆ ê, e * 1 < ∆(ê, e * 2 ). The equality ∆ ê, e * 1 = ∆(ê, e * 2 ) holds only when e * 1 =ê. The presented theoretical explanations explain the discrepancy between those presented in Figures 3 and 4  QνSM) are characterized by a larger value of the conditional maximum entropy estimate than that characteristic of the normalized probabilities of GνSM. Thus, the mathematical apparatus presented in Section 2 was empirically confirmed in Section 3.

empirical values of E
In addition, the results of the experiments presented in Section 3 confirmed the conclusion generalized by Expressions (17) and (18) that, for a linear stochastic model of variable small data estimation, entropy estimates E (1) s are always exponential functions. The results of measuring the "input" and "output" entities of the investigated process determine the form, and not the type, of the E (1) s -functions of the corresponding linear stochastic model of small data estimation.
The results shown in Figures 3 and 4 showed that a priori information about the initial values of the vectors p 0 = p 0 i , i = 1, 5, and q 0 = q 0 j , j = 1, 2, summarized in the corresponding named sets of p 0 {A,B,C} , q 0 {D,E} , has a significant effect on the E (1) {s, s} p ∪ i , q ∪ j , H({s, s}) estimates. In this context, the fact that the author's mathematical apparatus allows the calculation of the quasi-momentums of the characteristic parameters α * i , i = 1, 5, of both the GνSM and QνSM, as well as the taking into account of their variability υ * j ,j = 1, 2, caused by the measurement errors, is very relevant. From those visualized in Figures 5 and 6 of the data, it can be seen that the ε deviations from the values indicated in the vector α 0 caused by the variability of the measurements are most pronounced for the schemes p 0 C , q 0 D and p 0 C , q 0 E . These schemes are characterized by the fact that the essential parameters of the models are characterized by an uneven distribution (see Figure 1, "C"), and the influence parameters are characterized by both uneven (see Figure 2, "D") and uniform distributions (see Figure 2, "E"). For both schemes, we obtained: ε CD = 0.30, ε CE = 0.36; ε CD = 0.33, ε CE = 0.36. Therefore, for the considered example, the unevenness of the distribution of parameters p i , i = 1, 5 provided a significant contribution to the high value of errors ε. Reliable a priori information turned out to be very important in the entropy estimation of variable small data.

Conclusions
The article analytically summarizes the idea of applying the Shannon entropy maximization principle to sets that represent the results of observations of the "input" and "output" entities of the stochastic model for evaluating variable small data. To formalize this idea, a sequential transition from the likelihood function to the likelihood functional and the Shannon entropy functional is analytically described. Shannon's entropy characterizes the uncertainty caused not only by the probabilistic nature of the parameters of the stochastic data evaluation model but also by influences that distort the results of measurements of the values of these parameters. Accordingly, based on the Shannon entropy, it is possible to determine the best estimates of the values of these parameters for maximally uncertain (per entropy unit) influences that cause measurement variability. This postulate is organically transferred to the statement that the estimates of the probability distribution density of the parameters of the stochastic model of small data obtained as a result of Shannon entropy maximization will also take into account the fact of the variability of the process of their measurements. In the article, this principle is developed into the information technology of the parametric and non-parametric evaluation on the basis of Shannon entropy of small data measured under the influence of interferences.
The article also examines the structural properties of stochastic models for variable data evaluation, the parameters of which were represented by normalized or interval probabilities. At the same time, the inherent non-linearity of these models and errors in measuring the values of the "output" entity was taken into account.
The functionality and adequacy of the created mathematical apparatus are proven based on the empirical results obtained during the investigation of the linear stochastic model of evaluating specific variable small data.
The authors acknowledge that the research presented in the article is formulated in academic form. This circumstance complicates the applied use of the obtained results. At the same time, the developed methodological approach can be useful in various important applications. In particular, it concerns the assessment of software reliability, when the sample of data is usually not large due to the difficulties of reliably assessing them during the testing and operation of the system. In this case, the lack of testing data or information about failures during pilot software operation can be compensated for by analyzing the assumptions that are specific to the software and selecting appropriate models using assumption matrices [37]. Thus, studies that combine the analysis of small data and expert methods are interesting.
Another important application is in safety critical systems, which, due to multi-level reserving, have as a rule a low failure rate and small data about them. On the other hand, it is extremely important for such systems to have accurate or at least interval estimates of indicators with an acceptable range. For that, the described method could be combined with the traditional methods of reliability analysis and risk oriented assessing of safety indicators using formal and semi-formal methods [38].
In this regard, further research is proposed to formalize the obtained information technology on a UML basis. This will allow the future work to reach the stage of implementing the profile framework. In addition, it would be very interesting and useful from a practical point of view to combine Big and Small Data analysis to create universal or adaptable framework focusing on the assessment of data quality and their selection according to the quality indicator. Data Availability Statement: Most data is contained within the article. All the data available on request due to restrictions, e.g., privacy or ethical.