Memory-Based Evolutionary Algorithms for Nonlinear and Stochastic Programming Problems

: In this paper, we target the problems of ﬁnding a global minimum of nonlinear and stochastic programming problems. To solve this type of problem, we propose new approaches based on combining direct search methods with Evolution Strategies (ESs) and Scatter Search (SS) metaheuristics approaches. First, we suggest new designs of ESs and SS with a memory-based element called Gene Matrix (GM) to deal with those type of problems. These methods are called Directed Evolution Strategies (DES) and Directed Scatter Search (DSS), respectively, and they are able to search for a global minima. Moreover, a faster convergence can be achieved by accelerating the evolutionary search process using GM, and in the ﬁnal stage we apply the Nelder-Mead algorithm to ﬁnd the global minimum from the solutions found so far. Then, the variable-sample method is invoked in the DES and DSS to compose new stochastic programming techniques. Extensive numerical experiments have been applied on some well-known functions to test the performance of the proposed methods.


Introduction
In the real world, there are many challenging applications which naturally require involving optimization techniques to have the best configurations, and find the optimal resources [1]. In particular, these complex application problems are computationally expensive in a sense of resources and time and most of them contain noise. Furthermore, the objective functions associated with most of these problems are not differentiable and have a large scale of parameters. Therefore, trials to compute the derivatives using finite differences methods mostly fail.
Typically, such real complex systems are usually modeled using a series of simulation procedures to evaluate their output responses. Also, they are integrated with an optimization algorithm to find their optimal parameters. The integration of the stochastic optimization methods into simulation software packages to meet their design requirements is crucial. Thus, these techniques are called Simulation-Based Optimization (SBO) [2,3]. SBO consists of three components, design tools, precise soft simulators, and optimization techniques. There are many engineering fields using the SBO, such as aerospace [4][5][6], automotive [7,8], transportation [9], and naval [10]. One of the most difficult issue of such systems is that the objective function usually contain noise and the user should carry out several simulations to obtain an acceptable approximate value of the objective function at a single point [11].
Therefore, this research focuses on optimizing objective functions with high noise which is known as stochastic programming [12].
Usually, the term "Simulation-Based Optimization" is more common than "stochastic optimization" [11,13]. Highly non-linear and/or high dimensional real-world optimization problems require the SBO, since these problems involve uncertainties in the form of randomness. This problem is known as stochastic programming problem where the estimation of the uncertainty is carried out as a stochastic probability function. The considered problem is represented mathematically as follows: where X ⊆ R n is the search space and f is a function with decision variables x ∈ X, and F is the probability function of a random variable ω. The configuration of the simulation parameters to enhance the operation is the main challenge. Typically, the simulation process is complex, since the evaluation of its objective function is computationally expensive, the results are noisy, and the objective function does not have an exact gradient. Usually, these problems have multiple local optima, which prevents classical nonlinear programming techniques from solving such problems successfully all the time. Therefore, global search methods are the best candidates to solve these problems which can defined in the following nonlinear programming problem: where X ⊆ R n is the search space and f is a nonlinear function with decision variables x ∈ X. Therefore, this research aims to design new methods to deal with Problems (1) and (2).
Recently the metaheuristics techniques, such as Evolution Strategies (ESs) and Scatter Search (SS), can solve Problem (2) by applying intelligent and learned procedures [14][15][16][17]. Metaheuristics are powerful tools since they are robust and can fight to successfully find accurate solutions of a wide range of problem areas [14,17]. However, applying the metaheuristics for complex problems converges slowly since exploring the global search space is a stochastic process that needs a high computational cost. Additionally, they usually could not remark promising search directions utilizing local information. On the other hand, the convergence of local search methods is faster since they create mathematical or logical movements based on the local information to determine promising search directions. However, they can be easily entrapped in local minima.
Typically, numerical global optimization methods are classified into deterministic, stochastic (heuristic), and hybrid techniques [18,19]. Generally, deterministic methods solve problems that are formulated based on their analytical properties, and some theoretical assumptions. Also, when there is available prior information about the problem under investigation, Branch-and-Bound algorithms in the deterministic global optimization are the most capable methods for searching the global minima [20]. The deterministic methods can find a global minimum with a pre-defined tolerance in case of having exact computation over long run time. In this type of optimization, the range of all variables defines the search space. Besides, these methods are characterized by having a theoretical assurance of finding the global solution. This means the value of the objective function when applying a local optimum differs just by small value from the global one. Old deterministic algorithms suffer from some limitations such as loose bounds, and they cannot deal with practical problems with more than ten variables. However, some new deterministic methods have been improved to obtain good results, such as the Hit-and-Run method. Also, it is modified to deal with high dimensional problems such as the Molecular Distance Geometry Problem (MDGP) instances [21], but it consumes high computational time. These limitations of the deterministic methods have motivated the transition to the stochastic methods.
On the other hand, some of stochastic methods are characterized by a probabilistic convergence guarantee, which usually means the probability of its global solution is one within infinite time.
Furthermore, the stochastic methods search for the optimal solution of problems that have black-box formulations, awful behavior, and noisy evaluation functions [22]. The stochastic methods are the best candidate to search for optimal solutions when objective functions suffer from a deficiency of prior suppositions. These methods work by sampling the search space randomly. There are many techniques classified as stochastic methods such as Genetic Algorithms (GA), Simulated Annealing (SA), Tabu Search (TS), Particle Swarm Optimization (PSO), and others [23]. Typically, the stochastic search is a multistage method which consists of global and local stages. The performance of these two stages significantly affects the efficiency of the stochastic methods, which also subject to the quality of sampled points.
Each technique in the above two categories has certain advantages and drawbacks. Researchers have hybridized these techniques to enhance their performance, and make it faster and more robust [24,25]. An effective introduction to hybrid metaheuristic methods is explained in [26]. An example for these hybrid algorithm is proposed in [27], where an approximation algorithm and linear relaxations can be integrated with an ant colony-like algorithm to get a very effective and fast algorithm for multi-period capacity and routing path planning in network design. Another example of work which combines genetic algorithms with individual Quasi-Newton learning for global optimizations is shown in [28].
The considered stochastic problem is conquered by several optimizations and search techniques, such as the Variable-Sample methods (VS) [29]. These methods are classified as Monte-Carlo methods that convert unconstrained stochastic optimization problems to deterministic problems. Explicitly, an estimation of the sample performance measure containing the random variable is carried out using several sample points. The VS method is invoked by a random search algorithm called Sampling Pure Random Search (SPRS) where an average approximation method with a variable-sample size replaces the objective function in each iteration [29].
In this paper, we find the global solution of the stochastic programming problem by developing new hybrid methods combining metaheuristics techniques, such as Evolution Strategies, Scatter Search, direct search, and Variable-Sample (VS) methods. Uncertainties formed as randomness and high dimensional are the main difficulties associated with such stochastic programming problems. Two main steps are representing the proposed methods as follows: • First, new versions of ES and SS are proposed to enhance their performance in finding the solutions of the global optimization problem defined as Problem (2): -A new version of ESs called Direct Evolution Strategy (DES) [30] is applied. The DES method is composed by integrating the ESs with a new philosophy for applying the evolution process. Specifically, reproducing offspring in the DES method has two directions. The first one is to modify the mutation operation of ESs to reproduce offspring in a closed environment of their parents and use the parents' experience to direct the offspring towards promising directions. The other direction of reproducing offspring is achieved by modifying the recombination operation to reproduce offspring in a promising area slightly away from their parents' environment. This offspring has exploration duties that keep the diversity of its generation and the successive ones. This process of exploration is known in the literature as "Dispersal", which means the movement of organisms away from their point of origin [31]. More elements have been added in the DES method to achieve more diversification process and to overcome the drawback of ESs termination.

-
A new SS method is proposed by extending the regular SS method built based on Laguna and Martı [32] and Duarte et al. [33] and integrating it with the Gene Matrix (GM). The proposed method is called Directed Scatter Search (DSS). Particularly, a hybrid technique is applied to update the reference set of the SS, where the possible values of each variable are divided into sub-ranges stored in the GM [34]. The GM enhances the exploration process after applying a mutagenesis operator [34] by adding new diverse solutions to the search space. Mutagenesis operator works on accelerating the exploration process by altering some survival solutions.
• Second, the proposed metaheuristics techniques are combined with the Variable-Sample method to solve the simulation-based optimization problem defined as Problem (1) The details of the proposed non-linear programming techniques and its main components are explained in Sections 2 and 3. The variable sample methods are highlighted in Section 4. The proposed simulation-based optimization methods are introduced in Sections 5 and 6. Section 7 explains the main numerical experiments and how the parameters of the proposed methods are selected and tuned. Finally, the conclusion is presented in Section 8.

Directed Evolution Strategies
Hereafter, we explain how the ESs method is modified to produce the Directed Evolution Strategies (DES) [30]. Then, the DES method is combined with a variable-sample technique to compose a simulation-based optimization method. Next, we mention some remarks about the ESs which considered as initiatives for building the DES.
In the (µ/ρ,λ)-ES, all parents are discarded during the selection process, whether they are good or bad since a strict Darwinian selection is the approved selection mechanism. This selection mechanism is not the optimal one since we lose some good solutions when discarding all of them. On the other hand, the (µ/ρ+λ)-ES applies different selection mechanism where the next generation only takes µ which are the best individuals come from the parents and offspring. This type of selection may lead to immature convergence without guarantee diversity. Therefore, a new operation called "mutagenesis" is proposed. To maintain the diversity, the proposed mutagenesis alters some survival individuals to new unexplored areas. Besides, this new operation maintains the elitism works since it works within the "+" selection mechanism.
The ESs have an unbiased chance of reproducing children, which makes them are not like genetic algorithms (GAs). During recombination and mutation operations, there is an equal chance for all parents to reproduce children. Beside, the fittest works only during the selection operation. Moreover, the recombination operation in the standard ESs has a variable number ρ of mated parents, i.e., it supports the multi-parents recombination.
Since intelligent exploitation and exploration processes should be implemented in a good global search method, and since ESs have their promising mutation operation as an exploitation process, it is advisable to utilize a recombination mechanism which can serve as an intelligent exploration process. According to this concept, the recombination operation in the DES method has been constructed to diverse the search process towards new unexplored areas.
Another drawback in the ESs and EAs is termination criteria since there are no automatic termination criteria. In fact, in an early stage of the search process, the EAs may reach an optimal or near-optimal solution; however they could not terminate because they were not learned enough. Therefore, through successive generations, our proposed DES has termination measures to check the progress of the exploration process. Afterwards, to converge faster an exploitation process refines the best candidates obtained so far.
The remarks mentioned above have been taken into consideration while designing the proposed DES. Next explained components are integrated into the standard ESs algorithm to create the new DES method.

•
Gene Matrix and Termination. During the search process, GM [34] is applied to ensure the exploration of the search space. In the GM, each variable is divided to sub-ranges based on its possible values. In the real-coding representation, there are n variables or genes in each individual x in the search space. The GM checks the diversity by dividing the range of each gene to m sub-ranges. Initially, a zero matrix of size the n × m represents the GM where each element in each row denotes a sub-range of the equivalent gene. During the search process, GM's values are altered from zeros to ones. An example of GM in a two-dimension search space is shown in Figure 1. We achieve an advanced exploration process when the GM has all ones. But the search will widely explore the recently chosen sub-ranges with the help of recombination operation. Therefore, the advantages of the GM are the practical termination criteria and providing the search with diverse solutions as explained hereafter. • Mutation. The standard ESs mutation process is applied for generating mutated children [35,36]. Thus, a mutated child (x,σ) is attained from a parent (x,σ), where the i-th component of the mutated child (x,σ) is given asσ where typicaly the values of τ ∝ 1/ √ 2n and τ ∝ 1/ 2 √ n are selected as one.

•
Mutagenesis. For accelerating the exploration process, the N w worst individuals selected for the next generation are reformed by a more artificial mutation operation called "mutagenesis". Explicitly, a random zero-position (i, j) in GM is selected. This will explore the j-th partition of the variable x i range which has not been visited yet. Then, one of the selected individuals for mutagenesis will be changed by a random value for x i within this partition. The formal mutagenesis procedure is explained as follows in Algorithm 1: if GM is not full then 3: Select a zero-component (i, j) in GM randomly. 4: where l i , u i are the lower and upper bound of the variable x i , respectively, and r is a random number from (0, 1).

5:
Update GM ← (GM ij = 1). Recombination. The recombination operation is not applied for all parents, but with probability p r ∈ [0, 1), since it has an exploration tendency. Initially, ρ parents p 1 , . . . , p ρ will be randomly chosen from the current generation. The number of the partitions of each recombined child x equals ρ (ρ ≤ n) where partitions are taken from parents. Figure 2 displays an instant of 4 parents p 1 , p 2 , p 3 , and p 4 partitioned at the same positions into 4 partitions. Next, a recombined child is created by inheriting its four partitions from p 3 , p 1 , p 4 and p 2 , respectively. The DES recombination operation is described in Algorithm 2. if ρ ≤ n then 3: Divide each parent into ρ parts at the same positions, i.e., Reorder the index set {1, 2, . . . , ρ} randomly to be {o 1 , o 2 , . . . , o ρ }.

5:
Compose It is important to mention that the DES recombination procedure, defined to support the DES exploration process, is built as a standard type of the "single point crossover" and "two point crossover" [37]. Specifically in the high dimensional problems, the GM overcomes the complexity of such problem by not saving the information of the diverse sub-ranges of separate variables. Hence, as shown in the example in Figure 3a, it is possible to have misguided termination of the exploration process. However, this drawback, shown in Figure 3b, can be easily overcome by calling such recombination operation as in Procedure 2. • Selection. The same standard ESs "+" selection operation is applied for the DES to choose from P g some survival individuals and insert it into its next P g+1 . Though, to ensure diversity, the mutagenesis operator alters some of these individuals. • Intensification Process. After having all children, DES applies a local search starting from two characteristic children to improve themselves. These two children are the one with the best computed solution till now, and the other child is chosen after computing the objective function of each parent and his child and select the child with the highest difference in the current generation. Therefore, these characteristic are called children "the best child" and "the most promising child" respectively. The DES behaves like a "Memetic Algorithm" [38] after applying this intensification process which accelerates the convergence [39,40]. For more intensification to enhance the best solutions found so far, we apply derivative-free local search method as an efficient local search method. At the last stage, this intensification process can enhance convergence of EAs and prevent the EAs roaming around the global solution. Figure 4 shows the main steps of the DES method while its algorithm is explained as follows (Algorithm 3): Algorithm 3 DES Algorithm 1: Initial Population. Set the population parameters µ and λ, and calculate ν := λ/µ. Generate an . . , µ}. 2: Initial Parameters. Select the recombination probability p r ∈ [0, 1], and set values of ρ, N w , m and N elite . Set the gene matrix GM to be an empty matrix of size n × m. 3: Initialize the generation counter g = 0. 4: while the termination conditions are not met do 5: for j ← 1, µ do 6: Recombination. Choose a random number χ ∈ [0, 1]. 7: if χ > p r then 8: Select a parent (x j ,σ j ) from P g . 9: else 10: Select ρ parents p 1 , . . . , p ρ from P g . 11: Compute the recombined child (x j ,σ j ) using Recombination Procedure (Algorithm 2). 12: end if 13: Mutation. Use Equations (3) and (4) to compute a mutated children (x j,k ,σ j,k ), k = 1, . . . , ν. 14: Fitness. Evaluate the fitness functionF j,k = F(x j,k ), k = 1, . . . , ν. 15: end for 16: . . , ν}, to contain all children. 17: Selection. Set the next generation P g+1 to contain the best µ individuals from P g ∪ C g . 18: Mutagenesis. Use Mutagenesis Procedure (Algorithm 1) to alter the N w worst solutions in P g+1 . 19: Update the gene matrix GM.

Scatter Search with Gene Matrix
Hereafter, we explain the integration between the Gene Matrix (GM) explained in section 2, and the standard Scatter Search (SS) to compose the DSS method. The proposed DSS method involves the same five steps of the SS as follows:

2.
Improvement: Applying local search to improve these trial solutions.

3.
Reference Set Update: Updating the reference set containing the b best solutions found. the Solutions are classified based on the quality or the diversity.

4.
Solution Subset Generation: Dividing the reference set solutions into subsets to create combined solutions.

5.
Solution Combination: Combining solutions in the obtained subsets to create new trial solutions.
In step 3, the Reference Set (RefSet) in the proposed DSS has three parts. These parts are updated to keep the elitism, the diversity, while the GM and mutagenesis concepts [34] explained in Section 2 are applied to update the last part. The indicator of attaining an advanced exploration process is having GM full by ones. Therefore, GM assists the diversification process by supplying the search with diverse solutions using the mutagenesis operator (Algorithm 1), and it is used as a practical termination criteria.

Updating the Reference Set
In the basic procedure of SS based on the evaluation of the objective function, a solution with high value replaces the one with the worst value for updating the reference set.
Also, another part in the RefSet is kept preserving the solution diversity [32]. The standard update procedure in the SS is a reformed technique called the Update with Gene Matrix (UpGM) in the proposed DSS method. This new update technique increases the diversity, the exploration, and exploitation processes. All solutions from the combination ones already in the RefSet make up the new offspring Pool, i.e., Pool = Offspring ∪ RefSet. Algorithm 4 represents the formal steps for the process of updating the RefSet.

Diversity Control
In the standard SS to avoid trapping in local minima, the reference set has to be checked to circumvent duplicate solutions. However, applying this method cannot guarantee the diversity of the b 1 superior solutions while initializing the RefSet. Instead, a strict diversity check using the max-min criterion is applied to ensure the diversity of the b 2 solutions. In the initial RefSet, the chosen b 1 solutions can be exposed to a minimum diversity test as follows: The best solutions in the initial population P are chosen to be x 1 in the RefSet. Afterwards, x 1 is removed from P, and the next high-quality solution x in P is selected and added to the reference set only if it is far from x 1 . Therefore, the minimum distance between the current solutions in RefSet and the selected solution x controls the process of adding the next best solution in P at each step. This distance has to be at least as large as the threshold value to add this best solution.

DSS Algorithm
The main steps of the DSS method are shown in Figure 5 while it is formalized as described in the following Algorithm 5.

Algorithm 5 DSS Algorithm
Generate a solution x using the Diversification Generation Method. 4: if (x ∈ P || is not close to any member in P) then, 5: Use the Subset Generation Method to create solution subsets of size two. 12: Generate the offspring using the Combination Method. 13: Refine the offspring using the Improvement Method and add them to Pool Set. 14: Use the Update with GM Procedure (Algorithm 4) to update the RefSet. 15: end while 16: Refine the N elite best solutions using a local search method. In this method, we apply "fminsearch.m" which is a MATLAB function implementing the Nelder-Mead method. This function improves the intensification of the proposed method (Step 16 of Algorithm 5). In the experimental work, the maximum evaluations number of the objective function is used to terminate the search process.

Variable Sample Method
The Variable-Sample (VS) method is classified as a class of Monte-Carlo methods for dealing with unconstrained stochastic optimization problems. The stochastic optimization problem can be converted using the VS to a deterministic problem, where an estimation of the sample performance measure containing the random variable is carried out using several sample points. The VS method is usually called by the Sampling Pure Random Search (SPRS), which is a random search algorithm [29]. In the SPSR, a sample of average approximation which has a VS size scheme replaces the objective function. The SPSR formal algorithm is described in Algorithm 6 [29]. for i = 1 to N k do 6: Compute F(x, ω k i ).

Directed Evolution Strategies for Simulation-based Optimization
The DESSP method is a modified version of the DES method to handle the simulation-based global optimization problem. The modified version (DESSP) evaluates the fitness function values by employing the VS method (Algorithm 6). To illustrate how this can be done, Figures 4 and 6 show the main structures of the DES and DESSP methods, respectively. One can note that the main added components in the DESSP method are the fitness function evaluation using variable sample technique and updating the size of the sample points. The first added component about function evaluations is performed as in Step 4 of Algorithm 6 with sample size N. The parameter is configured through the following procedure shown as Algorithm 7.

Scatter Search for Simulation-Based Optimization
The DSS adapted to handle the same optimization problem using the following VS method.

Sampling and Fitness Evaluation
The VS technique is applied, as in Step 4 of Algorithm 6, to evaluate the objective (fitness) function, since it contains a random variable. The sample size parameter and its update are the main challenges with the fitness evaluation process. In the DSSSP method, the GM status can guide the configuration of the sample size parameter as described Sample Size Update Procedure (Algorithm 7).

DSSSP Algorithm
The DSS is modified to a new version called DSSSP, which tries to find the global optima in stochastic programming problems. Integrating the VS method into DSS is the main modification, Algorithm 6, to compute the fitness function values. The main steps in the DSSSP method are shown in Figure 7 and explained in Algorithm 8.

Algorithm 8 DSSSP Algorithm
1: P ← ∅ 2: repeat 3: Generate a solution x using the Diversification Generation Method. 4: if (x ∈ P || is not close to any member in P) then, 5: 6: end if 7: until the size of P is equal to PSize. 8: Use variable sampling with size N to compute the fitness values for all members in P. 9: Apply Update with GM Procedure (Algorithm 4) to generate an initial RefSet. 10: while the stopping criteria are not met do 11: Use the Subset Generation Method to create solution subsets of size two. 12: Generate the offspring using the Combination Method. 13: Refine the offspring using the Improvement Method and add them to Pool Set. 14: Use the Update with GM Procedure (Algorithm 4) to update the RefSet.

15:
Update the variable sample size N using Sample Size Update Procedure (Algorithm 7). 16: end while 17: Refine the N elite best solutions using a local search method.

Numerical Experiments
The proposed methods have been implemented using MATLAB, and some benchmark test functions have been applied to evaluate their performance. In this section, the configuration and the main numerical results prove the efficiency of these methods are shown.

Test Functions
For the nonlinear programming problem, we have tested the proposed algorithms using two sets of well-known test functions Set A, and Set B. Table 1 contains Set A [41][42][43], while Set B is listed in Table 2. These two sets of functions have enough diverse characteristics to test different difficulties that arise in global optimization problems. For the stochastic programming problem, seven test problems [44] are formulated from four documented test functions shown in Table 3, while their mathematical definitions are given in Appendix A [34].

Parameter Setting and Tuning
To have a complete description of the proposed algorithms, all setting parameters of the DES, DSS, DESSP, and DSSSP methods are discussed in this section. Tables 4 and 5 summarize these parameters, their definitions, and best values. The values stated in the literature are used to set some parameters. A tuning process is carried out to obtain standard settings, while the proper values of the other parameters have been set by running some preliminary numerical experiments. These standard settings do not as much as possible, depend on the problem. Different termination criteria are used in comparisons to stop the proposed methods. First, the automatic termination criterion based on the GM is discussed. We define γ ratio to be equal to the number of ones in GM divided by the total number of entities of GM. If γ is equal to 1 or closed to 1, then the method can learn that a wide exploration process has been achieved. Therefore, the algorithm can be stopped after satisfying the condition γ ≥ γ max , where γ max is equal to 1 or closed to 1. Figure 8 shows the results of applying the GM termination on the DESSP and DSSSP methods. In this figure, we allowed the algorithm to continue after reaching the termination point in order to notice any possible improvement after that. It is clear from the figure that there is almost no significant improvement after the GM termination point. The values of γ max are set equal to 0.8 and 1 for the DESSP and DSSSP methods, respectively. The difference in γ max values is a result because the DESSP method uses the GM to update its iterate solution sets more than that in the DSSSP method. Therefore, the DSSSP method reaches a full GM faster than the DSSSP method. In comparisons, we terminated the proposed methods according to the same termination criterion used in the compared methods in order to get a fair comparison [45,46]. Table 4. DES and DESSP parameters.

Parameter Definition
Best Value µ The population size 50 λ The offspring size 10µ ρ No. of mated parents min(n, 5)) p r The recombination probability 0.25 N min The initial value for the sample size 100 N max The maximum value for the sample size No. of function evaluations

Psize
The initial population size 100 b The size of the RefSet 10 b 1 No. of elite solutions in RefSet The initial value for the sample size 100 N max The maximum value for the sample size 5000 N elite No. of best solutions used in in the intensification step 1 N w No. of worst solutions updated by the mutagenesis process 3 m No. of partitions in GM 50

Numerical Results on Functions without Noise
This section contains the explanation of the experiments' implementation that evaluates the proposed algorithms and show their results. Initially, two experiments are carried out to test the DSS algorithm. The first experiment measures the effect of applying Gene Matrix to the standard SS. The experimental results are reported after applying the SS algorithm with and without using GM on the Set A test functions stated in Table 1. On the other hand, the second one compares the performance of the DSS and the standard SS method as well as other metaheuristics methods.
The optimality gap of the above experiments is defined as: where x and and x * are a heuristic solution, and the optimal solution, respectively. We consider x as a best solution if the average GAP ≤ 0.001. The results of the first experiment after updating RefSet with (DSS) and without using GM (standard SS) are reported in Table 6. These results are calculated after running each test function; ten independent runs with a maximum number of evaluating the objective function equals 20000 for each run. The results, shown in Table 6, illustrate both the average gap (Av. GAP) and the success rates (Suc.) of all functions. The results, especially for the higher dimensional problems, reveal that the DSS algorithm has steadily better performance than the standard SS in terms of Av. GAP and the ability to obtain a global solution. The average gap calculated using DSS algorithm, with the maximum number of function evaluation equals 50,000, is reported in Table 7 over the set of 40 functions listed as Set B in Table 2. The DSS algorithm closely reaches near to global solutions as proven from the results shown in Tables 6 and 7. A performance comparison is carried out in the second experiment between the DSS algorithm and standard Scatter Search (SS) algorithms [32] as well as the other two metaheuristics methods designed for the continuous optimization problems. These other two methods are the Genetic Algorithm (GA) for constrained problems (Genocop III) [47], and Directed Tabu Search (DTS) [43]. The results of the Av. GAP calculated over the 40 functions of Set B at several points during the search are shown in Table 8. The method [32] ignores the GAP for Problem 23; therefore, we also ignore the result of that problem and provide the results of 39 problems. Our proposed DSS algorithm performs better than the SS as designed in [32], since the GAP < 1.0; except problem 23 with GAP = 78.99516181. All the SS have GAP < 1, except problems 23, 26, 34, and 40, which have GAP values equal 118.4341, 9.9496, 2.2441, and 5.5033, respectively. Another comparison criteria is the number of solved problems. The DSS solves 33 problems after 50000 function evaluations, however the SS [32] just solves 30 problems even with increasing the number of the functions evaluations from 20000 up to 50000.
The results in Table 8 show a substantial improvement of the values of Av. GAP applying DSS algorithm compared to the other methods at a large number of function evaluation. These results demonstrate the standing of the DSS method on the exploration process, which requests a high cost of function evaluations to guarantee good coverage to the search space. An additional performance measure is the number of optimal solutions found while applying each method. Figure 9 illustrates the number of optimal solutions reached by each algorithm during the search process with a stopping condition of 50,000 objective function evaluations. The figure demonstrates a better performance of the DSS algorithm at a large number of function evaluations by solving 33 at 50,000 evaluations, while Genocop III, SS, and DTS successfully solve a total of 23, 30, and 32 problems, respectively.

Stochastic Programming Results
The main results of the DESSP and DSSSP methods are shown in Table 9, which includes the averages of the best and the average of the error gaps of the obtained solutions. These results obtained through 100 independent runs of the DESSP and DSSSP codes with a maximum number of function evaluations equal to 500,000 in each run. These results illustrate that the DSSSP could attain very close solutions for four test functions f 1 , f 3 , f 5 , and f 6 . The results for functions f 2 and f 4 are not so far from the optimal ones. However, due to the big size of function f 7 , the result of DESSP was not close to the optimal solution. Moreover, the DSSSP algorithm has revealed a best performance especially for high dimensional function f 7 . For four test functions f 1 and f 6 , the DSSSP can attain very close solutions, and the solutions for functions f 2 , f 3 , f 4 , and f 5 are near the optimal ones. The performance of the proposed methods is measured with more accurate metric called Normalized Distance [48], which is defined using three components as follows: • Normalized Euclidean Distance (∆x): wheref min is the best minimum value of the objective function obtained by the proposed methods, and f max , f min are the global maximum and minimum values of the objective function, respectively, within the search space.
• Total Normalized Distance (∆ t ): Table 10 shows the values of these metrics for the results obtained by the DESSP and DSSSP methods.
Averages of processing time for running the DESSP and DSSSP codes are shown in Table 11. These averages are recorded in seconds after running the proposed codes on iMac machine with 3.2 GHz Quad-Core Intel Core i5 processor and 24 GB 1600 MHz DDR3 memory. The processing time of the DESSP code is better than that of the DSSSP code as shown in Table 11. However, the statistical test results in Table 12 do not favor either method against the other at the significance level 5%. In order to clarify the performance difference between the proposed methods, we can use the Wilcoxon rank-sum test [49][50][51]. This test belongs to the class of non-parametric tests which are recommended in our experiments [52,53]. Table 12 summarizes the results of applying Wilcoxon test on the results obtained by the proposed methods, it displays the sum of rankings obtained in each comparison and the p-value associated. In this statistical test, we assume that d i is the difference between the performance scores of the two methods on i-th out of N results. Then, we rank the differences according to their absolute values. The ranks R + and R − are computed as follows: The null hypothesis of equality of means is rejected if the min(R + , R − ) is less than or equal to the value of the distribution of Wilcoxon for N degrees of freedom (Table B.12 in [51]).
Although the shown results indicate that the DESSP method is slightly better than the DSSSP method, the statistical test results in Table 12 show that there is no significant difference between the two methods at the significance level 5%. The DESSP results have been compared against the results of the standard evolution strategies (ES) method and one of the best version of it which is called Covariance Matrix Adaptation Evolution Strategy (CMA-ES) [54]. The comparison results are shown in Table 13, which includes the averages of the best function values obtained by each method within 100 independent runs. From Table 13, the DESSP method has shown superior performance against the compared methods. A comparison between the proposed methods and the Cross-Entropy method [44] is also carried out. Figures 10 shows the results, which promise potentials of the proposed methods in attaining better objective function values in some tested function. It is worthwhile to mention that the final intensification plays an important role in improving the obtained solutions, as shown in the method performance with function f 1 in Figure 10.

Conclusions
In this paper, novel algorithms are designed to search for a globally optimal solution or near-globally optimal solution of an objective function that having random noisy variables. The proposed methods combine a modified version of the scatter search and evolution strategies with a variable-sample method, which is a Monte-Carlo based to handle the random variables. Moreover, a memory element called gene matrix is invoked which could guide the search process to new diverse solutions and also help in terminating the algorithms. We have implemented these methods in MATLAB. The performance of these proposed methods is tested using several sets of benchmark test functions of nonlinear and stochastic programming problems. The obtained results proved that these novel methods have promising performance compared with some standard methods in the literature. As future work, many other metaheuristics can be adapted to deal with stochastic programming by invoking different sampling techniques in order to interface the problem of noise. In addition, the fuzzy and rough set theories can be modeled instead of sampling techniques to deal with the uncertainty of objective functions with noise.