Next Article in Journal
Context-Aware Mode Selection for 5G Multi-Hop Cellular Networks
Next Article in Special Issue
Efficient Iterative Process Based on an Improved Genetic Algorithm for Decoupling Capacitor Placement at Board Level
Previous Article in Journal
Wireless Power and Data Transmission System of Submarine Cable-Inspecting Robot Fish and Its Time-Sharing Multiplexing Method
Previous Article in Special Issue
Automatic Emotion-Based Music Classification for Supporting Intelligent IoT Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low-Cost Multi-Objective Optimization of Multiparameter Antenna Structures Based on the l1 Optimization BPNN Surrogate Model

1
School of Computer Science and Engineering, Central South University, Changsha 410083, China
2
School of Aeronautics and Astronautics, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Electronics 2019, 8(8), 839; https://doi.org/10.3390/electronics8080839
Submission received: 5 July 2019 / Revised: 22 July 2019 / Accepted: 24 July 2019 / Published: 26 July 2019

Abstract

:
The development of modern wireless communication systems not only requires the antenna to be lightweight, low cost, easy to manufacture and easy to integrate but also imposes requirements on the miniaturization, wideband, and multiband design of the antenna. Therefore, designing an antenna that quickly and effectively meets multiple performance requirements is of great significance. To solve the problem of the large computational cost of traditional multi-objective antenna design methods, this paper proposes a backpropagation neural network surrogate model based on l1 optimization (l1-BPNN). The l1 optimization method tends to punish larger weight values and select smaller weight values so as to preserve a small amount of important weights and reset relatively unimportant weights to zero. By using l1 optimization method, the network mapping structure can be automatically adjusted to achieve the most suitable and compact structure of the surrogate model. Furthermore, for multi-parameter antenna design problems, a fast multi-objective optimization framework is constructed using the proposed l1-BPNN as a surrogate model. The framework is illustrated using a miniaturized multiband antenna design case, and a comparison with previously published methods, as well as numerical validation, is also provided.

1. Introduction

With the rapid development of radio communication technologies, wireless communication devices are becoming more diverse and intelligent, especially portable communication devices. Antennas play a decisive role in the performance of wireless communication devices and are an integral component of wireless communication devices. However, the multifunctional requirement of modern antenna design is undoubtedly a challenge for antenna researchers. A classic case is the miniaturization of the antenna, which not only requires the antenna to have a small size but also strict requirements on the electrical performance of the antenna (such as reflection coefficient, gain, etc.) [1,2,3]. Therefore, efficiently designing modern antennas that meet the structural and performance requirements has always been a popular research topic in the field of antennas. Whether to promote the theoretical development of antenna design or apply it to practical engineering design, the research topic has very important value.
Through certain optimization strategies, the process of finding suitable antenna parameters in a defined design space so that the antenna can achieve predefined performance requirements is called antenna optimization design. Traditional antenna optimization design usually requires that the antenna designer have a deep understanding of antenna design principles and a wealth of work experience in adjusting or correcting optimization parameters based on a certain antenna structure. However, such design methods make antenna design time costly, and the design requirements for complex antennas are difficult to satisfy. Initially, numerical calculation methods were used to analyze antenna performance, including the moment method (MOM) and the finite element method (FEM). The analysis results of these methods were basically consistent with the physical test results of the antenna [4]. Numerical calculation methods combined with optimization methods, such as the gradient method and the quasi-Newton method [5], could solve structural parameters that met electrical performance requirements. Then, more powerful antenna performance analysis tools emerged; that is, various electromagnetic (EM) simulation tools, including the high-frequency structure simulator (HFSS) and CST, could not only quickly analyze the various performances of the antenna but also optimize the antenna parameters and reduce the time burden of the antenna design. EM tools often optimize only the single antenna structure parameters, but the relationship between antenna structure parameters and performance is very complex and interdependent and influences their constraints. Thus, when the EM simulation tool performs multi-objective optimization of multiparameter antenna structures, the design process redundancy, optimization ability and efficiency become poor.
When optimizing the antenna structure, the designer needs to translate the actual antenna design problem into a mathematical description (i.e., the objective function). The choice of the objective function has a great influence on the antenna optimization design process. Common antenna optimization design objectives include gain, return loss and reflection coefficient. Because there are many influencing factors in the actual antenna structure design problem, the design of the objective function has many different forms and functional states. Different from simple linear, unipolar, and differential mathematical problems, the objective functions of the actual antenna solving problem are mostly highly nonlinear, have multiple extremums and are indivisible, and are difficult to express explicitly by mathematical formulas. At the same time, antenna optimization design is mostly a multi-objective optimization problem, for example, antenna optimization design with a low reflection coefficient and small size. Therefore, traditional antenna design methods often have difficulty finding global optimal solutions in the design space. However, the rapid development of intelligent optimization algorithms and computer technology effectively solves the above problems. The multi-objective intelligent optimization algorithm embodies its powerful local search and global convergence ability when solving multiparameter, large solution space and complex objective function optimization problems, which has gradually attracted the attention of antenna researchers.
In this paper, to improve optimization efficiency and reduce the computational cost, we propose a fast multi-objective optimization method for multiparameter antenna structures based on a surrogate model. This method uses a backpropagation neural network (BPNN) as a surrogate model to replace the computationally expensive EM tools and proposes a BPNN of l1 optimization structure (l1-BPNN) for overcoming the inherent defects of BPNN to construct a high-precision and streamlined neural network (NN) surrogate model. This method allows us to obtain a set of Pareto-optimal designs at a very low computational cost. Simultaneously, the surrogate model combined with the multi-objective evolutionary algorithm allows us to obtain the Pareto-optimal design at a lower computational cost.
This paper is organized as follows. Section 2 investigates the related work based on multiparameter antenna design. Section 3 introduces the motivation for improving the BPNN surrogate model. Section 4 describes the proposed fast multi-objective optimization method for multiparameter antenna structures based on the l1-BPNN surrogate model. Section 5 uses a three-band planar monopole antenna to illustrate our method and verify the effectiveness of the proposed method. Finally, Section 6 summarizes the study and proposes future work.

2. Related Work

2.1. Evolutionary Algorithms

Evolutionary algorithms (EAs) are derived from the process of biological evolution. The basic principle is to simulate the evolution of biological populations through multigenerational genetic variation processes to achieve environmental adaptation. Evolutionary algorithms are generally divided into four steps: (1) initializing the population; (2) calculating the fitness of each individual in the population; (3) generating the next generation of populations by means of selection, crossover, and mutation; (4) judging the termination condition, and if it is satisfied, terminating the evolutionary algorithm; otherwise, return to step (2) (as shown in Figure 1). Evolutionary algorithms include four typical methods: the genetic algorithm (GA) [6,7], genetic programming (GP) [8], evolution strategy (ES) [9] and evolutionary programming (EP) [10]. Swarm intelligence (SI) algorithms are also a special type of EA and can be defined as the collective behavior of decentralized and self-organized swarms. SI algorithms include particle swarm optimization (PSO) [11], ant colony optimization [12], and the artificial bee colony (ABC) algorithm [13] (Figure 2). Other recently proposed evolutionary techniques include wind driven optimization (WDO) [14]; invasive weed optimization (IWO) [15,16] and the covariance matrix adaptive evolution strategy (CMA-ES) [17,18].
Due to their intrinsic parallelism and powerful global search capability, EAs have emerged as viable candidates for global optimization problems, and a large number of antenna design problems have been addressed in the literature in recent years by using EAs. EAs can be real or binary-coded. Real-coded EAs are commonly used in antenna structure optimization design problems based on size or shape. This type of design can adjust the size or shape parameters of the antenna structure by an optimization algorithm based on the specified initial configuration to achieve the given single or multiple antenna performance indexes. In [19], the researchers used a combination of GA and full-wave EM tools to design a Sierpinski gasket fractal microstrip antenna. A novel coplanar waveguide fed planar monopole antenna for multiband operation was designed using a PSO algorithm in conjunction with the MOM method in [20]. Moreover, many researchers have proposed improved optimization algorithms for antenna design. For example, S. Baskar et al. [21] proposed a comprehensive learning PSO (CLPSO) algorithm, combined with the SuperNEC EM analysis toolkit to optimize the array element spacing and unit length of the Yagi–Uda antennas. Y. Sato et al. [22] proposed an adaptive GA to realize the rapid design of the meander line antenna structure. D. Ding et al. [23] proposed a modified multi-objective evolutionary algorithm based on decomposition to design a quad-band double-sided bow-tie antenna. In addition, in binary-coded EAs, each individual is encoded as a binary string, and the population is updated by crossover, mutation and a selection operator, which is well suited for discrete antenna topology optimization problems. Compared with size optimization and shape optimization, topology optimization can automatically generate “holes” in the design domain, thereby evolving the topology configuration of the structure and weakening the dependence of the entire design process on the initial configuration. In [24], the binary coded GA was applied to linear and planar array synthesis with arbitrary radiation patterns. In [25], a multi-objective evolutionary algorithm based on decomposition combined with enhanced genetic operators (MOEA/D-GO) was used to design a fragment-type MIMO dual antenna with a high isolation structure. In [26], an improved binary particle swarm optimization (BPSO) algorithm was used to design multiband or multitrap ultrawideband fragment-type antennas.

2.2. Surrogate Models

To accurately compute the antenna performance, a high-fidelity EM simulation is often required. However, it is computationally intensive to perform an EM simulation for a candidate solution, which poses a great challenge on rapid antenna design [27]. Moreover, the actual antenna design problem is mostly a multi-objective optimization problem (MOP). The use of multi-objective evolutionary algorithms (MOEAs) has led to a sharp increase in the number of evaluations of the objective function, which greatly increases the time cost of antenna optimization design, making it possible for a multi-objective optimization design to take days or even weeks to complete. Therefore, it is worthwhile improving the optimization speed at the cost of small degradation of the calculation accuracy.
Fortunately, surrogate model techniques [28,29,30,31,32,33,34,35,36,37,38,39,40,41] have been proven to effectively avoid the huge computational cost of the EM-driven process. The use of surrogate models has been a recurrent approach adopted by the evolutionary computational community to reduce the fitness function evaluations required to produce acceptable results. The surrogate model does not address how EM simulation evaluates antenna performance. It captures the relationship between the relevant information of the input and output variables and simulates the behavior of complex systems based on a set of collected examples. The surrogate model is designed to form a black box that can be mapped between the antenna design parameters and its performance indicators (e.g., return loss, gain, and efficiency). The antenna analysis model aims to construct a mapping relationship between design parameters and performance indexes, while the antenna synthesis model is the mapping relationship between performance indexes and design parameters. The antenna synthesis model and analysis model are shown in Figure 3a,b, respectively.
The principal surrogate model techniques applied to antenna optimization problems are the kriging models [28], the support vector machines (SVMs) [29], the Gaussian processes [30] and the artificial neural networks (ANNs) [31]. The techniques can be divided into direct fitness replacement (DFR) methods and indirect fitness replacement (IFR) methods [32]. In the former group, the approximated fitness replaces the original fitness during the entire course of the EA process, while in the latter group, some, but not all processes (e.g., population initialization or EA operators) use the approximated fitness. Kriging is a spatial prediction method that belongs to the group of geostatistical methods. It is based on minimizing the mean squared error, and it describes the spatial and temporal correlation among the values of an attribute. In [33,34], S. Koziel et al. used the low-fidelity sampling points of the coarse network to construct the kriging surrogate model and then used the spatial mapping method to correct the kriging model to realize the ultrawideband (UWB) single-cone antenna and fast multitarget design of the plane quasi-Yagi antenna. On this basis, they proposed constructing a cooperative kriging (co-kriging) model by adding a limited number of high-fidelity EM response data in the iterative optimization process, thus effectively improving the prediction accuracy of the surrogate model. SVM draws inspiration from statistical learning theory and is a set of related supervised learning methods that analyzes data and recognizes patterns. SVM constructs a hyperplane or a set of hyperplanes in a high-dimensional space that can be used for classification, regression, or other tasks. In [35], the SVM is used as a surrogate model to design rectangular patch antennas and rectangular patch antenna arrays. The GP method is derived from the kriging model [36], which is capable of building the surrogate model while providing the approximation error (i.e., a value of uncertainty/confidentiality) of the predictions over the whole input space without the need for testing samples. A GP is completely specified by its estimation function and uncertainty function (as shown in Figure 4). The estimation function gives the expectation of the process, and the uncertainty function defines the covariance between the output random variables. Once the estimation and uncertainty functions are defined, GP can be handled following rules of probability as applied to multivariate Gaussian distributions.
In [37], the researchers constructed a gradient-enhanced GP-based surrogate model by exploiting gradient information from adjoint simulations to reduce the number of training sample points, and the method was used to design dielectric resonators and UWB antennas. ANNs are based on a nonlinear parametric model that can provide a ‘universal’ approximation capability for emulating the functions describing the behavior of complex systems. The definition of an ANN requires two steps: (i) the selection of the ANN architecture and (ii) its training. As a substitute for the fine model with high computational burden, ANN technology is rapidly developing in antenna design problems [38,39,40,41]. In [38], a radial basis function neural network (RBFNN) was used to estimate the directivity of a uniform linear array of collinear short dipoles and parallel short dipoles. In [39], multilayer perceptron NN (MLP-NN)-based model and differential evolution algorithm are used to optimize performance index for reconfigurable antenna. In [40], the BP algorithm-based MLP-NN as antenna analysis model is combined with the particle swarm optimization (PSO) algorithm to form a design program for making a custom fractal antenna, and a Sierpinski gasket and Koch monopole antennas were taken as the candidate antennas to verify effectiveness of the developed approach. In [41], a sparsely connected neural network optimized (SC-BPNN) by HPSO algorithm was proposed and used as a surrogate model to participate in planar miniaturized multi-band antenna design. The proposed SC-BPNN has a high prediction accuracy and a simplified network structure, but the comparison experiment results show that the method has a larger time cost compared with other BPNN-based surrogate models as the HPSO optimization process is time consuming.

3. Research Motivation

BPNN are based on a non-linear parametric model which is able to provide an ’universal’ approximation capability for emulating the functions describing the behavior of complex systems. In theory, a multi-layer BPNN can approximate any complex nonlinear function [42]. However, in order to avoid excessive computational complexity, a three-layer BPNN as shown in Figure 5 is usually chosen. The neurons between adjacent layers are connected in a fully connected manner, and each connection corresponds to a weight value. The output layer obtains the error between the actual output and the expected result, and then updates the weight in direction of the largest gradient according to the gradient descent algorithm. Finally, the error is forwarded to the input layer once to complete the update of the weight. The quadratic cost function C0 of conventional BPNN and the weight and bias update formula are as follows
C 0 ( w , b ) = 1 2 n x i y k ( x i ) Y k ( x i ) 2
w w = w η C 0 w b b = b η C 0 b
where xi is the input of ith neuron and represents an antenna structure design variable; n is the number of input neurons; y k and Y k is the output of kth neuron and expected result respectively; n o is the number of output neurons; w denotes the mapping relationship between adjacent nodes; w denote the updated link weight; b and b denote the biases between adjacent nodes and the updated biases, respectively; η is a learning factor. The BPNN surrogate model is designed to construct a black box that can be mapped between the antenna design parameters and its performance indicators.
Since BPNN has the ability to approximate nonlinear maps well and can simulate nonlinear functions with an arbitrary precision, it is very suitable to replace the EM tools in traditional antenna designs. However, the traditional BPNN has some shortcomings, which makes the construction of the surrogate model too expensive and the prediction accuracy not high. First, the network structure of the traditional BPNN can only adjust the number of neurons in the input layer, the hidden layer, and the output layer, and cannot fine-tune the links between the layers, resulting in a network structure that is not compact enough and wastes computing resources. Second, the BPNN generally has over-fitting phenomenon, mainly due to too few training sample points or too complex network structures, resulting in network learning noise data in training sets. Therefore, a simple, compact BPNN is designed to reflect the performance of a multi-parameter antenna structure in real time.
The major innovation contributions as summarized as follows: (1) A new l1-BPNN surrogate model is proposed, and l1 optimization [43] is used to automatically determine all mapping parameters and states in the BPNN model to achieve the most appropriate and compact structure of the surrogate model; (2) Based on the proposed l1-BPNN model, a computationally efficient analysis strategy is established for predicting the performance indexes of the particular antenna structure, and then a fast multi-objective optimization framework for multiparameter structures is formed.

4. Our Approach

4.1. l1-BPNN Model

Since BPNN uses the gradient descent algorithm, the stability of the network requires little learning efficiency, resulting in a slower network convergence speed. At the same time, the multi-layer network is often interfered by the local optimal solution, and whether the learning process falls into the local optimal solution is closely related to the initial parameters (weights w and thresholds b) of the network [44]. In addition, the network structure will have an impact on the overall learning ability. Too complex network structure can cause the learning process to be too slow and easily lead to overfitting phenomenon [41]. On the contrary, too simple network structure may lead to insufficient learning ability.
Therefore, we propose a new formulation using l1 optimization to automatically determine network structure and connection parameter. The l1 optimization has the distinctive property for feature selection within the training process. In the l1 optimization process, the optimization goal is to maximize g, and at the end of l1 optimization, some weights in the neural networks are zeros while others remain nonzero. Zero weights mean that the corresponding parts of the mapping can be ignored and deleted. The cost function of l1-BPNN model is expressed as follows
C = C 0 + λ N w | w |
g = 1 1 + C
where C0 is used to determine the updated direction and updated size of the connection parameters; C is the updated cost function and connection parameter is updated as shown in formula (5); N is the size of the training set, and λ > 0 is the normalized parameter; sgn() is used to return the sign of the parameter.
w w = w η C 0 w η λ N sgn ( w ) b b = b η C 0 b
l1 optimization increases the punishment for large weights, tending to small network weights, so that a small number of important connection parameters in the network are preserved, while relatively less important connections will be close to 0, which is inactive, eventually making the network structure sparse. An l1-BPNN model is given in Figure 6, where logsig( ) is the most commonly used S-type transfer function for hidden layers, and defined as
log sig ( x i ) = 1 1 + e x i , x i .
Compared with the traditional BPNN model with fixed mapping structure, the l1-BPNN model can adjust the mapping structure automatically in addition to the number of neurons to achieve the most suitable and compact structure of the surrogate model.

4.2. Fast Multi-Objective Antenna Optimization Framework Combining MOEAs and l1-BPNN Surrogate Model

In this section, we use the l1-BPNN surrogate model discussed in the previous section, rather than conventionally time-consuming EM simulations, to evaluate the antenna performance. The first step in building a surrogate model is to use Latin hypercube sampling (LHS) [45,46] to obtain a uniformly distributed and representative sample set S = [s1,s2,…,sq]T, and to import those sample sets S into the EM tool to achieve high-accuracy response set Y. The quality of S and Y is directly related to the generalization ability of the surrogate model, but the process is the most time consuming in the overall surrogate model construction. Then S and Y will be used as input layer data and output layer data to construct antenna surrogate model Rs(x), respectively. The l1 optimization method automatically adjusts the network mapping structure in this process. Thus, the whole multi-objective optimization framework is summarized as follows:
  • Predefine the design space;
  • Determine the number of neurons in each layer of NN and the antenna geometry vector x;
  • Sample design space using LHS and acquire the response set Y;
  • Adjust the network mapping structure by l1 optimization;
  • Construct an l1-BPNN surrogate model Rs(x);
  • Optimize the population by MOEA with an l1-BPNN surrogate model;
  • If termination condition is not satisfied then go to 6; else end optimization.
The flowchart of MOEA with an l1-BPNN surrogate model is shown in Figure 7. It is expected that our approach will greatly reduce the computational cost of the antenna optimization process and meanwhile speed up the convergence.

5. Verification Case Study and Discussions

This section will give an application example of a miniaturized planar triple-band antenna design to further illustrate the advantages of the fast multi-objective antenna optimization framework constructed by l1-BPNN model.

5.1. l1-BPNN Antenna Surrogate Model

The initial structure of the planar monopole antenna is shown in Figure 8. This antenna is formed by a fork-shaped radiator and a rectangle ground plane, which produces different resonant frequency bands to satisfy multiband applications. The antenna is printed on a Rogers RO4003(tm) substrate with a thickness of 0.5 mm, permittivity of 3.55, and loss tangent of 0.0027. Design variables are x = [   l l 1   l 2   l 3 l 4   l 5   w   w 1 w 2   g   ] T and initial ranges are given in Table 1 (Units: mm).
A high-precision surrogate model means that a large number of training sets and response sets need to be acquired, which makes the computational cost increase dramatically. The surrogate model we want to construct is low-cost and can respond to trends in performance indexes. Therefore, the LHS is introduced to sample design space and obtain a representative sample set S including 200 sample points. Next, we acquire the high fidelity response set Y of sampling points by using EM simulation software HFSS, S and Y are used as input layer data and output layer data to construct l1-BPNN surrogate model, respectively. As for the determination of the number of hidden layer nodes, although there are some empirical formulas for reference, there is no mature theory as a guide. In order to ensure the stability of the network, experimental tests are used to determine the number of hidden layer nodes. The node test interval is taken as [10, 20]. Then S and Y will be used as input layer data and output layer data to construct antenna surrogate model Rs(x), respectively. The l1 optimization method automatically adjusts the network mapping structure in this process.
Table 2 shows the fitness values and the number of network mapping of the three models (the proposed l1-BPNN over the traditional BPNN [40], the BPNN optimized by PSO (PSO-BPNN) [47]). From Table 2, we have the observation that the l1-BPNN generally has better fitness values and a simpler network mapping structure than other models. Through this experiment, we choose the network mapping structure when the number of hidden nodes is 19, and its connection parameters only account for 72% of other models.
Figure 9 shows the error value curves of the three models during training process. It can be observed that the number of iterations and the training error of the proposed model are smaller than the other two models, indicating that l1-BPNN has better convergence performance and training accuracy.
In addition, Figure 10 shows a comparison of the predictions of Kriging [33], traditional BPNN [40], PSO-BPNN [47], and l1-BPNN with the HFSS response results. Compared to Figure 10b–d, Figure 10a is more concentrated, indicating that the predicted structure of l1-BPNN is closer to the HSFF response. Table 3 clearly lists the time costs of different surrogate models and EM simulations. It is observed that the surrogate model is generally less time consuming.
In summary, our proposed l1-BPNN model can be used for performance prediction efficiently instead of EM simulation software and achieve a fast multi-objective optimization with the help of MOEAs for a predefined antenna geometry.

5.2. Pareto-Optimal Designs of a Planar Miniaturized Multiband Antenna

In this section, the MOEA/D [48] and l1-BPNN model are combined to achieve fast multi-objective optimization of the planar antenna geometry presented in Figure 9. Two design goals are (i) the values of S11 are lower than −10dB within the frequency bands of 2.40–2.60 GHz, 3.30–3.80 GHz, 5.00–5.85 GHz, covering the entire WLAN2.4/5.2/5.8 GHz and WiMAX3.5 GHz applications (objective F1); (ii) the size of the antenna structure is reduced to satisfy the need for antenna miniaturization in portable electric devices (objective F2). The objective function of F1 is specified as
F 1 = 1 N i = 1 n Q ( f i )
Q ( f i ) = { | S 11 ( f i ) | | S 11 ( f i ) | > 10 10 | S 11 ( f i ) | 10
where fi is the ith sampling frequency point within the given bands of operation; S 11 ( f i ) is the reflection coefficient of sampling point fi; N is the number of sampling frequency points. The objective function of F2 can be defined as
F 2 = w × l
The parameters of MOEA/D [48] are set as follows: the number of population is 100, and the maximum number of iterations is 150. Figure 11 displays the representation of the Pareto set of the optimized antenna at different iterations. It can be observed that as the number of iterations increases, the objective function becomes closer to a smaller size and a lower S11 value. The detailed antenna designs are given in Table 4. Furthermore, Figure 12 demonstrates the good fitting ability of the l1-BPNN model. The HFSS simulation curve of the Pareto-optimal designs well matches the l1-BPNN prediction values. In addition, S11 curves of all designs cover the target frequency band (red shadow) to meet the preset application requirements. Figure 12 shows how close the predicted value of the surrogate model we constructed is to the EM curve. Additionally, it is observed that the S11 curve is lower than −10 dB within the three frequency bands of 2.33–2.66 GHz, 3.05–3.80 GHz, and 5.06–5.96 GHz, simultaneously satisfying the WLAN and WiMAX applications.
In order to verify the feasibility and validity of l1-BPNN model, the results of fitness F1 and percentage error for those selected Pareto-optimal designs, which are calculated by different approaches, are listed in Table 5, and Figure 13 shows an error comparison histogram between simulation results and prediction results in Table 5. The predictive results 1 and predictive results 2 are given by traditional BPNN and l1-BPNN surrogate model strategy, respectively. The minimum error percentages are 3.58% and 1.33% for prediction results 1 (with an average error of 7.69%) and prediction results 2 (with an average error of 2.65%), respectively. Therefore, our proposed l1-BPNN is meaningfully better than the conventional BPNN with an acceptable error rate.
Furthermore, Table 6 lists the computational cost comparisons for the four antenna optimization methods. Scheme I is direct MOEA/D-based optimization without a surrogate model. Scheme II, Scheme III and Scheme IV are based on the traditional BPNN [40], SC-BPNN [41], and l1-BPNN surrogate model, respectively. As listed in Table 6, the time cost of Scheme IV represents only 1.26% of scheme I, scheme II is approximately 1.37% of the time cost of scheme I, and scheme III is approximately 1.51% of the time cost of scheme I. Figure 14 shows the time-consuming ratio of each part of the multi-objective optimization method using the different surrogate models. As shown in Figure 14b, since SC-BPNN requires a time-consuming HPSO optimization process, its model construction time cost is the largest among these three surrogate-assisted methods. On the contrary, as shown in Figure 14c, since our proposed model has a simplified network structure and requires fewer training times, its model construction time cost is the smallest among these three surrogate-assisted methods.

6. Conclusions

An improved BPNN surrogate model is developed in this paper to applied low-cost multi-objective multiparameter antenna optimization. To overcome the shortcomings of the traditional BPNN, an l1-BPNN model is proposed. Some modifications are given for better network performance, including automatically adjusting the network mapping structure by l1 optimization. Compared with other existing BPNN-based surrogate models, the proposed l1-BPNN model has smaller construction cost, better convergence performance, and prediction accuracy. Then, by integrating the l1-BPNN surrogate model with MOEAs, a low-cost antenna design method is established, which can quickly realize the optimal design of multi-objective and multi-parameter antenna structures. This technique has been illustrated using a planar antenna structure. Comparison with the previously published surrogate model scheme indicates significant savings of the overall optimization cost that can be achieved with our l1 optimization method.
Although the results from this study are very encouraging, further research will be conducted to address several issues. First, the connection parameters are decremented to zero at a constant speed during l1 optimization. When the absolute value of a weight is large, the speed of optimization is too slow. So, we intend to improve the l1 algorithm so that the weight reduction speed is related to the absolute value of the weight to improve the optimization efficiency. Second, the proposed method is a direct fitness surrogate method, which cannot be updated in the EA process. Therefore, we want to study the dynamically updated BPNN surrogate model, which aims to improve the utilization of EM data and further reduce its construction cost.

Author Contributions

Conceptualization, W.Q., J.D. and J.M.; methodology, J.D. and J.M.; software, W.Q.; validation, W.Q.; formal analysis, J.D. and J.M.; investigation, W.Q.; resources, J.D. and J.M.; data curation, W.Q.; writing—original draft preparation, J.D. and W.Q.; writing—review and editing, J.M.; visualization, W.Q. and J.D.; supervision, J.M.; project administration, J.M.; funding acquisition, J.M.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant No. 61801521 & 61471368, in part by the Natural Science Foundation of Hunan Province under Grant No. 2018JJ2533, and in part by the Fundamental Research Funds for the Central Universities under Grant No. 2018gczd014 & 20190038020050.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Milligan, T.A. Modern Antenna Design, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  2. Fujimoto, K. Mobile Antenna Systems Handbook, 3rd ed.; Artech House: Norwood, MA, USA, 2008. [Google Scholar]
  3. Dziunikowski, W. Multiple-Input Multiple-Output (MIMO) Antenna Systems. In Adaptive Antenna Arrays: Signals and Communication Technology; Chandran, S., Ed.; Springer: Berlin, Germany, 2004; pp. 259–273. [Google Scholar]
  4. McGrath, D.; Pyati, V. Phased array antenna analysis with the hybrid finite element method. IEEE Trans. Antennas Propag. 1994, 42, 1625–1630. [Google Scholar] [CrossRef]
  5. Choi, W.-S.; Sarkar, T. Phase-Only Adaptive Processing Based on a Direct Data Domain Least Squares Approach Using the Conjugate Gradient Method. IEEE Trans. Antennas Propag. 2004, 52, 3265–3272. [Google Scholar] [CrossRef]
  6. Goldberg, D.E. Genetic Algorithms in Search, Optimization and Machine Learning; Addison-Wesley: New York, NY, USA, 1989. [Google Scholar]
  7. Holland, J.H. Adaptation in Natural and Artificial Systems; The University of Michigan Press: Ann Arbor, MI, USA, 1975. [Google Scholar]
  8. Koza, J.R.; Poli, R. Genetic Programming. In Search Methodologies; Burke, E.K., Kendall, G., Eds.; Springer: Boston, MA, USA, 2005; pp. 127–164. [Google Scholar]
  9. Beyer, H.-G.; Schwefel, H.-P. Evolution strategies – A comprehensive introduction. Nat. Comput. 2002, 1, 3–52. [Google Scholar] [CrossRef]
  10. Fogel, D.B. Evolutionary Computation: Toward a New Philosophy of Machine Intelligence; IEEE Press: Piscataway, NJ, USA, 1995. [Google Scholar]
  11. Kennedy, J.; Eberhart, R.C. Particle swarm optimization. In Proceedings of the 1995 IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  12. Dorigo, M.; Stutzle, T. Ant Colony Optimization; The MIT Press: Cambridge, MA, USA, 2004. [Google Scholar]
  13. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  14. Bayraktar, Z.; Bossard, J.A.; Werner, D.H.; Komurcu, M. The Wind Driven Optimization Technique and its Application in Electromagnetics. IEEE Trans. Antennas Propag. 2013, 61, 2745–2757. [Google Scholar] [CrossRef]
  15. Zaharis, Z.D.; Lazaridis, P.I.; Cosmas, J.; Skeberis, C.; Xenos, T.D. Synthesis of a Near-Optimal High-Gain Antenna Array With Main Lobe Tilting and Null Filling Using Taguchi Initialized Invasive Weed Optimization. IEEE Trans. Broadcast. 2014, 60, 120–127. [Google Scholar] [CrossRef] [Green Version]
  16. Zaharis, Z.; Skeberis, C.; Xenos, T.D.; Lazaridis, P.; Cosmas, J. Design of a Novel Antenna Array Beamformer Using Neural Networks Trained by Modified Adaptive Dispersion Invasive Weed Optimization Based Data. IEEE Trans. Broadcast. 2013, 59, 455–460. [Google Scholar] [CrossRef] [Green Version]
  17. Boudaher, E.; Hoorfar, A. Electromagnetic design optimization using mixed-parameter and multiobjective CMA-ES. In Proceedings of the IEEE Antennas and Propagation Society International Symposium (APSURSI ’13), Orlando, FL, USA, 7–13 July 2013; pp. 406–407. [Google Scholar]
  18. BouDaher, E.; Hoorfar, A. Electromagnetic Optimization Using Mixed-Parameter and Multiobjective Covariance Matrix Adaptation Evolution Strategy. IEEE Trans. Antennas Propag. 2015, 63, 1712–1724. [Google Scholar] [CrossRef]
  19. Ghatak, R.; Poddar, D.; Mishra, R. Design of Sierpinski gasket fractal microstrip antenna using real coded genetic algorithm. IET Microwaves Antennas Propag. 2009, 3, 1133–1140. [Google Scholar] [CrossRef]
  20. Liu, W.-C. Design of a multiband CPW-fed monopole antenna using a particle swarm optimization approach. IEEE Trans. Antennas Propag. 2005, 53, 3273–3279. [Google Scholar]
  21. Baskar, S.; Alphones, A.; Suganthan, P.N.; Liang, J.J. Design of Yagi-Uda antennas using comprehensive learning particle swarm optimization. IEE Proc. Microw. Antennas Propag. 2005, 152, 340–346. [Google Scholar] [CrossRef]
  22. Sato, Y.; Campelo, F.; Igarashi, H. Meander Line Antenna Design Using an Adaptive Genetic Algorithm. IEEE Trans. Magn. 2013, 49, 1889–1892. [Google Scholar] [CrossRef] [Green Version]
  23. Ding, D.; Wang, G. Modified Multiobjective Evolutionary Algorithm Based on Decomposition for Antenna Design. IEEE Trans. Antennas Propag. 2013, 61, 5301–5307. [Google Scholar] [CrossRef]
  24. Marcano, D.; Duran, F. Synthesis of antenna arrays using genetic algorithms. IEEE Antennas Propag. Mag. 2000, 42, 12–20. [Google Scholar] [CrossRef]
  25. Ding, D.; Wang, L.; Wang, G. High-efficiency scheme and optimisation technique for design of fragment-type isolation structure between multiple-input and multiple-output antennas. IET Microw. Antennas Propag. 2015, 9, 933–939. [Google Scholar] [CrossRef]
  26. Dong, J.; Li, Q.; Deng, L. Design of Fragment-Type Antenna Structure Using an Improved BPSO. IEEE Trans. Antennas Propag. 2018, 66, 564–571. [Google Scholar] [CrossRef]
  27. John, M.; Ammann, M.J. Antenna Optimization with a Computationally Efficient Multiobjective Evolutionary Algorithm. IEEE Trans. Antennas Propag. 2009, 57, 260–263. [Google Scholar] [CrossRef]
  28. Sacks, J.; Mitchell, T.J.; Welch, W.J.; Wynn, H.P. Design and Analysis of Computer Experiments. Stat. Sci. 1989, 4, 409–423. [Google Scholar] [CrossRef]
  29. Vapnik, V.N. Statistical Learning Theory, Adaptive and Learning Systems for Signal Processing, Communications, and Control; John Wiley & Sons: New York, NY, USA, 1998. [Google Scholar]
  30. Rasmussen, C.; Williams, C. Gaussian Processes for Machine Learning; The MIT Press: Cambridge, MA, USA, 2008. [Google Scholar]
  31. Mao, J.; Mohiuddin, K.; Jain, A. Artificial neural networks: A tutorial. Computer 1996, 29, 31–44. [Google Scholar]
  32. Shi, L.; Rasheed, K. A Survey of Fitness Approximation Methods Applied in Evolutionary Algorithms. In Adaptation, Learning, and Optimization; Springer: Berlin, Germany, 2010; Volume 2, pp. 3–28. [Google Scholar]
  33. Koziel, S.; Ogurtsov, S. Multi-Objective Design of Antennas Using Variable-Fidelity Simulations and Surrogate Models. IEEE Trans. Antennas Propag. 2013, 61, 5931–5939. [Google Scholar] [CrossRef]
  34. Koziel, S.; Bekasiewicz, A.; Couckuyt, I.; Dhaene, T. Efficient Multi-Objective Simulation-Driven Antenna Design Using Co-Kriging. IEEE Trans. Antennas Propag. 2014, 62, 5900–5905. [Google Scholar] [CrossRef]
  35. Zheng, Z.; Chen, X.; Huang, K. Application of support vector machines to the antenna design. Int. J. RF Microw. Comput. Aided Eng. 2010, 21, 85–90. [Google Scholar] [CrossRef]
  36. Massa, A.; Oliveri, G.; Salucci, M.; Salucci, N.; Rocca, P. Learning-by-examples techniques as applied to electromagnetics. J. Electromagn. Waves Appl. 2018, 32, 516–541. [Google Scholar] [CrossRef]
  37. Ulaganathan, S.; Koziel, S.; Bekasiewicz, A.; Couckuyt, I.; Laermans, E.; Dhaene, T. Data-driven model based design and analysis of antenna structures. IET Microw. Antennas Propag. 2016, 10, 1428–1434. [Google Scholar] [CrossRef]
  38. Mishra, S.; Yadav, R.N.; Singh, R.P. Directivity Estimations for Short Dipole Antenna Arrays Using Radial Basis Function Neural Networks. IEEE Antennas Wirel. Propag. Lett. 2015, 14, 1219–1222. [Google Scholar] [CrossRef]
  39. Mahouti, P. Design optimization of a pattern reconfigurable microstrip antenna using differential evolution and 3D EM simulation-based neural network model. Int. J. RF Microw. Comput. Aided Eng. 2019, 29, e21796. [Google Scholar] [CrossRef]
  40. Anuradha, A.; Patnaik, A.; Sinha, S.N. Design of custom-made fractal multi-band antennas using ANN-PSO. IEEE Antennas Propag. Mag. 2011, 53, 94–101. [Google Scholar] [CrossRef]
  41. Dong, J.; Qin, W.; Wang, M. Fast Multi-Objective Optimization of Multi-Parameter Antenna Structures Based on Improved BPNN Surrogate Model. IEEE Access 2019, 7, 77692–77701. [Google Scholar] [CrossRef]
  42. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  43. Bandler, J.W.; Chen, S.H.; Daijavad, S. Microwave device modeling using efficient I1 optimization: A novel approach. IEEE Trans. Microw. Theory Tech. 1986, 34, 1282–1293. [Google Scholar] [CrossRef]
  44. Zhang, J.-R.; Zhang, J.; Lok, T.-M.; Lyu, M.R. A hybrid particle swarm optimization–back-propagation algorithm for feedforward neural network training. Appl. Math. Comput. 2007, 185, 1026–1037. [Google Scholar] [CrossRef]
  45. Stein, M. Large Sample Properties of Simulations Using Latin Hypercube Sampling. Technometrics 1987, 29, 143–151. [Google Scholar] [CrossRef]
  46. Owen, A.B. A Central Limit Theorem for Latin Hypercube Sampling. J. R. Stat. Soc. Ser. B 1992, 54, 541–551. [Google Scholar] [CrossRef]
  47. Dong, J.; Qin, W.; Li, Q.; Deng, L. Fast multi-objective antenna design based on improved back propagation neural network surrogate model. J. Electron. Inf. 2018, 40, 2712–2719. (In Chinese) [Google Scholar]
  48. Zhang, Q.F.; Li, H. MOED/A: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
Figure 1. Evolutionary algorithm flow chart.
Figure 1. Evolutionary algorithm flow chart.
Electronics 08 00839 g001
Figure 2. A diagram depicting the main families of evolutionary algorithms.
Figure 2. A diagram depicting the main families of evolutionary algorithms.
Electronics 08 00839 g002
Figure 3. The antenna analysis model (a) and synthesis model (b).
Figure 3. The antenna analysis model (a) and synthesis model (b).
Electronics 08 00839 g003
Figure 4. Gaussian processes specified by the estimation function and the uncertainty function.
Figure 4. Gaussian processes specified by the estimation function and the uncertainty function.
Electronics 08 00839 g004
Figure 5. The structure of a conventional three-layer BPNN.
Figure 5. The structure of a conventional three-layer BPNN.
Electronics 08 00839 g005
Figure 6. The structure of the proposed l1-BPNN.
Figure 6. The structure of the proposed l1-BPNN.
Electronics 08 00839 g006
Figure 7. Flowchart of the fast multi-objective antenna optimization framework combining the l1-BPNN surrogate model and MOEAs.
Figure 7. Flowchart of the fast multi-objective antenna optimization framework combining the l1-BPNN surrogate model and MOEAs.
Electronics 08 00839 g007
Figure 8. Geometry of the planar multiband antenna.
Figure 8. Geometry of the planar multiband antenna.
Electronics 08 00839 g008
Figure 9. Training error curves of BPNN, PSO-BPNN and l1-BPNN.
Figure 9. Training error curves of BPNN, PSO-BPNN and l1-BPNN.
Electronics 08 00839 g009
Figure 10. Comparison of actual versus predicted |S11| values when using (a) l1-BPNN, (b) PSO-BPNN, (c) BPNN, (d) kriging surrogate models.
Figure 10. Comparison of actual versus predicted |S11| values when using (a) l1-BPNN, (b) PSO-BPNN, (c) BPNN, (d) kriging surrogate models.
Electronics 08 00839 g010aElectronics 08 00839 g010b
Figure 11. The obtained representations of the Pareto set for the planar multiband antenna.
Figure 11. The obtained representations of the Pareto set for the planar multiband antenna.
Electronics 08 00839 g011
Figure 12. Simulated (—) and predicted (○) reflection responses for the designs of the PF in Figure 11 (from top left to bottom right): (a) x(1), (b) x(2) , (c) x(3) , (d) x(4) , (e) x(5) , (f) x(6).
Figure 12. Simulated (—) and predicted (○) reflection responses for the designs of the PF in Figure 11 (from top left to bottom right): (a) x(1), (b) x(2) , (c) x(3) , (d) x(4) , (e) x(5) , (f) x(6).
Electronics 08 00839 g012aElectronics 08 00839 g012b
Figure 13. Error Comparison Histogram between Simulation Results and Prediction Results in Table 5.
Figure 13. Error Comparison Histogram between Simulation Results and Prediction Results in Table 5.
Electronics 08 00839 g013
Figure 14. Time-consuming ratio of each part of the multi-objective optimization method using different surrogate models: (a) Scheme II, (b) Scheme III, and (c) Scheme IV.
Figure 14. Time-consuming ratio of each part of the multi-objective optimization method using different surrogate models: (a) Scheme II, (b) Scheme III, and (c) Scheme IV.
Electronics 08 00839 g014
Table 1. Initial Ranges of Design Parameters.
Table 1. Initial Ranges of Design Parameters.
Parameterll1l2l3l4
Range[36, 40][16, 19][10, 12.5][8.5, 10.5][2.8, 3.9]
Parameterl5ww1w2g
Range[9.5, 11.5][19, 24][6.5, 8.3][8.7, 11.2][1.8, 2.1]
Table 2. Comparison of Fitness Values for Predicting the S11 Values.
Table 2. Comparison of Fitness Values for Predicting the S11 Values.
nhl1-BPNNConventional BPNNPSO-BPNN
Fitness ValuesNumber of LinksFitness ValuesNumber of LinksFitness ValuesNumber of Links
100.93181540.88922750.9190275
110.94351860.87813010.8879301
120.94192350.89713270.9074327
130.94572540.90563530.9313353
140.94782100.90223790.9147379
150.93242480.91304050.9203405
160.92992590.89754310.9138431
170.93652820.90874570.9312457
180.95283110.89094830.9383483
190.95313670.89255090.9205509
200.95013890.90535350.9226535
Table 3. Computational Time of Various Surrogate Models and HFSS Simulations.
Table 3. Computational Time of Various Surrogate Models and HFSS Simulations.
ModelHFSSKrigingBPNNPSO-BPNNl1-BPNN
Total time (s)587.821.920.360.280.26
Average time(s)58.780.1920.0360.0280.026
Table 4. Planar Multiband Antenna: Selected Pareto-Optimal Designs.
Table 4. Planar Multiband Antenna: Selected Pareto-Optimal Designs.
Design x ( 1 ) x ( 2 ) x ( 3 ) x ( 4 ) x ( 5 ) x ( 6 )
F 1 [ d B ] −13.52−13.27−12.78−12.49−12.20−11.74
F 2 [ m m 2 ] 889.35862.14838.10801.27781.19754.46
l38.438.739.038.039.137.9
l117.518.317.117.816.617.5
l212.411.212.211.411.911.7
l310.29.99.610.39.410.2
l43.33.23.33.13.33.0
l59.79.69.910.110.49.8
w23.222.321.521.119.919.7
w16.57.17.37.27.47.5
w210.49.410.29.69.59.0
g2.01.91.92.01.91.9
Table 5. Comparison of Fitness Values F1 of Selected Pareto-Optimal Designs Obtained by Surrogate Model and HFSS.
Table 5. Comparison of Fitness Values F1 of Selected Pareto-Optimal Designs Obtained by Surrogate Model and HFSS.
Designx(1)x(2)x(3)x(4)x(5)x(6)
Simulation results−13.52−13.27−12.78−12.49−12.20−11.74
Predictive results I−14.41−11.71−13.75−11.39−13.15−11.32
Predictive results II−13.70−13.82−13.10−12.77−11.92−12.19
Error rate I (%)6.5911.767.598.817.793.58
Error rate II (%)1.334.142.502.242.293.38
Table 6. Comparison of Computational Cost among Different Antenna Optimization Schemes.
Table 6. Comparison of Computational Cost among Different Antenna Optimization Schemes.
Optimization ApproachNumber of EM SimulationsCPU Time/h
TotalRelative (%)
Scheme I15,000245.96100
Scheme II2003.381.37
Scheme III2003.711.51
Scheme IV2003.111.26

Share and Cite

MDPI and ACS Style

Dong, J.; Qin, W.; Mo, J. Low-Cost Multi-Objective Optimization of Multiparameter Antenna Structures Based on the l1 Optimization BPNN Surrogate Model. Electronics 2019, 8, 839. https://doi.org/10.3390/electronics8080839

AMA Style

Dong J, Qin W, Mo J. Low-Cost Multi-Objective Optimization of Multiparameter Antenna Structures Based on the l1 Optimization BPNN Surrogate Model. Electronics. 2019; 8(8):839. https://doi.org/10.3390/electronics8080839

Chicago/Turabian Style

Dong, Jian, Wenwen Qin, and Jinjun Mo. 2019. "Low-Cost Multi-Objective Optimization of Multiparameter Antenna Structures Based on the l1 Optimization BPNN Surrogate Model" Electronics 8, no. 8: 839. https://doi.org/10.3390/electronics8080839

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop