Next Article in Journal
Design of a Hierarchical Control Architecture for Fully-Driven Multi-Fingered Dexterous Hand
Previous Article in Journal
A Novel Improved Dung Beetle Optimization Algorithm for Collaborative 3D Path Planning of UAVs
Previous Article in Special Issue
ADVCSO: Adaptive Dynamically Enhanced Variant of Chicken Swarm Optimization for Combinatorial Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Evolutionary Neural Architecture Search Algorithm Without Training

1
Software College, Northeastern University, Shenyang 110167, China
2
College of Computer Science and Engineering, Ningxia Institute of Science and Technology, Shizuishan 753000, China
3
Lightspeed Studios, Tencent Co., Ltd., Shenzhen 518000, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(7), 421; https://doi.org/10.3390/biomimetics10070421
Submission received: 22 May 2025 / Revised: 19 June 2025 / Accepted: 27 June 2025 / Published: 29 June 2025
(This article belongs to the Special Issue Exploration of Bio-Inspired Computing)

Abstract

Neural Architecture Search (NAS) has made significant advancements in autonomously constructing high-performance network architectures, capturing extensive attention. However, a key challenge of existing NAS approaches is the intensive performance evaluation, leading to significant time and computational resource consumption. In this paper, we propose an efficient Evolutionary Neural Architecture Search (ENAS) method to address this issue. Specifically, in order to accelerate the convergence speed of the algorithm and shorten the search time, thereby avoiding blind searching in the early stages of the algorithm, we drew on the principles of biometrics to redesign the interaction between individuals in the evolutionary algorithm. By making full use of the information carried by individuals, we promoted information exchange and optimization between individuals and their neighbors, thereby improving local search capabilities while maintaining global search capabilities. Furthermore, to accelerate the evaluation process and minimize computational resource consumption, a multi-metric training-free evaluator is introduced to assess network performance, bypassing the resource-intensive training phase, and the adopted multi-metric combination method further solves the ranking offset problem. To evaluate the performance of the proposed method, we conduct experiments on two widely adopted benchmarks, NAS-Bench-101 and NAS-Bench-201. Comparative analysis with state-of-the-art algorithms shows that our proposed method identifies network architectures with comparable or better performance while requiring significantly less time.

1. Introduction

The field of machine learning is undergoing an unprecedented boom [1], driven by pioneering networks such as Res-Net [2], VGG [3], and DenseNet [4]. These networks have played a pivotal role in yielding breakthrough results across a spectrum of application domains, encompassing image classification [5], medical image analysis [6], vegetation remote sensing [7]. But the success achieved by neural networks in these fields relies on the expert experience and domain-specific knowledge possessed by neural network designers, and this knowledge is not easily and quickly obtainable. The process of adjusting neural networks also leads to a significant waste of time and resources due to repeated experiments. This makes the manual design of neural networks challenging and inefficient [8].
To address the challenges arising from the manual design of neural networks, researchers have developed neural architecture search algorithms (NAS) [9], with the anticipation of autonomously obtaining excellent neural network architectures for the target domain, without the need for manual intervention. Neural architecture search primarily encompasses three aspects [10]: search space, search strategy, and performance evaluation. Search algorithms explore the search space to discover networks with superior performance, which is evaluated during the performance evaluation phase. Therefore, neural network architecture search can be modeled as an optimization problem aiming to achieve optimal performance. NAS has also found applications in various domains, such as medical image segmentation [11], video super resolution [12], facial expression recognition [13].
Based on different search strategies, neural network architecture search algorithms can be categorized into three types: reinforcement learning (RL)—based [14], gradient descent (GD)—based [15], and evolutionary computation(EC)—based [16]. In RL-based algorithms [17,18], the actions of the agent correspond to the generation of neural network architectures. The action space corresponds to the search space, and the reward corresponds to the accuracy of the designed architecture. Through the reward, RL can train the controller to prefer architectures with higher accuracy in the next iteration. The extensive evaluation process makes RL-based methods consume significant computational resources, resulting in high search costs. Despite recent efforts to improve through multi-agent parallelism [19], the performance remains unsatisfactory. In comparison to random search and reinforcement learning, GD-based methods exhibit better efficiency in search. This approach, through continuous relaxation, allows researchers to transform the discrete search space into a continuous space, enabling the use of gradient descent to search for architectures. Evidently, GD-based methods offer a substantial advantage in terms of computational speed. However, these methods are sensitive to initialization, implying the need for strong domain knowledge when constructing the initial search space. Additionally, due to their inherent alternating optimization characteristics, they may lead to performance collapse [20,21].
Compared to other methods, EC-based methods are less restricted by search space requirements, more automated, and insensitive to gradients, and have been widely adopted as search algorithms. Specifically, [22], as a pioneer of ENAS, adopts genetic algorithms as a search algorithm. The previous tournament selection operator was modified in [23] to discard older individuals in the population. Two levels of particle swarm optimization are used to evolve the whole network architecture in [24,25] proposes a new encoding method and crossover operator adapted to the need of variable length. In these methods, the main computational resource consumption is focused on the evaluation process of the network, and some improvements have been proposed in the subsequent methods, ref. [26] adopts an incremental evaluation to shorten the evaluation process, ref. [27] adopts the weight inheritance to reduce the cost of the evaluation, and [28] reduces the number of evaluations by the similarity evaluation of the individuals. However, in most of these works, there is an improvement of the operator without considering the relationship between individuals in the population and the utilization of the relevant information carried by the individuals in the problem. This means that in the early stages of the search, the search space is less utilized and time is wasted in the process of blind exploration.
Furthermore, when EC is employed as a search algorithm, it usually involves a large number of performance evaluation processes, which leads to high search costs that hinder the development of related research, among the algorithms mentioned above [22,23], consumed thousands of GPU Days (the number of GPUs used multiplied by the number of days searched) to get the results of the search. Subsequent methods have been proposed to try to solve this problem, e.g., weight inheritance [29]; early stopping [30]; low fidelity [31]; however, in these methods, training evaluation of the individuals is still required, and the acceleration results are not satisfactory, and usually the search process still takes several GPU days. Evaluation methods that do not require training have been applied to the evaluation process in NAS with the development of neural network interpretability. These methods [32,33] avoids the expensive training process by scoring the network at the initialization stage of the network, but in neural networks, the factors that affect the performance are more complex. Therefore the accuracy of the evaluation is not good. Meanwhile, in previous work, these methods only generate initial populations to reduce part of the evaluation process or only combine with simple evolutionary algorithms [34]. This means that the search process is slower to converge, less capable of global search, and consumes more time and computational resources. Meanwhile, in previous methods, the network is usually evaluated using only a single evaluation metric [32,35], which is not comprehensive enough to assess the network performance. Alternatively, multiple network metrics are used for assessment, but the rankings are directly added to combine the metrics [36], resulting in a situation where the rankings in multiple metrics conflict and the impact of the worse metrics is canceled out by the better metrics. Moreover, ref. [37] have reviewed the progress of zero-cost NAS and highlighted open challenges such as metric instability, search space bias, and poor generalization across datasets.
In order to solve the problems mentioned in the above content, we design a new efficient evolutionary neural network architecture search method without training for automatic neural network architecture search, which improves the search efficiency while ensuring the performance of the search results. Specifically, the method proposed in this paper is improved in terms of both algorithm and evaluation method. In the algorithm, the role of individuals in the population is emphasized, and the interactions among individuals in the population are enhanced, so that the relevant problem information carried by individuals can be further utilized and diffused through competition and cooperation among individuals to promote faster convergence of the algorithm. In the evaluation method, all the training and evaluation processes are omitted, and the performance of the network is evaluated at the initialization stage of the network. In order to improve the accuracy of the evaluation, the evaluation function is improved to assess the network performance more comprehensively from three aspects to reduce the possibility of misjudgment of the fitness function for the network performance, and a new way of combining the indicators is proposed to solve the problem of ranking discrepancy, which overall improves the consistency of the ranking between the fitness function and the actual performance.
The contributions of the proposed algorithm are summarised as follows.
  • In order to improve the search efficiency of the algorithm, the proposed algorithm improves the interaction between the individuals, which drives the whole evolution in the direction of goodness and makes greater use of the problem information carried by each individual. This advancement facilitates faster convergence of the search algorithm.
  • We propose a fitness function based on the combination of multiple zero-cost proxies that evaluates network architecture performance from multiple aspects and combines them. Compared with previous approaches, this paper considers more comprehensive performance influences, combines them in a more effective way, and evaluates them more accurately.
  • Experiments on widely adopted benchmarks demonstrate that this paper’s method outperforms existing NAS methods and is able to find better performing network architectures in a shorter period of time. The effectiveness of this paper’s method is also demonstrated by ablation study.
The remainder of this paper is organized as follows. Section 2 provides an overview of the related work in the field of neural architecture search. In Section 3, we present the methodology employed in this research, providing a detailed description of the proposed method. Section 4 presents the experimental setup, results, and analysis, comparing them with previous studies and discussing their implications. Finally, Section 5 we will summarize the whole work and discuss the future directions for further research.

2. Related Works

In this section, we review related work in the field of neural architecture search, focusing on two key aspects: EC-based NAS methods and accelerated evaluation strategies.

2.1. EC-Based NAS

Due to their high flexibility, simplicity in concept, strong search capabilities, NAS methods based on Evolutionary Computation are widely embraced [38].
In EC-based NAS approach, a set of architectures is updated by continuous iteration to get the final network architecture. Typically, population initialization is performed to get the initial population, one or more parents are selected to generate new individuals through crossover and mutation operators, and well-performing individuals will be added to the population.
The variations among algorithms primarily stem from their approaches to constructing these evolutionary operations. LargeEvo [22] is often regarded as the pioneering EC-based NAS algorithm, employing genetic algorithms for automated search. In [23], aging evolution was introduced, Discarding older individuals during evolution forever keeps the population young and facilitates the exploration of more individuals. Ref. [39] Fusing the two operators while discarding the oldest and worst individuals. Furthermore, algorithms design their own crossover and mutation operators. Refs. [40,41,42] employ single-point crossovers [43], while [44] utilizes disruptive crossovers. Or have multiple crossover operators at the same time and randomly choose the type of crossover when an individual makes a crossover [45]. Mutations are usually performed at two levels [46], modifying the types of basic structures in the architecture and modifying the connectivity relationships. In [47,48], RNNs are trained to guide mutation operations. In addition to using GA as a search algorithm, recent studies have begun to use DE as a search algorithm because the DE algorithm has the advantages of fewer parameters and faster convergence. The standard DE algorithm is used as the search strategy in [49], which proves the superiority of DE applied in NAS. Ref. [50] combines the single-path coding in PSO [51] with DE, while adding the quadratic crossover [50] combines single-path coding in PSO [51] with DE, while adding secondary crossover. The coding strategy divided into two parts is designed in [52], and the local variation [53] applies DE to NAS by combining the individual selection strategy with the environment selection strategy in [54]. In all these works, DE has shown good performance, but the problem-related information carried by individuals in the population has not been further exploited, so it is necessary to further improve the DE algorithm to enhance the performance of the NAS algorithm.

2.2. Methods for Accelerated Evaluation

The development of neural network architecture search methods aims to obtain better performing networks in a shorter time, and a large amount of time in the architecture search process is consumed in the evaluation process [55], so it is particularly important to accelerate the evaluation process.
Early stopping strategy is used in [30], which is different from the previous full training, the training process in the early stopping strategy stops after a few epochs, and the training time is shortened by decreasing the number of epochs for training. Ref. [56] is accelerated by low fidelity estimation, which employs a small number of datasets to train to get the performance estimation. Ref. [57] avoided training of the network by performance predictor and predicted to get the performance of the network, but the performance predictor usually needs training process with data. Ref. [58] through network morphism the children inherit the weights of the parents, thus reducing the small part of training and achieving speedup. Through weight-sharing [20], a large supernet is obtained by training first, which makes the child network inherit weights from the supernet without training. However, all of these acceleration methods still leave the training process more or less intact, so there is still room for development for acceleration.
An ideal approach for accelerating evaluation is to obtain network performance without training, termed as zero-shot or training-free methods. In such methods, there is no need to train the network, and no trained agent is required for evaluation, making it extremely cost-effective. These approaches primarily evaluate networks from two aspects: parameters and structure. Parameter evaluation is mainly used for pruning, assessing the importance of parameters, such as fisher [59] and SynFlow [60]. Architecture-based methods, on the other hand, leverage the theory of Neural Tangents Kernel (NTK) to score networks based on attributes that enhance structural performance, as seen in NASWOT [32] and TE-NAS [33]. However, most zero-cost proxies do not accurately predict neural network performance.
In the process of combining with these metrics, different search strategies are used, for example, pruning is used as a search strategy in ZCNAS [61] and TE-NAS [33], Bayesian optimization is used in BANANA [62], evolutionary algorithms are used in NASWOT to warm-up subsequent searches, REA [23] is used as a search strategy in TEG-NAS [36], and TNASSE [63] uses the search economic algorithm as a search strategy, as stated before, in the process of combining with metrics that do not require training, most of the evolutionary algorithms used are simple evolutionary algorithms or are only used as a warm-up session, so the potential of combining evolutionary algorithms with metrics that do not require training is worth exploring.

3. Approach

The proposed method in this paper aims to search for neural network architectures with high accuracy within a sho-rter time consumption and lower computational consumption. This section explains the proposed method in detail. Specifically, we will begin by introducing the definitions of the environment and the local environment. Subsequently, we present the definition of individuals, exemplifying the encoding method for individuals within the NAS-Bench-201 search space. Following this, we introduce the overall algorithmic flow. Further, we delve into comprehensive explanations of the operations involved in the algorithm. Finally, we introduce the improved evaluation method proposed in this paper.

3.1. Problem Definition

The goal of this paper is to search for a high-performing neural network architecture a A from a predefined search space A , without requiring any network training during the search. Formally, this can be formulated as the following optimization problem:
a * = arg max a A F ( a )
where F ( a ) is the performance estimation function that evaluates architecture a without full training (e.g., using zero-cost proxies), and  A is the discrete architecture search space defined by a specific benchmark, such as NAS-Bench-101 or NAS-Bench-201.
Our objective is to efficiently explore A to find the architecture a * with the highest predicted performance, by designing a novel evolutionary strategy incorporating both individual interactions and multi-source evaluation metrics.

3.2. Environment Definition

In order to enhance the interaction between individuals and further utilize the problem information carried by individuals, we adopt a different approach from traditional evolutionary algorithms. Specifically, we define a separate environment in which individuals interact with their neighbors.
The environment can be conceptualized as a grid-like structure resembling an alternating black and white chessboard (Figure 1). The size of the environment is determined by the integer L s i z e ( L s i z e is specified as an even number), which corresponds to L s i z e × L s i z e . Each individual is positioned at a fixed chessboard cell, and the size of the environment is equal to the number of chessboard cells, which is also equal to the number of individuals. As shown in Figure 1, if  L s i z e is 4, the environment size would be 4 × 4 , and correspondingly, there would be 16 individuals located at 16 chessboard cells.
Individuals located on chessboard cells only interact with their neighbors, and information is diffused through these interactions. In other words, each individual operates within its local environment, which consists of the individual itself and its four neighboring individuals of the environment. As shown in Figure 2, for non-edge individuals (depicted in light green), their neighbors include the four individuals located above, below, to the left, and to the right. When individuals are located at the edges (as depicted in dark yellow in the figure), it can be considered that each row and column forms a cyclic queue. Therefore, for an individual A i , j , its neighbors A i , j n e i g h b o r s = { A i , j , A i , j , A i , j , A i , j }.where
i = i 1 i 1 L s i z e i = 1 j = j 1 j 1 L s i z e j = 1 i = 1 i = L s i z e i + 1 i L s i z e j = 1 j = L s i z e j + 1 j L s i z e
Since each individual can only perceive its local environment, competitive and cooperative behaviors occur solely between an individual and its neighbors. individuals interact with their neighbors to transmit information to them. In this way, information spreads throughout the entire environment. It can be observed that the model of local environment interaction is closer to the real evolutionary mechanisms found in nature than the population model in traditional genetic algorithms.

3.3. Individual Definition

In our approach, each individual represents a potential solution to the NAS problem. The quality of an individual is determined by its fitness function value, where higher values indicate better performance and a higher likelihood of survival in the environment. Thus, individuals strive to maximize their fitness function levels. We will delve into the fitness function details in upcoming sections.
When conducting a search, it is necessary to encode individuals to facilitate the subsequent search process. In NAS-related work, the search space is often divided into macro-skeleton and micro-cell levels. The macro-skeleton remains constant, focusing the encoding on the cell. As an example, we’ll discuss our encoding strategy using NAS-Bench-201 [64].
In NAS-Bench-201, the cell structure consists of four nodes, numbered 0, 1, 2, and 3, as shown in Figure 3. These nodes are connected in a unidirectional manner, resulting in six possible connections. This means the encoding length for a cell is 6.
The cell encoding describes operations between different nodes. Each digit in the encoding corresponds to a specific node pair, as illustrated in Figure 4.
Therefore, the search space on NAS-Bench-201 can be expressed as:
A = { A | A = ( a 1 , , a n ) } , n = 6 , a i [ 0 , 5 ] , i = 1 , 2 , 3 , , n
We employ a real number encoding method, as depicted in Figure 5, which involves mapping real number parameters back to operators during decoding. To maintain diversity and facilitate exploration, each digit in the encoding is a real number within the range of 0 to 5. Using real numbers instead of integers prevents the population from becoming too homogeneous, as integers would lead to numerous duplicates in the population and hinder the effectiveness of Differential Evolution (DE) in exploring the solution space.
Map ( a i ) = a i , a i [ 0 , 5 ]
The result of map indicates the type of operation between two nodes. Specifically, zeroize, skip-connect, 1 × 1 convolution, 3 × 3 convolution, and 3 × 3 average pooling are represented by the values 0, 1, 2, 3, and 4, respectively.
For example, let’s consider the cell shown in Figure 6. The corresponding encoding for this cell is [0.45, 1.35, 2.85, 1.64, 3.21, 4.82], and after the mapping operation of the Formula (3), the result obtained is [0, 1, 2, 1, 3, 4], indicating that the operations between the nodes are as follows: from the node 0 to node 1, a zeroize operation is applied; from the node 0 to node 2, a skip-connection is applied; from the node 1 to the node 2, a 1 × 1 convolution is applied; from node 0 to node 3, a zeroize operation is applied; from node 1 to the node 3, a 3 × 3 convolution is applied; and from node 2 to the node 3, an average pooling operation is applied.
This encoding strategy allows for a concise representation of the cell structure and the operations between nodes, facilitating the exploration and evaluation of different cell architectures in NAS-Bench-201.

3.4. Overall Introduction

In this section, the overall process of an efficient search algorithm proposed for evolutionary neural network architecture search is presented.
In the context of neural network architecture search, each individual represents a distinct network, and the population exists within the environment previously described. We represent the environment in which the individuals reside as L, where the size of the environment is determined by the integer L s i z e , corresponding to L s i z e × L s i z e . This implies that the population size remains constant in each generation, with a fixed size of L s i z e × L s i z e . After completing the search in each generation, a total of L s i z e × L s i z e neural networks are retained. We denote L t as the environment at the t-th generation, B e s t t is the best individual among L 0 , L 1 , L 2 , …, L t , and  C B e s t t is the best individual in L t .
Detailed explanations of the following operational operators will be provided in Section 3.5.
Initialization: The proposed method commences with the random initialization of the environment L 0 . In this process, random sampling within the search space is utilized to select L s i z e × L s i z e neural networks, forming the initial population. Additionally, the individual with the highest fitness function value in the initial population is identified and designated as Best0. The method for computing fitness will be detailed in subsequent sections.
Neighborhood Competition: Subsequently, individuals in the environment interact with each other, facilitating information exchange. We apply the neighborhood competition operator for each individual in the environment L t . This operator makes interaction between the individual and its local environment, leading to the exchange of information. Upon completion of the neighborhood competition, the current environment state is denoted as L t + 1 / 3 .
Global Exploration: After completing the neighborhood competition, we employ the differential operator for global exploration. Within  L t + 1 / 3 , involving mutation and crossover operators. Perform the mutation operator to each individual A i , j , yielding the mutated individual A i , j m u t a n t . Subsequently, perform the crossover operator to generate the trial individual A i , j t r i a l . If  Fitness ( A i , j ) < Fitness ( A i , j t r i a l ) , replace A i , j with A i , j t r i a l . By conducting global exploration, we can uncover superior neural network architectures, advancing the overall quality of the population. Once the process is finished for all individuals within L t + 1 / 3 , it yields the environment L t + 2 / 3 .
Self-Learning: After completing the differential global exploration, we perform self-learning on the best individual C B e s t t + 2 / 3 in the current population L t + 2 / 3 . Self-learning allows the individual to autonomously explore the surrounding area for better individuals. Here, we employ a simple greedy algorithm for self-learning. After completing the self-learning, we obtain a new individual C B e s t t + 1 , and the current environment state is recorded as L t + 1 . In other words, at this point, we have completed one iteration of the population.
The evolution process continues until the termination condition is met. Ultimately, the best network architecture is decoded from the individual B e s t t .

3.5. Evolutionary Operators

Neighborhood competition and neighborhood crossover operators facilitate competition and cooperation, respectively. Mutation and self-learning operators implement knowledge-based behaviors.

3.5.1. Neighborhood Competition Operation

There exists a competitive relationship among the individuals. For an individual A i , j , in its local environment, there are only five individuals: individual A i , j , and its four neighbors A i , j n e i g h b o r s , the individual with the highest fitness value is denoted as M i , j .
M i , j = arg max { Fitness ( A ) A A i , j n e i g h b o r s }
There is a competitive relationship among all of them, like Figure 7 and an individual can survive in its local environment only when it has the maximum fitness value among them:
Fitness ( A i , j ) > Fitness ( M i , j )
If the individual A i , j satisfies (5), it can continue to survive in its current position. When an individual fails to satisfy (5), it will be removed from the environment. This process is similar to the selection operator in swarm intelligence algorithms. However, in swarm intelligence algorithms, the selection operator operates on the entire population, while the competitive operator occurs only among the individuals and their neighbors in this context.
In addition to the competitive relationships, there are also cooperative relationships among the individuals, which allow individual individuals to increase their fitness values. After the competitive operator is executed, if an individual fails to meet the fitness requirement, it will be removed from the environment, creating a corresponding vacancy. Individuals cooperate to generate new individuals, filling the vacant positions. New individuals can be generated through two different strategies. In the first strategy:
A i , j n e w = M a x i , j + Rand ( 0 , 1 ) × ( M a x i , j A i , j )
where A i , j n e w represents the new individual generated to replace A i , j , M a x i , j represents the individual with the highest fitness value in the local environment L i , j where A i , j previously existed, and  r a n d ( 0 , 1 ) denotes a random number drawn from the uniform distribution between 0 and 1.
The second strategy is as follows:
A i , j n e w = M a x i , j + 1 t × G ( 0 , 1 )
where t represents the current generation, and G denotes Gaussian noise. According to the assumption that “Good solutions are often found near good solutions”. In the operator, as the generation increases, the degree of mutation decreases, focusing the search near good solutions.
The two strategies for generating new individuals have different tendencies. Compared to the second strategy, the first one tends to emphasize cooperation between individuals. This is because we believe that among the killed individuals, there may also be factors beneficial to evolution. Thus, the first strategy combines the best individual with the current individual to generate a new individual, moving closer to the best individual in the local environment based on the current individual. The second strategy, on the other hand, leans more towards utilizing the best individual in the local environment to enhance the overall fitness value.
The overall neighborhood competition operator is shown in Algorithm 1.
Algorithm 1 Neighborhood competition operations
Require: 
A i , j R n : individual at position ( i , j ) ;
A i , j neighbors R n : neighbor individuals;
P s [ 0 , 1 ] : selection probability
Ensure: 
A i , j R n : updated individual at position ( i , j )
    1:
M i , j ← None
    2:
for each  A  in  A i , j n e i g h b o r s do
    3:
    if  Fitness ( A ) > Fitness ( M i , j )  then
    4:
         M i , j A
    5:
    end if
    6:
end for
    7:
if  Fitness ( A i , j ) < Fitness ( M i , j )   then
    8:
    if  Rand ( 0 , 1 ) < P s  then
    9:
          N e w I n d i v i d u a l Strategy1( )
  10:
                                                         ▹ Use Strategy 1 to generate new individual
  11:
    else
  12:
          N e w I n d i v i d u a l Strategy2( )
  13:
                                                         ▹ Use Strategy 2 to generate new individual
  14:
    end if
  15:
else
  16:
     N e w I n d i v i d u a l A i , j                                 ▹ Retain the original individual
  17:
end if
  18:
Return  N e w I n d i v i d u a l as the updated individual A i , j

3.5.2. Global Exploration

New offspring are brought into existence via the implementation of mutation operations for each individual, employing a specific mutation strategy. In our works, we traverse through every individual denoted as A i , j residing on the grid. When a randomly generated number U ( 0 , 1 ) falls below a predefined threshold P m , we proceed to initiate a mutation operation on that specific individual. The mutation operation is performed by selecting any two individuals from the environment, excluding the individual A i , j , identified as A k , l and A m . n , ( i , j ) ( k , l ) ( m , n ) . These selected neighbors and individual A i , j itself, collectively form the three parents essential for executing the differential mutation operation. This process is used to generate a new individual as follows:
A i , j m u t a n t = A i , j + F · ( A k , l A m , n )
where A i , j m u t a n t is the mutant individual generated for individual A i , j in the environment, F is the scaling factor (which usually takes values within the range [0, 1]).
After the mutation, a crossover operation is applied to each individual A i , j and its corresponding mutant individual A i , j m u t a n t to generate a trial individual A i , j t r i a l . We use a simple binomial crossover, which chooses the value for each dimension i from A i , j with probability P c and from A i , j m u t a n t otherwise.
a i , j t r i a l = a i , j , rand ( 0 , 1 ) P c a i , j m u t a n t , o t h e r w i s e .

3.5.3. Self-Learning Operation

In addition to interacting with the environment, an individual can self-learn to improve its fitness value. According to [65], in the algorithm, individuals can improve the quality of their offspring through local refinement to achieve lifelong learning. Therefore, in our algorithm, we employ the corresponding local search operator to achieve individual self-learning.
In self-learning, the goal is to improve the ability of future generations by obtaining a better solution in the current environment than the current one, without necessarily reaching the current global optimum. Therefore, in this case, we use a simple greedy algorithm for self-learning.
The overall algorithm flow, as depicted in Algorithm 2, unfolds as follows. After executing the preceding operators, the algorithm identifies the best solution within the current population, which becomes the input for the local search as the current individual. The algorithm then applies the appropriate mutation operator to the current individual, generating the corresponding neighbor individual. Within the neighbor individuals, it seeks out the individual with the highest fitness value. If this individual’s fitness value surpasses that of the current individual, the current individual is replaced with the one with the highest fitness value. This process shares similarities with the competition operator discussed earlier, with the distinction lying in how the retained individuals are generated. Additionally, the mutation operator employed here differs from the one used earlier.
In this context, the mutation operator is applied to neural network architecture design. Specifically, it involves a single-point mutation on the neural network architecture. Compared to other operators, single point mutation retains a higher degree of information about the current individual, which is in line with our aim of using self-learning operators. The mutation operator randomly selects a position within the encoding and replaces it with a random number from the range [0, 5]. This implies that within the network architecture, one operation will be replaced. As depicted in the Figure 8, an original skip connection in the network architecture is replaced with a 1 × 1 average pooling operation.
Algorithm 2 Self-learning operation
Require: 
C B e s t t + 2 / 3 R n : initial solution
Ensure: 
C B e s t t + 1 R n : result solution
    1:
C u r r e n t S o l u t i o n C B e s t t + 2 / 3
    2:
C u r r e n t F i t n e s s Fitness ( C u r r e n t S o l u t i o n )
    3:
while stop condition not meet do
    4:
     N e i g h b o r s GenerateNeighbors ( C u r r e n t S o l u t i o n )
    5:
     B e s t F i t n e s s C u r r e n t F i t n e s s
    6:
    for  N e i g h b o r in N e i g h b o r s  do
    7:
         N e i g h b o r F i t n e s s Fitness ( N e i g h b o r )
    8:
        if  N e i g h b o r F i t n e s s  >  B e s t F i t n e s s  then
    9:
              B e s t N e i g h b o r N e i g h b o r
  10:
            B e s t F i t n e s s N e i g h b o r F i t n e s s
  11:
        end if
  12:
    end for
  13:
    if  B e s t F i t n e s s  >  C u r r e n t F i t n e s s  then
  14:
         C u r r e n t S o l u t i o n B e s t N e i g h b o r
  15:
         C u r r e n t F i t n e s s B e s t F i t n e s s
  16:
    else
  17:
        break
  18:
    end if
  19:
end while
  20:
C B e s t t + 1 C u r r e n t S o l u t i o n
  21:
Return  C B e s t t + 1

3.6. Fitness Function

In evolutionary algorithms, the evaluation of individuals within a population relies on fitness functions, which indicate an individual’s adaptability to a specific environment, reflecting the quality of the represented solution. Correspondingly, each optimization problem may employ a distinct fitness function. In the context of NAS, classification accuracy is conventionally employed as the fitness measure for individuals. However, obtaining classification accuracy for a network typically entails costly network training. The aim of this paper is to discover high-performing neural network architectures without the need for network training during the search process. To address this challenge, we introduce a novel fitness function that eliminates the requirement for training during the search process, thereby better aligning with the aforementioned goals.
According to [33,66], NTK can be used to measure the effectiveness of optimizing a network using gradient descent. however the constant is computationally expensive and time consuming. According to [67,68], the constant NTK can be approximated on the mini-batch by the trace paradigm and the gradient flow, which greatly saves computational time. Therefore, when measuring the trainability of the network, we use the approximated NTK computed on mini-batch to get the performance of the network, being written as:
S T = d l 1 b j = 1 b G r a d A L x j 2 2
where d is the size of the dataset indicating the total number of samples, l is the Lipschitz constant of the loss function used for normalizing the gradient’s contribution, b is the mini-batch size representing the number of samples used per iteration, Grad A L x j is the gradient of the loss function L with respect to the architecture A for the j-th data point x j , and · 2 denotes the L 2 norm representing the Euclidean length of the gradient.
It can be seen that the trace norm of NTK is related to the Frobenius norm of the Jacobian matrix [33], so a higher S T may have more parameters, which means better fitting ability, but the generalization ability may be poor. In contrast, we expect to obtain a neural network that is easy to converge while also having better generalization. Therefore, we expect a metric that can measure the network’s generalizability, that is, the performance that remains good when the network encounters unseen data, and a plain idea is that my network’s generalization is good when it thinks that two different inputs correspond to the same output. Therefore when evaluating the generalization performance of a neural network, we propose the following method:
First, for a given input image x, we generate a new image x noisy by adding Gaussian noise. Inputting these two images separately into the neural network, we focus on and compare their feature maps output at the last pooling layer. Let the outputs of the last pooling layer of the network be F ( x ) and F ( x noisy ) , respectively, and these feature maps can be regarded as the mapping of the function F on the input.
To quantify the differences between the feature maps, we use the Euclidean distance to define the distance between them:
S G = i , j [ F ( x ) ( i , j ) F ( x noisy ) ( i , j ) ] 2
This distance metric can help us to evaluate the difference in the response of the neural network to the addition of Gaussian noise to the input image, thus providing a quantitative analysis of the network’s generalization ability, especially in an untrained initial state.
The last factor in the fitness function is used to describe the network’s expressiveness, which refers to the network’s ability to approximate complex functions [69]. Higher network expressiveness leads to better overall performance. In this paper, we utilize the S E [70] to assess expressiveness. The S E consists of two components. The first part characterizes the network’s expressiveness using Gaussian complexity:
ϕ ( x ) = l o g ( E x , ϵ f ( x ) f ( x + α ϵ ) F )
where x , ϵ N ( 0 , 1 ) , F stands for Frobenius norm, α is an adjustable hyper parameter, e.g., 0.01.
The second part is used for scale-sensitive problems, and the scale-insensitive effect is achieved by scaling the first part by the product of the variance statistics of the BatchNorm layer, So the complete S E is:
S E = ϕ ( x ) + i l o g ( j σ i j 2 m
σ i j is the standard deviation of the ith channel in ith BN layer with m output channels.
The rationale for selecting these three specific components is to create a holistic evaluation metric. S T serves as a proxy for the network’s trainability, indicating how readily the architecture can be optimized. S G is designed to estimate generalization by measuring the network’s robustness to input perturbations, a key quality for performance on unseen data. Finally, S E quantifies the network’s expressiveness, or its capacity to model complex functions. By synergizing these three complementary aspects, our fitness function aims to provide a more accurate and robust ranking of architectures compared to relying on any single zero-cost proxy.
The last question is how to combine the three indicators. In multi-criteria evaluations, we encountered the issue where each criterion S T , S G , and S E correlates differently with the ground truth, causing significant discrepancies in a model’s rankings across different metrics. Simple normalization and summation, or independently ranking and summing each criterion, might lead to extreme rankings offsetting each other, failing to accurately reflect the model’s overall performance. To address this, we chose to process the rankings of the criteria using logarithmic and exponential transformations. This method mitigates the impact of extreme rankings by first applying a logarithmic transformation to the rankings, followed by an exponential reversal, resulting in a more stable and balanced aggregate score.
S ( i ) = M { T , G , E } e log ( Rank ( S M ( i ) ) + ϵ )
where ϵ is a small positive constant to avoid logarithms of zero. This approach effectively provides a balanced composite ranking, accommodating the varying influences of each criterion on the overall assessment.

4. Experiments and Results

To validate the effectiveness of the proposed algorithm, experiments were conducted on two widely adopted benchmarks: NAS-Bench-101 [71] and NAS-Bench-201 [64]. In this section, we will first introduce these two benchmarks, the corresponding datasets, and peer competitors in the field. Then, we will describe the relevant parameters used in the experiments and finally present the experimental results and analysis.

4.1. Benchmark Datasets

NAS-Bench-101 and NAS-Bench-201 are two widely used benchmarks in the field of neural network architecture search. These benchmarks provide researchers with standardized and pre-evaluated architectures to facilitate the comparison and evaluation of different search algorithms.
In both NAS-Bench-101 and NAS-Bench-201, the search space is constructed based on cells, which are fundamental building blocks of neural networks. These cells are formed by stacking repeated modules, resembling the structure of well-designed manually crafted networks. This fixed search space allows researchers to explore and evaluate a diverse range of architectures.
In NAS-Bench-101, the search space consists of 3 stacked cells, forming a neural network. Each cell is composed of 7 nodes, where the first node is fixed as the input node and the last node is fixed as the output node. The maximum number of edges in a cell is limited to 9. Each node in the cell represents an operation, and the available operations include 3 × 3 convolution, 1 × 1 convolution, and 3 × 3 max pooling. The benchmark provides a collection of 423,624 candidate neural networks, and these architectures are trained and evaluated on the CIFAR10 dataset.
On the other hand, NAS-Bench-201 defines a search space with 4 nodes and 6 edges. The edges represent the possible operations between two nodes, and the available options include zeroize, skip-connect, 1 × 1 convolution, 3 × 3 convolution, and 3 × 3 average pooling. With the fixed size of the cell and the predetermined set of operations, there are a total of 5 6 = 15625 candidate cell architectures. The benchmark provides training data and performance evaluations on CIFAR10, CIFAR100, and imagenet16-120 datasets for all these candidate architectures.
While both benchmarks serve the purpose of evaluating neural network architectures, there are some differences between NAS-Bench-101 and NAS-Bench-201. NAS-Bench-101 focuses on training architectures on a single dataset (CIFAR10), but it offers a larger search space compared to NAS-Bench-201. On the other hand, NAS-Bench-201 provides training data and performance evaluations on multiple datasets, allowing researchers to assess the generalization capability of architectures.

4.2. Peer Competitors

To validate the effectiveness of the proposed method, we conducted comparisons with other state-of-the-art techniques in our experiments. As discussed in Section 2, the methods chosen for comparison should cover RL, EA, and gradient-based approaches. Within these categories, most gradient-based methods have achieved acceleration through weight sharing, which is a prominent research direction in NAS. Consequently, the methods selected for comparison are categorized into “weight sharing” and “non-weight sharing”. Additionally, as we employ a training-free evaluation approach in this paper, we also compared our results with training-free methods. Furthermore, we included manually designed neural networks and random search as baseline approaches in the comparison.
Specifically, the first category of non-weight sharing work includes manually designed neural networks (ResNet) [2], random search with parameter sharing (RSPS) [72], regularized evolution for architecture search (REA) [23], and Bayesian optimization with hyperband (BOHB) [73], Assisted Regularized Evolution (AREA) [32].This basically covers all non-weight-sharing search methods mentioned in the related works.
The second category of methods is known as weight-sharing based methods. These methods include various techniques such as differentiable architecture search (DARTs-v1, DARTs-v2) [15], subgraph-based differentiable architecture search (GDAS) [74], neural architecture search via self-evaluated template network (SETN) [75], efficient neural architecture search (ENAS) [76], beta-decay regularization for differentiable architecture search [77] ( β -DARTS). Among these methods, DARTs serves as the baseline for all the mentioned works.
The third category comprises training-free methods that avoid the computational costs associated with network training, allowing a more focused search process. This category includes methods such as Guided Evolutionary Neural Architecture Search (GEA) [78], Neural Architecture Search Without Training (NASWOT) [32], Training-free NAS (TE-NAS) [33], TEG-NAS [36], TNASSE [63], Green Neural Architecture Search (KNAS) [79], and Zero-Cost Proxies for NAS (ZCNAS) [61]. Among them, TE-NAS and ZCNAS use pruning as a search strategy, NASWOT, KNAS use random search as a search strategy. TNASSE, GEA, TEG-NAS, use evolutionary algorithms as a search strategy.

4.3. Parameter Setting

The experiments are implemented using Python 3.8 and PyTorch 1.7.0 on a PC with a single Nvidia RTX 3060 GPU. The hyperparameters (such as learning rate, batch size, etc.) are set according to the configurations provided in [64,71], to eliminate the influence of hyperparameters in the comparative experiment results. The parameters used in the algorithm are kept mostly consistent with the reference [49,80], population size set of 16, and 20 generations. To avoid the algorithm becoming a random search, the differential rate and crossover rate are set to 0.6 and 0.45, designed to foster a healthy balance between exploring new regions of the search space (exploration) and refining existing solutions (exploitation). A higher differential rate encourages the generation of diverse candidate solutions, while the crossover rate ensures that beneficial architectural motifs are preserved and propagated. In the self-learning operator, the Gaussian noise was configured with a mean of 0 and a standard deviation of 1. This standard normal distribution is a typical choice for introducing small, unbiased random perturbations, enabling fine-grained local search around promising architectures. Following the common conventions in evolutionary algorithms. In the self-learning operator, the number of neighbors is 8 and the number of iterations is 5. To avoid any errors in the code execution, the results of the peer competitors in the experiments are taken directly from the relevant published papers.

4.4. Results and Analysis

4.4.1. Overall Results

Because the goal is to be able to search for better performing neural network architectures faster, the focus of the comparison will be on the algorithm’s search time and the network performance obtained from the search (in the case of classification tasks, this is represented as classification accuracy). The search time will be expressed in terms of GPU days/hours/seconds, calculated by multiplying the number of GPUs used for the search by the search time. For example, if the search takes 3 days on two GPUs, the GPU days would be 6; if the search takes 60 s on two GPUs, the GPU seconds would be 120. The actual running time of the algorithm determines whether it is measured in GPU days or GPU seconds, but it should be consistent for comparison purposes.
We conducted four experiments with different random seeds to obtain experimental results and report the final averages and standard deviations. Similar to previous work, our approach for architecture search was performed on CIFAR-10. We obtained the final network encodings, which were then used to evaluate the performance on relevant datasets. In NAS-Bench-201, the relevant datasets included CIFAR-10, CIFAR-100, and ImageNet-16-120. In the table, we also present the test accuracy and validation accuracy for each dataset. Validation accuracy is obtained by training the model on the training dataset and evaluating it on the validation dataset. Test accuracy, on the other hand, is obtained after completing both the training and validation phases, through evaluation on a separate test dataset. Typically, these two accuracy metrics may not align.
Table 1 shows the results of the proposed algorithm in this paper compared with other well-performing competitors on NAS-Bench-201. From top to bottom, the first column consists of five non-weight sharing algorithms. Among them, ResNet is designed manually, while the other three are obtained through neural architecture search algorithms. The difference lies in the search strategies used, namely random search, Bayesian optimization, and evolutionary algorithm. The second column consists of six weight sharing algorithms, where most of them use gradient-based search strategies. The third column consists of seven training-free algorithms, where the difference lies in the zero-shot proxies used and the search strategies employed. Most of the competitors’ metrics are directly taken from the NAS-Bench-201 paper.
Validation accuracy and test accuracy are both provided in the table, but in the subsequent sections, the term “accuracy” as mentioned will follow the convention in related work, referring to test accuracy. Search accuracy and search cost are the metrics of our primary interest.
The precision of validation and testing are provided in the table, but in the following text, we will refer to the accuracy mentioned in related work, which refers to test accuracy. Search precision and search cost are the indicators we care about. We created a scatter plot as Figure 9 using the test accuracy and search cost data from Table 1 on CIFAR-10 to facilitate a clearer comparison of the algorithms. From the plot, we can see that in terms of search cost, our method outperforms all non weight sharing and weight-sharing algorithms, and is better than all training-free algorithms except NASWOT. However, compared with NASWOT, our algorithm achieves better network accuracy, and the standard deviation of the results is smaller, indicating that the algorithm is more stable. In terms of accuracy, the algorithm proposed in this paper also shows excellent performance, only lagging behind DARTs, but the time we need is considerably shorter compared to DARTs. Furthermore, an interesting finding is that among the weight sharing methods compared, almost all of them, except DARTs, have poor accuracy. This may be due to the presence of too many jump connections in the common architecture of these algorithms, while the structure on NAS-Bench-201 was not specifically designed to fulfill this condition. As a result, many incorrect evaluations occurred, leading to performance crashes. We conducted another experiment on NAS-Bench-101 to validate the effectiveness of the proposed method on a larger dataset and compared it with two non-weight sharing, two weight sharing, and three training-free algorithms. The results are shown in Table 2. In terms of accuracy, ENAS achieved the best result, but at the same time, it consumed the most time among these algorithms. All training-free algorithms have an advantage in search cost. Meanwhile, among all training-free algorithms, our algorithm achieved the best accuracy.Another interesting finding is that the standard deviation of the training-free method is larger than that of the method requiring training, both on NAS-bench-201 and NAS-bench-101. We speculate that this is probably due to the fact that there is always a gap between the evaluation results given by the zero-cost proxy used in the training-free method and the true accuracy.
To quantitatively validate the performance differences observed in Table 1, we conducted an extensive experimental evaluation with 10 independent runs (N = 10) for our algorithm and key baseline methods. To statistically compare the performance, we employed the Wilcoxon signed-rank test, a powerful non-parametric method ideal for the typical data distributions found in NAS experiments. The analysis was performed on the paired test accuracy results from the CIFAR-10 dataset, using a two-tailed test with a standard significance level of α = 0.05. The results of our statistical analysis are summarized in Table 3.
The analysis revealed several key findings. First, the proposed method showed statistically significant and substantial improvements over the NASWOT and TEG-NAS baseline methods. Since the p-value was well below the 0.002 threshold and accompanied by a large effect size (r > 0.7), this proves that the proposed method is not a random coincidence.
Secondly, comparison with REA shows that there is no statistically significant difference between the two (p > 0.3). This result indicates that our method is statistically comparable to the current leading method in terms of final model accuracy. The moderate effect size indicates that, despite some minor differences, the overall performance capabilities are comparable. However, in terms of time cost, the method proposed in this paper is far superior.

4.4.2. Abalation Study

In this section, we will delve into the details of each component of the proposed algorithm through ablation experiments, aiming to gain a deeper understanding of their specific impact on the final results. First we further validate the effectiveness of the proposed method on two benchmarks, NAS-Bench-201 and NAS-Bench-101, and then conduct experiments on NAS-Bench-201 to prove the validity of the components in the adopted evaluation approach.
First we compare the method proposed in this paper with the method that retains only the global search operator on two benchmarks. The specific results are displayed in the last two rows of Table 1 and Table 2, and ours- denotes the result of keeping only the global search operator. From the results, it can be seen that the search speed is faster when only the global search operator is retained, and also the results of the search are comparable with the other algorithms in the table, which indicates that the evaluation method we used and its combination are effective. Compared with the method proposed in this paper, the average accuracy has decreased, while the standard deviation has increased but not significantly, which may be due to the smaller dimension of NAS-bench-201, the same experiment is conducted on NAS-bench-101 with a larger dimension, and it can be seen that there is a significant difference in the standard deviation, and at the same time the average accuracy is lower than that of the original algorithm, and it is reasonable to judge that our proposed algorithm is more stable while improving the final result of the algorithm.
The next thing we want to further verify is that there is about the validity of the parts of the fitness function with the combination method. First we use the algorithm proposed in this paper, but replace the fitness function with S T , S G and S E respectively to conduct experiments on the CIFAR10 dataset of NAS-Bench-201, and the experimental results are shown in Table 4.
Table 4 shows the specific results of the ablation experiments. As can be seen from the table, the best results, both in terms of mean precision and standard deviation, were achieved using a combination of three metrics compared to one metric alone. However, there is also some competition in the results obtained when one metric is used alone, which shows that there is some correlation between all three metrics involved and the groundtruth, as well as proving the effectiveness of the algorithm used.
To further validate the validity and reliability of the proposed fitness function as well as the combination approach, we randomly selected 200 network architectures from NAS-Bench-201. To prevent the results from being influenced by chance, we conducted 50 rounds of random sampling. Specifically, we used Kendall’s Tau correlation coefficients to measure the rank correlation between predicted and ground truth values.Kendall’s Tau compares the order of each pair of data points between the two variables and checks that their orders agree. If the order of the two variables is consistent, it indicates a high degree of correlation. Conversely, if the order is different, it indicates a lower correlation. In this study, higher Kendall’s Tau values indicate a higher degree of similarity between the rankings based on the fitness function and the actual classification accuracy rankings. Meanwhile, we compare the combination using the combination proposed in this paper and the combination by adding the ranks directly, and give the correlation Kendall’s Tau in Figure 10.
The results are shown in Figure 10. The horizontal axis represents the number of experiments, 50 in total, and the vertical axis represents the Kendall’s Tau correlation coefficients.The blue data corresponds to the combination of the rank values directly added together, while the yellow data corresponds to the combination proposed in this paper. As can be seen from the figure, Kendall’s Tau remains relatively stable in all 50 experiments, and the Kendall’s correlation coefficient corresponding to the combination method adopted in this paper is 0.66. This noteworthy consistency indicates the robustness and reliability of the value function. Furthermore, in the proposed algorithm, the agent’s evolution towards a better solution partially depends on the current optimal solution. This means that the agent considers the current optimal solution when trying to improve itself. By doing so, the agent can utilize the knowledge and insight gained from the current optimal solution to guide its search for better solutions. Therefore, we also tested the top 200 networks in terms of accuracy in the benchmark and the results show that Kendall’s Tau has a correlation coefficient of 0.68. This indicates that the value function used is reliable. The Kendall’s coefficient is higher for the combination approach used in this paper compared to the combination approach of adding the ranks directly, which further proves the validity of the combination approach used in this paper.
To further understand the dynamic behavior of the proposed algorithm, we analyze the convergence trend during the search process on the NAS-Bench-201 benchmark using the CIFAR-10 dataset.
As illustrated in Figure 11, the validation accuracy of the best-found architecture improves steadily as the number of generations increases. Specifically, the initial population at generation 0 achieves an average validation accuracy of approximately 84.1%. In the early stages, the performance improves rapidly, reaching around 89.3% by generation 10. This rapid ascent is attributed to the neighborhood competition mechanism, which effectively guides the search towards promising regions of the search space by leveraging local peer interactions.
In the later stages, the improvement becomes more gradual, with the final best architecture achieving a validation accuracy of 91.53% and a test accuracy of 94.49%. This stable convergence is facilitated by the self-learning mechanism, which enables local refinement around the current best architecture using stochastic neighbor-based exploration.
Overall, the convergence curve exhibits a typical evolutionary optimization pattern: an initial rapid gain followed by diminishing improvements as the search approaches a local optimum. The smooth and monotonic improvement observed validates the effectiveness and stability of the proposed hybrid strategy combining both global exploration and local exploitation.

5. Conclusions

In this research, we introduce an efficient search algorithm for evolutionary neural network architectures while employing a score function as a zero-cost proxy for network evaluation. This not only enhanced the algorithm’s search capabilities but also reduced the evaluation process overhead, thus improving search efficiency. Experimental results indicate that our proposed method significantly reduces time consumption compared to other algorithms, while maintaining search performance, thanks to the designed algorithm and the employed score function. In future work, we aim to further enhance the algorithm by improving the encoding methods to better integrate with local search and enhance the accuracy and applicability of the score function in various scenarios.
While our experiments on NAS-Bench-101 and NAS-Bench-201 provide a controlled and reproducible validation, it is worth briefly discussing the broader applicability of our method. The proposed framework is designed for general-purpose use. Its core components—the evolutionary algorithm and the training-free fitness function—can be directly applied to a new, real-world task by simply defining a domain-specific search space and providing a small sample of the target data. This makes our approach particularly valuable for industrial or scientific applications with massive datasets, where the cost of traditional training-based NAS is a major barrier. Our method effectively democratizes NAS by making it computationally affordable and highly efficient.

Author Contributions

Conceptualization, Y.A. and C.Z.; methodology, Y.A.; software, Y.A., J.S. and Y.Y.; validation, Y.A., J.S. and B.S.; formal analysis, Y.A.; investigation, Y.A. and C.Z.; writing, Y.A. and J.S.; writing—review and editing, C.Z.; supervision, C.Z.; funding acquisition, C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Key Project of National Natural Science Foundation of China (U1908212), the Local Science and Technology Development Fund Project Under the Guidance of the Central Government of China (1653137155953), the Taking Lead Science and Technology Research Project of Liaoning (2021jh1/10400006) and National Natural Science Foundation of China (62302064).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the datasets used in the paper are publicly available.

Conflicts of Interest

Author Jintao Shao were employed by the company Lightspeed Studios, Tencent Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
NASNeural Architecture Search
RLReinforcement Learning
EAEvolutionary Algorithm
BOBayesian Optimization

References

  1. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 1–74. [Google Scholar] [CrossRef]
  2. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef]
  3. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  4. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  5. Elngar, A.A.; Arafa, M.; Fathy, A.; Moustafa, B.; Mahmoud, O.; Shaban, M.; Fawzy, N. Image classification based on CNN: A survey. J. Cybersecur. Inf. Manag. 2021, 6, 18–50. [Google Scholar] [CrossRef]
  6. Haque, K.F.; Haque, F.F.; Gandy, L.; Abdelgawad, A. Automatic detection of COVID-19 from chest X-ray images with convolutional neural networks. In Proceedings of the 2020 International Conference on Computing, Electronics & Communications Engineering (iCCECE), Southend, UK, 17–18 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 125–130. [Google Scholar]
  7. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  8. Hutter, F.; Kotthoff, L.; Vanschoren, J. Automated Machine Learning: Methods, Systems, Challenges; Springer Nature: Cham, Switzerland, 2019. [Google Scholar]
  9. Ren, P.; Xiao, Y.; Chang, X.; Huang, P.Y.; Li, Z.; Chen, X.; Wang, X. A comprehensive survey of neural architecture search: Challenges and solutions. ACM Comput. Surv. (CSUR) 2021, 54, 1–34. [Google Scholar] [CrossRef]
  10. Elsken, T.; Metzen, J.H.; Hutter, F. Neural architecture search: A survey. J. Mach. Learn. Res. 2019, 20, 1997–2017. [Google Scholar]
  11. Xu, S.; Quan, H. ECT-NAS: Searching Efficient CNN-Transformers Architecture for Medical Image Segmentation. In Proceedings of the 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA, 9–12 December 2021. [Google Scholar] [CrossRef]
  12. Liu, S.; Zheng, C.; Lu, K.; Gao, S.; Wang, N.; Wang, B.; Zhang, D.; Zhang, X.; Xu, T. EVSRNet: Efficient Video Super-Resolution with Neural Architecture Search. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Nashville, TN, USA, 20–25 June 2021. [Google Scholar] [CrossRef]
  13. Li, S.; Li, W.; Wen, S.; Shi, K.; Yang, Y.; Zhou, P.; Huang, T. Auto-FERNet: A Facial Expression Recognition Network With Architecture Search. IEEE Trans. Netw. Sci. Eng. 2021, 8, 2213–2222. [Google Scholar] [CrossRef]
  14. Zoph, B.; Le, Q.V. Neural architecture search with reinforcement learning. arXiv 2016, arXiv:1611.01578. [Google Scholar]
  15. Liu, H.; Simonyan, K.; Yang, Y. DARTS: Differentiable Architecture Search. arXiv 2018, arXiv:1806.09055. [Google Scholar]
  16. Liu, Y.; Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G.; Tan, K.C. A survey on evolutionary neural architecture search. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 550–570. [Google Scholar] [CrossRef]
  17. Zhong, Z.; Yan, J.; Wu, W.; Shao, J.; Liu, C.L. Practical Block-wise Neural Network Architecture Generation. arXiv 2018, arXiv:1708.05552. [Google Scholar] [CrossRef]
  18. Li, Y.; Dong, M.; Xu, Y.; Wang, Y.; Xu, C. Neural architecture tuning with policy adaptation. Neurocomputing 2022, 485, 196–204. [Google Scholar] [CrossRef]
  19. Lopes, V.; Carlucci, F.M.; Esperança, P.M.; Singh, M.; Gabillon, V.; Yang, A.; Xu, H.; Chen, Z.; Wang, J. MANAS: Multi-agent neural architecture search. arXiv 2019, arXiv:1909.01051. [Google Scholar] [CrossRef]
  20. Xie, L.; Chen, X.; Bi, K.; Wei, L.; Xu, Y.; Wang, L.; Chen, Z.; Xiao, A.; Chang, J.; Zhang, X.; et al. Weight-Sharing Neural Architecture Search: A Battle to Shrink the Optimization Gap. ACM Comput. Surv. 2022, 54, 1–37. [Google Scholar] [CrossRef]
  21. Yang, A.; Esperança, P.; Carlucci, F. NAS evaluation is frustratingly hard. arXiv 2019, arXiv:1912.12522. [Google Scholar]
  22. Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.; Tan, J.; Le, Q.; Kurakin, A. Large-Scale Evolution of Image Classifiers. arXiv 2017, arXiv:1703.01041. [Google Scholar]
  23. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q. Regularized Evolution for Image Classifier Architecture Search. arXiv 2018, arXiv:1802.01548. [Google Scholar] [CrossRef]
  24. Huang, J.; Xue, B.; Sun, Y.; Zhang, M.; Yen, G.G. Particle swarm optimization for compact neural architecture search for image classification. IEEE Trans. Evol. Comput. 2022, 27, 1298–1312. [Google Scholar] [CrossRef]
  25. Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G.; Lv, J. Automatically Designing CNN Architectures Using the Genetic Algorithm for Image Classification. IEEE Trans. Cybern. 2020, 50, 3840–3854. [Google Scholar] [CrossRef]
  26. Xue, Y.; Zha, J.; Pelusi, D.; Chen, P.; Luo, T.; Zhen, L.; Wang, Y.; Wahib, M. Neural Architecture Search with Progressive Evaluation and Sub-Population Preservation. IEEE Trans. Evol. Comput. 2024. [Google Scholar] [CrossRef]
  27. Yuan, G.; Xue, B.; Zhang, M. An evolutionary neural architecture search method based on performance prediction and weight inheritance. Inf. Sci. 2024, 667, 120466. [Google Scholar] [CrossRef]
  28. Xue, Y.; Zha, J.; Wahib, M.; Ouyang, T.; Wang, X. Neural architecture search via similarity adaptive guidance. Appl. Soft Comput. 2024, 162, 111821. [Google Scholar] [CrossRef]
  29. Yang, Z.; Wang, Y.; Chen, X.; Shi, B.; Xu, C.; Xu, C.; Tian, Q.; Xu, C. CARS: Continuous Evolution for Efficient Neural Architecture Search. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar] [CrossRef]
  30. Suganuma, M.; Kobayashi, M.; Shirakawa, S.; Nagao, T. Evolution of Deep Convolutional Neural Networks Using Cartesian Genetic Programming. Evol. Comput. 2020, 28, 141–163. [Google Scholar] [CrossRef] [PubMed]
  31. Liu, P.; El Basha, M.D.; Li, Y.; Xiao, Y.; Sanelli, P.C.; Fang, R. Deep Evolutionary Networks with Expedited Genetic Algorithms for Medical Image Denoising. Med. Image Anal. 2019, 54, 306–315. [Google Scholar] [CrossRef]
  32. Mellor, J.; Turner, J.; Storkey, A.; Crowley, E. Neural Architecture Search without Training. arXiv 2020, arXiv:2006.04647. [Google Scholar]
  33. Chen, W.; Gong, X.; Wang, Z. Neural Architecture Search on ImageNet in Four GPU Hours: A Theoretically Inspired Perspective. arXiv 2012. [Google Scholar]
  34. Wu, M.T.; Lin, H.I.; Tsai, C.W. A Training-free Genetic Neural Architecture Search. In Proceedings of the 2021 ACM International Conference on Intelligent Computing and Its Emerging Applications, Jinan, China, 28–29 December 2021. [Google Scholar] [CrossRef]
  35. Yang, J.; Liu, Y.; Wang, W.; Wu, H.; Chen, Z.; Ma, X. PATNAS: A Path-Based Training-Free Neural Architecture Search. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 47, 1484–1500. [Google Scholar] [CrossRef]
  36. Chen, W.; Gong, X.; Wu, J.; Wei, Y.; Shi, H.; Yan, Z.; Yang, Y.; Wang, Z. Understanding and accelerating neural architecture search with training-free and theory-grounded metrics. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 46, 749–763. [Google Scholar] [CrossRef]
  37. Li, G.; Hoang, D.; Bhardwaj, K.; Lin, M.; Wang, Z.; Marculescu, R. Zero-shot neural architecture search: Challenges, solutions, and opportunities. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 7618–7635. [Google Scholar] [CrossRef]
  38. White, C.; Safari, M.; Sukthanker, R.; Ru, B.; Elsken, T.; Zela, A.; Dey, D.; Hutter, F. Neural architecture search: Insights from 1000 papers. arXiv 2023, arXiv:2301.08727. [Google Scholar]
  39. Zhu, H.; An, Z.; Yang, C.; Xu, K.; Zhao, E.; Xu, Y. EENA: Efficient Evolution of Neural Architecture. arXiv 2019, arXiv:1905.07320. [Google Scholar]
  40. Singh, A.; Saha, S.; Sarkhel, R.; Kundu, M.; Nasipuri, M.; Das, N. A Genetic Algorithm based Kernel-size Selection Approach for a Multi-column Convolutional Neural Network. arXiv 2019, arXiv:1912.12405. [Google Scholar]
  41. Hu, Y.; Sun, S.; Li, J.; Wang, X.; Gu, Q. A novel channel pruning method for deep neural network compression. arXiv 2018, arXiv:1805.11394. [Google Scholar]
  42. Ahmed, A.A.; Darwish, S.M.S.; El-Sherbiny, M.M. A novel automatic CNN architecture design approach based on genetic algorithm. In Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2019, Cairo, Egypt, 26–28 October 2019; Springer: Cham, Switzerland, 2020; pp. 473–482. [Google Scholar]
  43. Deb, K. An introduction to genetic algorithms. Sadhana 1999, 24, 293–315. [Google Scholar] [CrossRef]
  44. Sapra, D.; Pimentel, A.D. An evolutionary optimization algorithm for gradually saturating objective functions. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference, Cancún, Mexico, 8–12 July 2020. [Google Scholar] [CrossRef]
  45. Wen, L.; Gao, L.; Li, X.; Li, H. A new genetic algorithm based evolutionary neural architecture search for image classification. Swarm Evol. Comput. 2022, 75, 101191. [Google Scholar] [CrossRef]
  46. Peng, Y.; Song, A.; Ciesielski, V.; Fayek, H.; Chang, X. Fast Evolutionary Neural Architecture Search by Contrastive Predictor with Linear Regions. In Proceedings of the Genetic and Evolutionary Computation Conference, Lisbon, Portugal, 15–19 July 2023; pp. 1257–1266. [Google Scholar]
  47. Maziarz, K.; Khorlin, A.; Laroussilhe, Q.; Gesmundo, A. Evolutionary-Neural Hybrid Agents for Architecture Search. arXiv 2018, arXiv:1811.09828. [Google Scholar]
  48. Chu, X.; Zhang, B.; Ma, H.; Xu, R.; Li, Q. Fast, Accurate and Lightweight Super-Resolution with Neural Architecture Search. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021. [Google Scholar] [CrossRef]
  49. Awad, N.; Mallik, N.; Hutter, F. Differential evolution for neural architecture search. arXiv 2020, arXiv:2012.06400. [Google Scholar]
  50. Wang, B.; Sun, Y.; Xue, B.; Zhang, M. A hybrid differential evolution approach to designing deep convolutional neural networks for image classification. In Proceedings of the AI 2018: Advances in Artificial Intelligence: 31st Australasian Joint Conference, Wellington, New Zealand, 11–14 December 2018; Proceedings 31. Springer: Cham, Switzerland, 2018; pp. 237–250. [Google Scholar]
  51. Wang, B.; Sun, Y.; Xue, B.; Zhang, M. Evolving deep convolutional neural networks by variable-length particle swarm optimization for image classification. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–8. [Google Scholar]
  52. Huang, J.; Xue, B.; Sun, Y.; Zhang, M. EDE-NAS: An Eclectic Differential Evolution Approach to Single-Path Neural Architecture Search. In Proceedings of the Australasian Joint Conference on Artificial Intelligence, Perth, WA, Australia, 5–8 December 2022; Springer: Cham, Switzerland, 2022; pp. 116–130. [Google Scholar]
  53. Rampavan, M.; Ijjina, E.P. Brake light detection of vehicles using differential evolution based neural architecture search. Appl. Soft Comput. 2023, 147, 110839. [Google Scholar] [CrossRef]
  54. Shang, R.; Zhu, S.; Ren, J.; Liu, H.; Jiao, L. Evolutionary neural architecture search based on evaluation correction and functional units. Knowl.-Based Syst. 2022, 251, 109206. [Google Scholar] [CrossRef]
  55. Zhou, X.; Qin, A.K.; Gong, M.; Tan, K.C. A survey on evolutionary construction of deep neural networks. IEEE Trans. Evol. Comput. 2021, 25, 894–912. [Google Scholar] [CrossRef]
  56. Na, B.; Mok, J.; Choe, H.; Yoon, S. Accelerating neural architecture search via proxy data. arXiv 2021, arXiv:2106.04784. [Google Scholar]
  57. Baker, B.; Gupta, O.; Raskar, R.; Naik, N. Accelerating Neural Architecture Search using Performance Prediction. In Proceedings of the International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  58. An, Y.; Zhang, C.; Zheng, X. Knowledge reconstruction assisted evolutionary algorithm for neural network architecture search. Knowl.-Based Syst. 2023, 264, 110341. [Google Scholar] [CrossRef]
  59. Theis, L.; Korshunova, I.; Tejani, A.; Huszar, F. Faster gaze prediction with dense networks and Fisher pruning. arXiv 2018, arXiv:1801.05787. [Google Scholar]
  60. Tanaka, H.; Kunin, D.; Yamins, D.; Ganguli, S. Pruning neural networks without any data by iteratively conserving synaptic flow. arXiv 2020, arXiv:2006.05467. [Google Scholar]
  61. Abdelfattah, M.S.; Mehrotra, A.; Dudziak, Ł.; Lane, N.D. Zero-cost proxies for lightweight nas. arXiv 2021, arXiv:2101.08134. [Google Scholar]
  62. White, C.; Neiswanger, W.; Savani, Y. Bananas: Bayesian optimization with neural architectures for neural architecture search. Proc. AAAI Conf. Artif. Intell. 2021, 35, 10293–10301. [Google Scholar] [CrossRef]
  63. Wu, M.T.; Lin, H.I.; Tsai, C.W. A training-free neural architecture search algorithm based on search economics. IEEE Trans. Evol. Comput. 2023, 28, 445–459. [Google Scholar] [CrossRef]
  64. Dong, X.; Yang, Y. Nas-bench-201: Extending the scope of reproducible neural architecture search. arXiv 2020, arXiv:2001.00326. [Google Scholar]
  65. Neri, F.; Cotta, C. Memetic algorithms and memetic computing optimization: A literature review. Swarm Evol. Comput. 2012, 2, 1–14. [Google Scholar] [CrossRef]
  66. Shin, Y.; Karniadakis, G. Trainability of ReLU Networks and Data-Dependent Initialization. J. Mach. Learn. Model. Comput. 2020, 1, 39–74. [Google Scholar] [CrossRef]
  67. Shu, Y.; Cai, S.; Dai, Z.; Ooi, B.C.; Low, B.K.H. NASI: Label-and data-agnostic neural architecture search at initialization. arXiv 2021, arXiv:2109.00817. [Google Scholar]
  68. Wang, C.; Zhang, G.; Grosse, R. Picking winning tickets before training by preserving gradient flow. arXiv 2020, arXiv:2002.07376. [Google Scholar]
  69. Xiong, H.; Huang, L.; Yu, M.; Liu, L.; Zhu, F.; Shao, L. On the number of linear regions of convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 13–18 July 2020; pp. 10514–10523. [Google Scholar]
  70. Lin, M.; Wang, P.; Sun, Z.; Chen, H.; Sun, X.; Qian, Q.; Li, H.; Jin, R. Zen-nas: A zero-shot nas for high-performance image recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 347–356. [Google Scholar]
  71. Ying, C.; Klein, A.; Christiansen, E.; Real, E.; Murphy, K.; Hutter, F. Nas-bench-101: Towards reproducible neural architecture search. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 7105–7114. [Google Scholar]
  72. Li, L.; Talwalkar, A. Random search and reproducibility for neural architecture search. In Proceedings of the Uncertainty in Artificial Intelligence, PMLR, Virtual, 3–6 August 2020; pp. 367–377. [Google Scholar]
  73. Falkner, S.; Klein, A.; Hutter, F. BOHB: Robust and efficient hyperparameter optimization at scale. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 1437–1446. [Google Scholar]
  74. Dong, X.; Yang, Y. Searching for a robust neural architecture in four gpu hours. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1761–1770. [Google Scholar]
  75. Dong, X.; Yang, Y. One-shot neural architecture search via self-evaluated template network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3681–3690. [Google Scholar]
  76. Pham, H.; Guan, M.; Zoph, B.; Le, Q.; Dean, J. Efficient neural architecture search via parameters sharing. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 4095–4104. [Google Scholar]
  77. Ye, P.; Li, B.; Li, Y.; Chen, T.; Fan, J.; Ouyang, W. b-darts: Beta-decay regularization for differentiable architecture search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10874–10883. [Google Scholar]
  78. Lopes, V.; Santos, M.; Degardin, B.; Alexandre, L.A. Guided Evolutionary Neural Architecture Search With Efficient Performance Estimation. arXiv 2022, arXiv:2208.06475. [Google Scholar] [CrossRef]
  79. Xu, J.; Zhao, L.; Lin, J.; Gao, R.; Sun, X.; Yang, H. KNAS: Green neural architecture search. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 11613–11625. [Google Scholar]
  80. Elsken, T.; Metzen, J.; Hutter, F. Simple and efficient architecture search for Convolutional Neural Networks. In Proceedings of the International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
Figure 1. Diagram of environment for L s i z e = 4 .
Figure 1. Diagram of environment for L s i z e = 4 .
Biomimetics 10 00421 g001
Figure 2. Diagram of local environment. (left): individual is not on the edge. (right): individual is on the edge.
Figure 2. Diagram of local environment. (left): individual is not on the edge. (right): individual is on the edge.
Biomimetics 10 00421 g002
Figure 3. Architecture of NAS-Bench-201.
Figure 3. Architecture of NAS-Bench-201.
Biomimetics 10 00421 g003
Figure 4. Encoding of NAS-Bench-201.
Figure 4. Encoding of NAS-Bench-201.
Biomimetics 10 00421 g004
Figure 5. Mapping of real numbers to operators.
Figure 5. Mapping of real numbers to operators.
Biomimetics 10 00421 g005
Figure 6. Example of encoding.
Figure 6. Example of encoding.
Biomimetics 10 00421 g006
Figure 7. Neighborhoodcompetition operator.
Figure 7. Neighborhoodcompetition operator.
Biomimetics 10 00421 g007
Figure 8. Mutate of self-learning operation.
Figure 8. Mutate of self-learning operation.
Biomimetics 10 00421 g008
Figure 9. Test accuracy VS search cost in NAS-Bench-201.
Figure 9. Test accuracy VS search cost in NAS-Bench-201.
Biomimetics 10 00421 g009
Figure 10. Kendall coefficients corresponding to 200 randomly sampled architectures with 50 experiments on NAS-bench-201, with the methodology of this paper in yellow and the rankings directly summed in blue.
Figure 10. Kendall coefficients corresponding to 200 randomly sampled architectures with 50 experiments on NAS-bench-201, with the methodology of this paper in yellow and the rankings directly summed in blue.
Biomimetics 10 00421 g010
Figure 11. Convergence cure on CIFAR-10.
Figure 11. Convergence cure on CIFAR-10.
Biomimetics 10 00421 g011
Table 1. Experimental results compared on NAS-Bench-201.
Table 1. Experimental results compared on NAS-Bench-201.
AlgorithmCIFAR10CIFAR100ImageNet16CostType
ValidTestValidTestValidTestGPUsec
ResNet90.8393.9770.4270.8644.5343.63-manual
RSPS84.16 ± 1.6987.66 ± 1.6959.00 ± 4.6058.33 ± 4.3431.56 ± 3.2831.14 ± 3.888007Random search
BOHB90.82 ± 0.5393.61 ± 0.5270.74 ± 1.2970.85 ± 1.2844.26 ± 1.3644.42 ± 1.4912,000BO
REA91.22 ± 0.2593.97 ± 0.3172.36 ± 1.0772.14 ± 0.8645.09 ± 0.9245.55 ± 1.0212,000EA
AREA91.20 ± 0.2793.95 ± 0.3971.95 ± 0.9971.92 ± 1.2945.70 ± 1.0545.40 ± 1.1412,000EA
ENAS37.51 ± 3.1953.89 ± 0.5813.37 ± 2.3513.96 ± 2.3315.06 ± 1.9514.84 ± 2.1014,059weight sharing
DARTs-v139.77 ± 0.0054.30 ± 0.0015.03 ± 0.0015.61 ± 0.0016.43 ± 0.0016.32 ± 0.0010,890weight sharing
DARTs-v239.77 ± 0.0054.30 ± 0.0015.03 ± 0.0015.61 ± 0.0016.43 ± 0.0016.32 ± 0.0029,901weight sharing
GDAS90.00 ± 0.2193.51 ± 0.1371.14 ± 0.2770.61 ± 0.2641.70 ± 1.2641.84 ± 0.9028,926weight sharing
SETN82.25 ± 5.1786.19 ± 4.6356.86 ± 7.5956.87 ± 7.7732.54 ± 3.6331.9 ± 4.0731,010weight sharing
β -DARTS91.55 ± 0.0094.36 ± 0.0073.49 ± 0.0073.51 ± 0.0046.37 ± 0.0046.34 ± 0.0011,520weight sharing
GEA91.26 ± 0.2093.99 ± 0.2372.62 ± 0.7772.36 ± 0.6645.97 ± 0.7246.04 ± 0.6726,911training-free
KNAS89.63 ± 1.7889.92 ± 1.8962.60 ± 3.6962.73 ± 3.8429.90 ± 8.1929.76 ± 8.348010training-free
ZCNAS90.21 ± 0.8993.43 ± 0.6470.71 ± 1.9870.94 ± 1.9342.09 ± 3.9842.44 ± 4.0612,915training-free
TNASSE89.99 ± 0.3293.38 ± 0.3570.39 ± 0.8170.47 ± 0.7642.95 ± 1.6643.44 ± 1.71549training-free
NASWOT89.69 ± 0.7392.96 ± 0.8169.86 ± 1.2169.98 ± 1.2243.95 ± 2.0544.44 ± 2.10306training-free
TE-NAS-93.21 ± 0.68-70.21 ± 1.16-43.93 ± 2.141558training-free
TEG-NAS-91.00 ± 0.33-70.10 ± 1.47-44.45 ± 0.759939.5training-free
Ours91.53 ± 0.2494.27 ± 0.2473.02 ± 0.8273.19 ± 0.5946.36 ± 0.7346.70 ± 0.52510training-free
Ours-90.83 ± 0.3193.21 ± 0.3871.29 ± 0.8971.76 ± 0.7844.58 ± 0.8044.31 ± 0.84472training-free
Table 2. Experimental results compared on NAS-Bench-101.
Table 2. Experimental results compared on NAS-Bench-101.
AlgorithmCIFAR10Cost
ValidTestGPUsec
ResNet90.8393.97-
AREA93.67 ± 0.4893.02 ± 0.4812,000
REA93.63 ± 0.5093.02 ± 0.5212,000
DARTs-V183.50 ± 0.1182.74 ± 0.0320,483
ENAS93.83 ± 0.2893.34 ± 0.2622,325
NASWOT92.47 ± 5.2591.95 ± 5.18140
ZCNAS91.72 ± 1.8091.29 ± 1.8218,940
TEG-NAS-92.52 ± 1.302808
Ours93.69 ± 0.5793.01 ± 0.52455
Ours-92.27 ± 1.1791.89 ± 1.13401
Table 3. Statistical analysis of performance differences.
Table 3. Statistical analysis of performance differences.
Comparisonp-ValueEffect Size (r)Significance ( α = 0.05)
Ours vs. NASWOT<0.0020.82Yes
Ours vs. TEG-NAS<0.0020.71Yes
Ours vs. REA>0.30.58No
Table 4. Ablation study on NAS-Bench-201 CIFAR10 with different indicators.
Table 4. Ablation study on NAS-Bench-201 CIFAR10 with different indicators.
S T S G S E S T + S G + S E
93.52 ± 0.3192.73 ± 0.5493.02 ± 0.4194.27 ± 0.24
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

An, Y.; Zhang, C.; Shao, J.; Yan, Y.; Sun, B. An Efficient Evolutionary Neural Architecture Search Algorithm Without Training. Biomimetics 2025, 10, 421. https://doi.org/10.3390/biomimetics10070421

AMA Style

An Y, Zhang C, Shao J, Yan Y, Sun B. An Efficient Evolutionary Neural Architecture Search Algorithm Without Training. Biomimetics. 2025; 10(7):421. https://doi.org/10.3390/biomimetics10070421

Chicago/Turabian Style

An, Yang, Changsheng Zhang, Jintao Shao, Yuxiao Yan, and Baiqing Sun. 2025. "An Efficient Evolutionary Neural Architecture Search Algorithm Without Training" Biomimetics 10, no. 7: 421. https://doi.org/10.3390/biomimetics10070421

APA Style

An, Y., Zhang, C., Shao, J., Yan, Y., & Sun, B. (2025). An Efficient Evolutionary Neural Architecture Search Algorithm Without Training. Biomimetics, 10(7), 421. https://doi.org/10.3390/biomimetics10070421

Article Metrics

Back to TopTop