Next Article in Journal
An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection
Previous Article in Journal
Enhanced Whale Optimization Algorithm with Novel Strategies for 3D TSP Problem
Previous Article in Special Issue
EAB-BES: A Global Optimization Approach for Efficient UAV Path Planning in High-Density Urban Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Value Patents Recognition with Random Forest and Enhanced Fire Hawk Optimization Algorithm

1
Key Laboratory of Ecological Safety and Sustainable Development in Arid Lands, Northwest Institute of Eco-Environment and Resources, Chinese Academy of Sciences, Lanzhou 730000, China
2
Key Laboratory of Knowledge Computing and Intelligent Decision, Lanzhou 730000, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(9), 561; https://doi.org/10.3390/biomimetics10090561 (registering DOI)
Submission received: 18 July 2025 / Revised: 18 August 2025 / Accepted: 21 August 2025 / Published: 23 August 2025
(This article belongs to the Special Issue Biomimicry for Optimization, Control, and Automation: 3rd Edition)

Abstract

High-value patents are a key indicator of new product development, the emergence of innovative technology, and a source of innovation incentives. Multiple studies have shown that patent value exhibits a significantly skewed distribution, with only about 10% of patents having high value. Identifying high-value patents from a large volume of patent data in advance has become a crucial problem that needs to be addressed urgently. However, current machine learning methods often rely on manual hyperparameter tuning, which is time-consuming and prone to suboptimal results. Existing optimization algorithms also suffer from slow convergence and local optima issues, limiting their effectiveness on complex patent datasets. In this paper, machine learning and intelligent optimization algorithms are combined to process and analyze the patent data. The Fire Hawk Optimization Algorithm (FHO) is a novel intelligence algorithm suggested in recent years, inspired by the process in nature where Fire Hawks capture prey by setting fires. This paper firstly proposes the Enhanced Fire Hawk Optimizer (EFHO), which combines four strategies, namely adaptive tent chaotic mapping, hunting prey, adding the inertial weight, and enhanced flee strategy to address the weakness of FHO development. Benchmark tests demonstrate EFHO’s superior convergence speed, accuracy, and robustness across standard optimization benchmarks. As a representative real-world application, EFHO is employed to optimize Random Forest hyperparameters for high-value patent recognition. While other intelligent optimizers could be applied, EFHO effectively overcomes common issues like slow convergence and local optima trapping. Compared to other classification methods, the EFHO-optimized Random Forest achieves superior accuracy and classification stability. This study fills a research gap in effective hyperparameter tuning for patent recognition and demonstrates EFHO’s practical value on real-world patent datasets.

1. Introduction

High-value patents recognition, or patent evaluation, or patent value assessment, refers to the process of identifying patents that hold high technical, market, economic, legal or strategic value [1]. High-value patents are a key indicator of new product development, the emergence of innovative technology, and a source of innovation incentives [2]. These patents carry important innovations and technological breakthroughs, and have a significant influence on the development of business and society [3]. From an enterprise perspective, the identification and acquisition of high-value patents not only guide technological development and optimize resource allocation but also prevent redundant research efforts and decrease risks associated with innovation. Moreover, such patents enhance technological barriers and protect core competitive advantages, thereby promoting sustainable growth and long-term profitability. At the national level, and for science and technology management departments, accurately identifying high-value patents supports the efficient allocation of resources and informs the development of effective policy strategies [4]. However, low-value patents consume a lot of resources for innovation, which may cause firms to innovate more slowly if their low-value patents are not converted or commercialized. This would impede a nation’s industrial development, technological advancement, productivity growth, and economic prosperity [5,6]. Therefore, the high-value patents recognition from a huge number of patents is crucial for the cultivation, application, protection and management of high-value patents in government departments, scientific research institutions and enterprises, and is also a hot topic of academic research [7,8].
With the progress of complex networks and machine learning methods, it is possible to identify high-value patents from large-scale data [9]. Machine learning can effectively deal with large-scale, high-dimensional and structured patent data, extract deep-seated features of patents, and improve the accuracy of patent value identification [10,11,12]. Hido et al. [13] introduce a machine learning and text mining-based tool that assesses patent application quality by computing a score, which predicts the likelihood of approval, and outperforms conventional methods by utilizing a large dataset and a new statistical prediction model. Wu et al. [14] established a patent quality classification model using support vector machine, self-organizing map and kernel principal component analysis, and developed a patent quality analysis and classification system based on it, and the experimental results showed the advantages of the method. Based on multiple relevant metrics that can be accessed immediately after patent issuance, Lee et al. [15] provide a machine learning method for early detection of emergent inventions. Their method can help with responsive technology forecasting and planning, as demonstrated by the pharmaceutical technology case. A deep learning analytical approach was also used by Trappey et al. [16] to automate the evaluation of patent value in the Internet of Things domain. Using the provided dataset, they first identified important patent value indicators using principal component analysis. A Deep Neural Network (DNN) was utilized to forecast values after a total of 11 value indicators were chosen. The outcomes demonstrated that the DNN model outperformed the conventional Back-Propagation Neural Network technology in terms of accuracy. Kwon et al. [17] address the identification of promising inventions by using patent-based machine learning techniques that incorporate the quality of knowledge accumulation as an input variable, finding it to be the most important predictor when compared to other traditional indicators. Hu et al. [1] experimented with five machine learning algorithms based on multidimensional value patent portfolios for the high-value patents identification. The Random Forest approach performs the best overall, according to the results. Liu et al. [18] proposed a multi-task learning framework that unifies the identification of high-value patents and standard-essential patents, leveraging the mutual reinforcement of both tasks. The proposed model, which uses both structured and embedded textual features of patents, significantly outperforms single-task learning models with regard to accuracy, recall, precision, and F1 measure across both balanced and imbalanced datasets.
However, a critical challenge remains in these machine learning approaches: the performance heavily depends on hyperparameter tuning, which is often performed manually or through conventional search strategies such as grid search, which are time-consuming and prone to suboptimal solutions. Existing intelligent optimization algorithms used for this purpose, such as Genetic Algorithm (GA) and Particle Swarm Optimization (PSO), still face issues like slow convergence and getting trapped in local optima when applied to real-world patent datasets.
In a patent evaluation, the patent lifetime has been used as a substitute for the commercial potential of a patent. Choi et al. [19] suggest a method to assess the commercial potential of individual patents by applying the feed-forward neural network model to forecast the probability that a patent will remain valid until its maximum expiry date and develop a patent business potential evaluation system at the end. Kumar et al. [20] developed a hybrid model that combines binomial regression and multi-class classification to accurately predict the renewal life of Indian patents, addressing the challenge of unusual renewal life distribution and achieving 90% accuracy.
By comparing multiple solutions until an optimal or satisfying response is discovered that produces a higher accuracy score than the previous one, optimization algorithms are used to increase the efficiency of the machine learning area [21]. Machine learning algorithms often require manual hyperparameter selection to achieve optimal performance. This brings difficulties, including more complexity and the requirement for expert knowledge to tune parameters effectively [22,23]. Intelligent optimization algorithms like Genetic Algorithm and Particle Swarm Optimization can be used to find optimal hyperparameters, improving the efficiency and performance of neural networks without exhaustive manual tuning [24,25,26].
Fire Hawk Optimizer (FHO) is a novel intelligence algorithm proposed in recent years, inspired by the process in nature where Fire Hawks capture prey by setting fires [27]. FHO demonstrates superior performance and exceptional results in addressing structural engineering design problems [28,29]. This paper combined the FHO algorithm and Random Forest and carried out a high-value patents recognition experiment based on the patent data from the Patyee database. The primary goal of this study is to propose an improved optimization algorithm and validate its effectiveness through the representative real-world application of high-value patent recognition. Nevertheless, the canonical FHO algorithm suffers from poor convergence speed and vulnerability to local optima, limiting its effectiveness for complex hyperparameter tuning in machine learning tasks. There is a lack of research focusing on enhancing FHO to overcome these shortcomings in the context of patent data analysis. Regarding the issues of the Fire Hawk optimization algorithm’s poor convergence speed and vulnerability to local optimization [30], an enhanced Fire Hawk optimization algorithm called EFHO is proposed. The EHFO improved the canonical FHO by integrating four strategies of adaptive tent chaotic mapping, hunting prey, adding the inertial weight, and an enhanced flee strategy to address the weakness of FHO development. Extensive benchmark experiments on 23 well-known test functions demonstrate that EFHO achieves faster convergence, higher optimization accuracy, and better stability compared to FHO, PSO, GWO, and WOA. Further tests on large-scale problems (up to 1000 dimensions) confirm EFHO’s robustness and dimension insensitivity. While other intelligent optimization algorithms could potentially be applied for hyperparameter tuning, many face challenges like slow convergence or getting trapped in local optima, which EFHO effectively addresses. Applying EFHO to optimize Random Forest hyperparameters for high-value patent recognition on a real-world patent dataset, the EFHO-RF model outperforms other classifiers in recognition accuracy and classification stability, demonstrating EFHO’s practical effectiveness and broad applicability in this domain. The remainder of this paper is structured in the following way. Section 2 introduces the original Fire Hawk Optimization Algorithm. The enhanced Fire Hawk Optimization Algorithm is presented in Section 3. Simulation experiments are described in Section 4. Section 5 explains the application of identifying high-value patents with Random Forest. Finally, the conclusion is presented in Section 6.

2. Fire Hawk Optimization Algorithm

FHO is a novel intelligence optimization algorithm suggested in recent years. The Fire Hawk in the algorithm refers to birds such as brown falcon, black kite and whistling kite. These birds are called Fire Hawk because they find and capture prey by setting fire while hunting in the wild. In the FHO algorithm, the candidate solutions in the search space are mapped to the three categories: the main fire, the prey and the Fire Hawks. The concrete steps of the optimization algorithm are as follows:
(1)
Compute the fitness value for all solution candidates. The best solution candidate in the global searching space is presumed to be the main fire, with the better n solution candidates considered as the Fire Hawks, and the remaining solutions referred to as prey.
(2)
Calculate the length between the prey and the Fire Hawks, and assign each prey to the nearest Fire Hawk based on the distance, thus establishing a territory for each Fire Hawk. The determination of distance relies on the following equation:
D i j = ( x 2 x 1 ) 2 + ( y 2 y 1 ) 2 , i = 1 , 2 , , n j = 1 , 2 , , m ,
where D i j represents the length between the ith Fire Hawk and the jth prey, m represents the number of prey, n represents the number of Fire Hawks, ( x 1 , y 1 ) represents the coordinate of the Fire Hawk, ( x 2 , y 2 ) represents the coordinate of the prey.
(3)
From the main fire, the Fire Hawks gather burning sticks and set fire in their specific territory to force the prey to hastily flee, then move to a new location. Meanwhile, some Fire Hawks are excited to use the sticks that are burning in other Fire Hawks’ territories. The equation of location updating procedures is as follows:
F H i n e w = F H i + ( r 1 × M F r 2 × F H N e a r ) , i = 1 , 2 , , n
where F H i n e w represents the updated location vector of the ith Fire Hawk ( F H i ), M F represents the globally optimal location in the searching space that is regarded as the main fire, F H N e a r represents another Fire Hawk in the searching space, r 1 and r 2 are evenly distributed random integers between 0 and 1 that determine how Fire Hawks will migrate toward the main fire or other occupied territories.
(4)
The prey in the Fire Hawk’s territory start to run away once it sets fire; they may run away, hide, or mistakenly run in the direction of the Fire Hawk. During the location update process, these actions can be expressed by the following equation:
P R q n e w = P R q + ( r 3 × F H l r 4 × S L l ) , l = 1 , 2 , , n . q = 1 , 2 , , r .
where P R q n e w represents the updated location vector of the qth prey, which is denoted as PR q that the lth Fire Hawk, which is denoted as F H l is encircling, SL l represents a safe location in the territory of F H l , r 3 and r 4 are evenly distributed random integers between 0 and 1 that are used to determine how prey will move toward the safe location and the Fire Hawks. The mathematical presentation of S L l is formulated as follows:
S L l = q = 1 r P R q r , q = 1 , 2 , , r l = 1 , 2 , , n
where the P R q is the qth prey that F H l is encircling.
(5)
Additionally, the prey may run out of the current Fire Hawk’s territory, potentially to another Fire Hawk’s territory, or to a safer location. The position update equation takes these actions into consideration.
P R q n e w = P R q + ( r 5 × F H O t h e r r 6 × S L ) , l = 1 , 2 , , n . q = 1 , 2 , , r .
In this equation, P R q n e w represents the updated location vector of the qth prey which is denoted as P R q that the lth Fire Hawk which is denoted as F H l is encircling, F H O t h e r represents another Fire Hawk in the searching space, S L represents a safe location beyond the territory of F H l , r 5 and r 6 are evenly distributed random integers between 0 and 1 that are used to determine how prey will approach the safe location beyond the territory or the other Fire Hawks. The equation of S L is as follows:
S L = j = 1 m P R j m
where the P R j is the jth prey in the search space.
(6)
Return to step 1, loop until certain conditions are met to obtain the global best solution, and the algorithm ends.
A growing body of research has applied FHO to diverse real-world and engineering problems, demonstrating its versatilibiomimetics-10-00561ty and effectiveness. In construction project management, Shishehgarkhaneh et al. [28] utilized FHO within a Building Information Modeling framework to balance multiple resource trade-offs such as time, cost, quality, risk, and environmental impact. Their results highlighted FHO’s capability to yield competitive and exceptional solutions in multi-objective scheduling problems. Similarly, Hosseinzadeh et al. [29] applied FHO to enhance security and energy efficiency in wireless sensor networks through a trust-based routing protocol, achieving superior performance in network reliability and energy consumption.
Algorithmic improvements of FHO itself have also been explored. Ashraf et al. [30] introduced novel swarm initialization techniques leveraging quasi-random sequences to boost convergence rates and diversity, significantly outperforming the standard FHO. Baweja et al. [31] proposed the Levy Flight-based Fire Hawk Optimizer, enhancing exploration and reducing premature convergence, with experimental validations confirming improved optimization outcomes on standard benchmarks.
In the domain of human-computer interaction and activity recognition, Alonazi et al. [32] integrated FHO with deep learning for hyperparameter tuning, resulting in enhanced classification accuracy and robustness for human activity detection. In renewable energy modeling, Said et al. [33] employed a modified version of FHO to precisely extract photovoltaic parameters in solar cell models, outperforming other metaheuristic algorithms and demonstrating superior prediction accuracy. In energy system modeling, Khajuria et al. [34] applied a modified FHO to identify unknown parameters in solid oxide fuel cell models, achieving highly accurate parameter estimation across varying temperatures and pressures.

3. Enhanced Fire Hawk Optimization Algorithm

Although recent FHO variants have introduced improvements such as quasi-random sequence initialization [30], Levy flight strategies [31], and domain-specific hybridization [32,33], these methods generally focus on enhancing either the exploration phase or the exploitation phase in isolation. They often fail to simultaneously maintain population diversity and balance global–local search throughout the optimization process, particularly in high-dimensional or complex search spaces.
The EFHO improved the canonical FHO by integrating four strategies of hunting prey, adaptive tent chaos mapping, adding the inertial weight, and an enhanced flee strategy to address the weakness of FHO development. These four strategies are designed to work in a complementary manner, jointly strengthening both exploration and exploitation while preserving population diversity. This holistic enhancement aims to overcome premature convergence and improve robustness across a wider range of problem types.

3.1. Adaptive Tent Chaos Mapping Strategy

In the original algorithm FHO, the initialization strategy is to randomly distribute throughout the entire space, which has high randomness and uneven distribution, leading to a lack of population diversity. Introducing an adaptive tent chaotic mapping strategy to provide a more uniform distribution is beneficial for obtaining high-quality initial populations. The following equation represents the adaptive tent chaos mapping, assuming that the sequence’s beginning values are random numbers between 0 and 1 [35].
s i + 1 = 1 r a n d ( 0 , 1 ) s i = 0 1 2 * s i 0 < s i < 0.5 2 * ( 1 s i ) 0.5 < s i < 1
In this equation, s i + 1 represents the i + 1th mapping value, and the s i represents the ith mapping value. The following is the modified initialization equation.
X j i ( i n i t i a l ) = X j min + s i * ( X j max X j min ) ,   i = 1 , 2 , , N j = 1 , 2 , , d
In this equation, X j i ( i n i t i a l ) represents the jth dimension value of the ith object in the initial population, X j min represents the minimum value on the jth dimension, X j max and represents the maximum value on the jth dimension.

3.2. Hunting Prey

The FHO algorithm lacked the hunting prey part for Fire Hawk. To broaden the scope of the search and further strengthen the capability of local development, the part of hunting prey was added. After the prey flees, the fitness value is calculated based on the position after fleeing, and the prey that has the best fitness value is selected for hunting. The hunting method refers to the spiral attack strategy of the whale optimization algorithm [36], and the equation is as follows:
F H l n e w = D l × e b k × cos ( 2 π k ) + P R l b e s t
where PR l b e s t represents the prey with the best position after fleeing, D l represents the length between the prey ( PR l b e s t ) and the Fire Hawk ( F H l ), the parameter b represents the shape parameter of the logarithmic spiral; this paper sets b at 1, and the parameter k represents an arbitrary number between −1 and 1.

3.3. Adding the Inertia Weight

This paper adds the inertia weight to the equation for setting fire to the Fire Hawks. The changed equation is as follows:
F H l n e w = ω × F H l + ( r 1 × G B r 2 × F H N e a r )
The inertia weight plays an important role in enhancing the search accuracy and accelerating the convergence speed of solutions [37,38]. In the early stage, larger inertia weight facilitates a more robust global search capability, whereas in the later stage, smaller inertia weight has stronger exploitation capability [39]. Consequently, it is imperative to adjust the inertia weight in a non-linear decreasing fashion. The following is the equation of inertia weight:
ω = 1 + sin ( e t T + a ) b
where T represents the total iteration number, t represents the current iteration number, and a and b are constants (i.e., a = 0.5 and b = 20).
To visually illustrate the features of the inertia weight function suggested in this paper, Figure 1 of this paper shows the function curve of the inertia weight, where the iteration number is shown on the x-axis with a maximum of 50, and it can be observed that the inertia weight gradually decreases from 0.1 to 0 as the iteration number increases.

3.4. Enhanced Flee Strategy

There are two flee strategies in the original algorithm, where all prey implemented both flee strategies in sequence. In order to increase population diversity, some prey implemented the first flee strategy while others implemented the second flee strategy. The Equations (3) and (5) are combined as follows:
P R q n e w = P R q + ( r 3 × F H l r 4 × S P l ) τ 0.5 P R q + ( r 5 × F H A l t e r r 6 × S P ) τ > 0.5
where τ is evenly distributed random integers between 0 and 1.
In addition, this paper introduces the Levy flight [40,41] for the first flee strategy and the t-distribution perturbation [42,43] strategy for the second flee strategy. The equation of the Levy flight for the first flee strategy is as follows:
P R f l e e 1 L = P R f l e e 1 b e s t × l e v y ( D )
where P R f l e e 1 b e s t is the best solution candidate in the prey’s implemented first flee strategy, l e v y ( D ) represents the Levy flight function, and D is the dimension. After the levy flight, a greedy selection is made between the old and new positions, choosing the one that has better fitness value for the next iteration. The following is the greedy selection equation:
P R f l e e 1 b e s t = P R f l e e 1 b e s t f ( P R f l e e 1 b e s t ) < f ( P R f l e e 1 L ) P R f l e e 1 L f ( P R f l e e 1 b e s t ) f ( P R f l e e 1 L )
Levy flying, which combines random movement over long and short distances, is a technique to imitate the random flight of animals. During the initial phase of the algorithm, long distances can expand the search range and explore discoveries, which is beneficial for enhancing population diversity and reducing the risk of local optima. In the latter stage of the algorithm, the range of the global optimal solution is basically determined, and short distances can improve the accuracy of the solutions, enabling the algorithm to converge to the global best.
The equation of t-distribution perturbation for the second flee strategy is as follows:
P R f l e e 2 T = P R f l e e 2 b e s t × ( 1 + t _ d i s t r i b u t i o n ( t ) )
where P R f l e e 2 b e s t is the best solution candidate in the prey that implemented the second flee strategy. The t-distribution function is a distribution function that is symmetric about the y-axis, and its two bounding distributions are the standard Cauchy and Gaussian distributions; this function usually contains only one parameter, called degrees of freedom. The degrees of freedom are the iteration number t in this paper, and the following is the equation of the probability density function.
f ( x ) = Γ ( t + 1 2 ) π t Γ t 2 1 + x 2 t t + 1 2 ,   < x < +
In the early stage with a small iteration number, the t-distribution gains the ability by approaching the Cauchy distribution. In the middle of the iteration, the t-distribution balances the convergence and population variety of the algorithm by falling between the Gaussian distribution and the Cauchy distribution. In the later stage, the t-distribution gets close to the Gaussian distribution as the iteration number is large, which facilitates the local exploration ability. After the perturbation, the greedy selection model is also used:
P R f l e e 2 b e s t = PR f l e e 2 b e s t f ( P R f l e e 2 b e s t ) < f ( P R f l e e 2 L ) P R f l e e 2 L f ( P R f l e e 2 b e s t ) f ( P R f l e e 2 L )
When the prey flees, the algorithm performs local exploitation around the fire hawk. As the iteration number increases, most of the prey may stay around the fire hawk, causing the algorithm to stagnate and be unable to further solve the global optimal solution. During the flee stage of prey, using the t-distribution perturbation strategy on the current location can prevent prey from staying around the fire hawk and reduce the likelihood of algorithmic stagnation. The convergence speed and the capacity to deviate from local optima were further improved.
The following is a description of the EFHO algorithm steps:
Step 1: Initialization.
The dimension Dim, maximum iterations of number T, and population size pop are set. The population is initialized by Equation (8). Set the current iteration number t as 1.
Step 2: Evaluation.
Compute fitness value for all individuals, the individual that has the best fitness value is considered as the main fire, select better n individuals as the Fire Hawks, make the remaining individuals as prey.
Step 3: Determining Territory.
For each Fire Hawk, compute the length between each Fire Hawk and each prey by Equation (1), and determine its territory where the preys belong.
Step 4: Setting Fire.
The Fire Hawks set fire and adjust the position according to Equation (10).
Step 5: Prey flees.
Some prey flee by the first strategy by Equation (12) and calculate the fitness for each prey. The prey with the best fitness value after fleeing then implements the Levy flight according to Equations (13) and (14). Some prey flee by the second strategy and calculate the fitness for each prey. The prey with the best fitness value after fleeing then implements the t-distribution perturbation according to Equations (15) and (17).
Step 6: Hunting prey.
The Fire Hawks hunt prey and adjust their position according to Equation (9).
Step 7: Termination.
When the termination condition is reached, the optimal individual position and fitness value are outputted, and the algorithm terminates; if not, proceed to Step 2.

4. Simulation Experiment

Two experiments are conducted in this section. The first experiment tested the capability of the EFHO algorithm by comparing it with several closely related algorithms, and the second experiment tested the scalability of the enhanced algorithm for large-scale problems.

4.1. Experiment Setup

To evaluate the proposed algorithm’s convergence rate and optimization precision, 23 well-known test functions of CEC2005 [44] are used for comparison experiments. In these functions, F1 to F7 are unimodal and commonly used as benchmark functions for evaluating search accuracy. Optimization algorithms often struggle to converge to the global optimum for these functions. F8 to F13 are multi-modal functions used to assess global search performance and have multiple local extreme points. F14 to F23 are benchmark functions that are fixed-dimension multi-modal. To obtain unbiased results, the maximum iteration number was set to 500 and the population size was set to 30, for all functions in all comparisons. For each function, all algorithms were independently run 30 times.
All experiments were carried out on a personal computer with Intel(R) Core(TM) i9-12900H CPU, 32.00 GB RAM and Windows 11 operating system, and MATLAB R2023b was used to implement all algorithms.

4.2. Performance Comparison

In this paper, FHO, PSO (Particle Swarm Optimization), GWO (Grey Wolf Optimization) and WOA (Whale Optimization Algorithm) are selected for comparison with EFHO to evaluate the performance of EFHO. Since some of the 23 functions do not have an optimal value of 0, this paper calculates fitness by taking the absolute value of the discrepancy between the computed result and the actual optimal value of the function. This approach converts all optimal values to 0. This paper uses the best value (Best), mean value (Mean), and standard deviation (Std) to evaluate the performance of all algorithms. Table 1 displays the test results; the best result for each function is boldfaced.
According to Table 1, for F1, F2, F3, F4, F9 and F11, EFHO can achieve the ideal value in theory. For F5, F6, F7, F8 and F13, the outcomes of EFHO outperform the other four algorithms. For F10, the best values obtained by the EFHO algorithm are the same as those of the FHO algorithm, superior to the other three algorithms. The above results demonstrate the remarkable accuracy and stability of EFHO.
To assess the statistical significance of the performance differences between EFHO and the baseline algorithms, the Wilcoxon signed-rank test was conducted for each of the 23 benchmark functions. Table 2 summarizes the number and proportion of functions where EFHO achieved a significantly better result (p < 0.05) compared to each baseline. The detailed test results are provided in Appendix A.
As shown in Table 2, EFHO achieved statistically significant improvements (p < 0.05) over PSO, GWO, and WOA in more than half of the benchmark functions, and outperformed FHO in 39.13% of the cases. While the proportion of significant wins varies across comparisons, these results still indicate that EFHO consistently delivers competitive or superior performance against the baseline algorithms across a wide range of test functions.
Figure 2 displays the five algorithms’ convergence curves for each test function, which better illustrates the benefit of EFHO. It is demonstrated that among them, the EFHO convergence rate is relatively quick. The EFHO convergence curves are smooth and decline quickly for the majority of functions.
In conclusion, the EFHO algorithm demonstrated superior convergence, stability, and accuracy compared to the original FHO and the other four algorithms, which adequately validates the feasibility of the suggested method for FHO improvement in this paper.

4.3. EFHO’s Scalability Test for Large-Scale Problems

This paper further confirms whether the EFHO is scalable when it comes to solving large-scale problems, as real-world engineering applications often encounter large-scale optimization challenges. Because the dimensions of F14 to F23 are fixed, the paper selects test functions F1 to F13 and respectively sets the dimensionality of EFHO to 500 and 1000. Table 3 displays the high-dimensional experiment results.
Table 3 shows that even for functions with 500 and 1000 dimensions, EFHO still has a good accuracy. Especially for F1, F2, F3, F4, F9 and F11, EFHO can achieve the ideal value 0 in theory. The results of EFHO for other functions are in line with those of 500 and 1000 dimensions, suggesting that EHO is not dimension-sensitive during problem-solving. The above analysis shows that EFHO performs well when handling large-scale problems since it is not significantly impacted by a significant increase in dimension while maintaining high accuracy.

5. High-Value Patents Recognition with Random Forest

Leo Breiman and Adele Cutler first presented the Random Forest ensemble learning method in 2001 [45]. It addresses classification and regression problems by constructing multiple decision trees and then improving prediction accuracy, generalization ability, and resistance to overfitting through averaging (for regression problems) or majority voting (for classification problems) [46]. Random Forest’s basic idea is to construct many decision trees, where each tree is independent and the features within each tree are randomly selected, thereby reducing the model’s variance [47]. To get the final prediction result, Random Forest averages or votes on each tree’s prediction results. Random Forest works well for a variety of regression and classification problems, especially for complex, high-dimensional datasets and problems that require handling a large number of features. Due to its excellent generalization ability and resistance to overfitting, Random Forest performs well in many practical applications [48].
The Random Forest model has a number of hyperparameters that can be tuned by the user to optimise its performance, such as the number of trees, the minimum number of samples required in a node, and the number of features to be considered for each segmentation. These hyperparameters, also called tuning parameters, control the construction and training process of the model. Because optimal tuning parameter values depend on the dataset being used, they must be carefully chosen [49]. Many researchers employ intelligent optimization algorithms to tune the hyperparameters of random forests, seeking the best hyperparameters for the current dataset [50,51,52]. This paper employs the EFHO algorithm to optimize the hyperparameters of the random forest and conducts application research for the high-value patents identification based on it.

5.1. Dataset

The patent dataset used in this paper is collected from the Patyee database. Patyee is a comprehensive database providing extensive information on patents, including over 180 million in-depth processed patent records from more than 171 countries, regions, and organizations worldwide. This paper retrieves Chinese invention patents that have received awards, such as the China Patent Award, have been published for over ten years, and are still valid, using them as positive samples for the high-value patents dataset, totaling 10,638 patents. The choice of awarded patents as high-value patents is based on the award attribute provided by the Patyee database, which compiles official records from the China National Intellectual Property Administration and other authoritative sources, and is supported by prior research [5,18] with two studies in our references also adopting this approach. Given that such awards are granted through rigorous expert review, the accuracy of this labeling is expected to be very high.
Additionally, to construct the negative sample set, this study retrieves Chinese invention patents with a patent value of 100, published for over ten years, and still valid, totaling 18,030 patents. After removing the awarded patents, the remaining 17,734 patents are used as negative samples. The choice of value 100 is mainly driven by two considerations. First, the number of non-awarded patents is extremely large, reaching several million, making it necessary to narrow the search with specific criteria. Selecting those with a value of 100 results in a set size close to that of the positive samples, which facilitates a relatively balanced dataset. Second, the value score is an internal evaluation index calculated by the Patyee database using a proprietary method that is not publicly disclosed. A score of 100 is simply one of the discrete values in this system and is used here as a filtering attribute, not as a classification threshold. While such database-calculated values do not necessarily indicate truly high-value patents, awarded patents are determined through expert review and can be considered genuinely high-value. In this study, patents with a value of 100 but without awards are treated as negative samples, not to imply that they have low value, but to establish a clear and consistent labeling criterion in which only awarded patents are considered positive. This approach minimizes ambiguity in the definition of high-value patents and avoids dependence on database-specific scoring methods, ensuring consistency in the training labels. Additionally, this paper selects both positive and negative samples from patents that have been published for over ten years, in order to exclude the influence of the publication time.
Numerous studies on patent value indexes have been conducted, taking into account various scales and embracing a wide range of perspectives [53]. In paper 18, researchers used the title and abstract of patents as textual features, which were converted into vectors using the BERT model and then concatenated with other structural features to serve as the model’s input for training. The final accuracy was only a little over sixty percent, which is lower than the accuracy obtained by using only structural features for training [1]. This indicates that textual features contribute little to the identification of high-value patents and might even act as noise. Therefore, this paper does not use textual features but selects the most commonly used patent value indicators as the features of the dataset. Table 4 lists the 14 structural features that are present in the dataset.

5.2. Data Preprocessing

An essential stage in machine learning is data preprocessing. Enhancing data quality, eliminating noise and outliers, and guaranteeing consistency and completeness involves cleaning, converting, and standardizing data.
Missing data can lead to biased results, reduce the statistical power of analyses, and affect the overall performance of machine learning models. There are two approaches to handling missing data: deletion or imputation. There are many methods for imputation, with the most common being replacing missing values with the mean or median. This paper firstly performed a missing data check on the dataset and found that every row and column contains data, so no missing value treatment is necessary.
The dataset used in this paper was split 80:20, with the training set being the remaining 80% and the testing set being 20%. Prior to the split, the dataset was randomly shuffled to ensure a balanced distribution of the data. Before training, this paper also performed normalization on the data, transforming the values of each feature to a value between 0 and 1. The equation of normalization is as follows:
v n o r m = v v min v max v min
where v represents the raw value. v min represents the minimum value, v max represents the maximum value. v n o r m represents the transformed value after normalization, which will be within the range of 0 to 1.

5.3. Experiments and Results

  • Experiment Setup
This paper optimized the Random Forest using the EFHO algorithm and conducted experimental verification based on the patent dataset. This paper also selects Naive Bayes (NB), Back Propagation Neural Network (BP), Support Vector Machine (SVM), and Logistic Regression (LR) for comparison. This article uses MATLAB for experiments; each model was implemented using built-in MATLAB functions. The NB model was called using the fitcnb function, the LR model using the fitclinear function, the SVM model using the fitcsvm function, the RF model using the TreeBagger function, and the BP neural network using the feedforwardnet function; the hidden layer’s node number is set to 8.
In this experiment, the objective function is the mean square error of the classification result, and the formula for the mean square error is as follows:
M = 1 n i = 1 n ( x i x ^ i ) 2
where n represents the amount of data in the training or testing set, x i represents the true value, and x ^ i represents the predicted value of the model. The algorithm terminates when either the maximum iteration number is exceeded or the objective function falls below the specified threshold.
2.
Results
This paper separately applies the LR model, the NB model, the RF model, the SVM model, the BP model and the RF model optimized using EFHO (EFHO-RF) to high-value patents recognition. To compare the optimization effects, this paper also optimized the BP model using the EFHO algorithm as EFHO-BP. Table 5 displays the results of the experiment. Five-fold cross-validation was used to evaluate each model by indicators such as accuracy, recall, precision, F1 measure and AUC. The AUC means the Area Under the Curve (AUC) of the Receiver Operating Characteristics (ROC) curve.
According to the results of the experiment, the EFHO-RF model performs best, with all indicators, except for precision, being at the highest level. All indicators are superior to the pre-optimization RF model. The accuracy improved from 95.8% before optimization to 96.6%, the recall improved from 91.8% to 96.3%, the precision increased from 96.7% to 97%, the F1 measure increased from 94.2% to 96.3%, and the AUC improved from 98.8% to 99%. The EFHO-BP also performed well, with accuracy, recall, and F1 measure surpassing all models except EFHO-RF. However, its precision and AUC were lower than those of the RF model before optimization, and its precision was even lower than that of the BP model before optimization. The experimental results indicate that EFHO can find better hyperparameters for the Random Forest, providing a desirable solution for the problem. In this classification task, EFHO contributed to improving the Random Forest’s use of 14 patent-level indicators (Table 4) by identifying hyperparameter settings that better balance feature relevance and model complexity. While a separate feature importance ranking was not conducted, the selected features are widely recognized in patent analytics as being strongly associated with patent value, based on both the literature and our long-term practical experience.
In order to assess the discriminative ability of the model, this paper also examined the ROC curve. The AUC for the optimized Random Forest was 99.0%, demonstrating a high level of performance. The ROC curve shows a marked increase in true positive rates while keeping low false positive rates, further validating the robustness of our optimization approach. These results highlight the potential of combining metaheuristic algorithms like FHO with machine learning methods to enhance classification tasks, providing a more reliable and efficient model for practical applications. The ROC curves before and after optimization are shown in Figure 3. This suggests that the Random Forest model’s classification performance is excellent, and the EFHO algorithm can help improve the classification performance.
To make the identification results more transparent, this paper further conducted a feature importance analysis using the out-of-bag (OOB) permutation method. The results are shown in Figure 4. As displayed, the claims number (CLMSN), the countries number belongs to the same family (NCSE), and the assignments number (RAQ) ranked as the top three most influential features. This highlights the central role of patent scope, international family coverage, and technology transfer activities in distinguishing high-value patents. By contrast, features such as the reexamination number (NR) and the applicant’s number (PAN) showed limited or negligible importance in classification. These findings indicate that the EFHO-RF model identifies high-value patents by leveraging a meaningful combination of legal, citation, and family-related attributes. Presenting feature importance thus makes the identified results more visible and interpretable, complementing the performance evaluation and providing additional insights into the determinants of patent value.

6. Conclusions

This paper introduces a novel enhanced Fire Hawk optimization algorithm called EFHO. The EFHO improves upon the canonical FHO by incorporating four key strategies: adaptive tent chaotic mapping, hunting prey, the addition of inertial weight, and an enhanced flee strategy. These modifications address the shortcomings of the original FHO, such as weak convergence and limited exploration capabilities. By enhancing these aspects, EFHO provides a more robust and efficient optimization method.
While EFHO is a general intelligent optimization algorithm, this study uses high-value patent recognition primarily as a representative application scenario to validate its effectiveness. The challenges inherent in such real-world applications demand optimization algorithms with faster convergence and stronger global search ability than traditional methods. Existing hyperparameter tuning techniques often suffer from inefficiency and suboptimal solutions in such scenarios.
This paper used the EFHO algorithm to optimize the Random Forest model’s hyperparameters in the chosen application scenario. The resulting EFHO-RF classification model leverages the strengths of both EFHO and Random Forest, leading to superior performance. EFHO effectively tuned key Random Forest hyperparameters such as the number of trees, the minimum number of samples required in a node, and the number of features to be considered for each segmentation, leading to better utilization of informative features and reduced overfitting. The optimization process improved the model’s balance between feature relevance and complexity, resulting in more stable and accurate classification outcomes. Experimental results demonstrate that this model achieves a high level of accuracy on the selected patent dataset, significantly outperforming traditional methods. As awarded patents are determined through expert review, the identified positives inherently represent truly valuable patents, aligning with the ultimate goal of high-value patent identification. Thus, this research not only advances optimization algorithm methodology but also demonstrates its applicability through a representative real-world task, highlighting the method’s relevance and effectiveness in applied settings.
Beyond the empirical results, this study offers broader scientific and practical value. The proposed EFHO-RF framework provides a replicable approach for addressing other machine learning tasks. Its successful validation in the chosen application scenario offers methodological insights for similar applied analytics and decision-support contexts.
However, certain limitations should be acknowledged. The dataset is limited to Chinese invention patents, and the positive samples are defined solely by award status, which, although highly accurate, may not fully capture all possible interpretations of patent value. In addition, the EFHO-RF model has yet to be tested on datasets from other jurisdictions or with different feature spaces, which may affect its generalizability. It should also be noted that high-value patent recognition is essentially a complex classification problem. By leveraging EFHO’s strong capability in hyperparameter optimization and Random Forest’s robust classification ability, the EFHO-RF framework provides a suitable and effective solution for this type of task. Future work will include cross-domain validation, incorporation of alternative indicators of patent value, and application of EFHO to diverse and heterogeneous datasets.

Author Contributions

All authors contributed to the study conception and design. X.Y. proposed the overall research framework and supervised the project. Patent data collection and analysis were performed by H.L. The first draft of the manuscript was written by S.W., and all authors commented on previous versions of the manuscript. X.Y. also revised and finalized the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Social Science Foundation of China under Grant 19CTQ007 and in part by the Gansu Provincial Natural Science Foundation under Grant 23JRRA581.

Data Availability Statement

The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.

Acknowledgments

The authors thank the anonymous reviewers for their insightful comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Wilcoxon signed-rank test results.
Table A1. Wilcoxon signed-rank test results.
EFHO vs. FHO (p)EFHO vs. PSO (p)EFHO vs. GWO (p)EFHO vs. WOA (p)
F19.1269 × 10−79.1269 × 10−79.1269 × 10−79.1269 × 10−7
F29.1269 × 10−79.1269 × 10−79.1269 × 10−79.1269 × 10−7
F39.1269 × 10−79.1269 × 10−79.1269 × 10−79.1269 × 10−7
F49.1269 × 10−79.1269 × 10−79.1269 × 10−79.1269 × 10−7
F52.7455 × 10−39.1269 × 10−79.1269 × 10−79.1269 × 10−7
F62.4836 × 10−69.1269 × 10−71.2377 × 10−69.1269 × 10−7
F79.1269 × 10−79.1269 × 10−79.1269 × 10−71.0106 × 10−6
F86.4886 × 10−69.1269 × 10−79.1269 × 10−72.7389 × 10−6
F91.0000 × 1009.1269 × 10−78.9229 × 10−75.0000 × 10−1
F101.0000 × 1009.1269 × 10−78.6624 × 10−73.8176 × 10−6
F111.0000 × 1009.1269 × 10−79.7656 × 10−45.0000 × 10−1
F121.0000 × 1001.0621 × 10−27.1267 × 10−69.9493 × 10−1
F139.1269 × 10−79.1269 × 10−79.1269 × 10−79.1269 × 10−7
F141.0000 × 1001.0000 × 1009.9550 × 10−11.0000 × 100
F158.2267 × 10−18.7948 × 10−16.5960 × 10−19.8112 × 10−1
F161.0000 × 1001.0000 × 1001.0000 × 1001.0000 × 100
F171.0000 × 1001.0000 × 1001.0000 × 1001.0000 × 100
F189.9996 × 10−11.0000 × 1009.9999 × 10−11.0000 × 100
F199.9837 × 10−11.0000 × 1001.0000 × 1001.0000 × 100
F201.0000 × 1001.0000 × 1001.0000 × 1001.0000 × 100
F219.4338 × 10−13.0623 × 10−28.1727 × 10−17.2097 × 10−2
F239.5215 × 10−18.1727 × 10−11.0000 × 1001.0621 × 10−2

References

  1. Hu, Z.; Zhou, X.; Lin, A. Evaluation and identification of potential high-value patents in the field of integrated circuits using a multidimensional patent indicators pre-screening strategy and machine learning approaches. J. Informetr. 2023, 17, 101406. [Google Scholar] [CrossRef]
  2. He, C.; Shi, F.; Tan, R. Evaluation and cultivation method of high-tech value patents for mechanical products. PLoS ONE 2024, 19, 0298144. [Google Scholar] [CrossRef]
  3. Artz, K.W.; Norman, P.M.; Hatfield, D.E.; Cardinal, L.B. A longitudinal study of the impact of r&d, patents, and product innovation on firm performance. J. Prod. Innov. Manag. 2010, 27, 725–740. [Google Scholar] [CrossRef]
  4. Deng, N.; Zhang, J. Hmfm: A method for identifying high-value patents by fusing multiple features. Comput. Mater. Contin. 2024, 82, 1–10. [Google Scholar] [CrossRef]
  5. Wang, S.; Zhou, H.; Zhao, T. Configuration paths to high-value patents: Evidence from patents winning the china patent awards. Scientometrics 2024, 129, 2633–2658. [Google Scholar] [CrossRef]
  6. Rizzo, U.; Sterzi, V. Original but of low value: The paradox of science-industry collaborative patents. Rev. D’economie Ind. 2023, 183, 113–142. [Google Scholar]
  7. Liu, L.-j.; Cao, C.; Song, M. China’s agricultural patents: How has their value changed amid recent patent boom? Technol. Forecast. Soc. Change 2014, 88, 106–121. [Google Scholar] [CrossRef]
  8. Huang, K.G.-L.; Huang, C.; Shen, H.; Mao, H. Assessing the value of china’s patented inventions. Technol. Forecast. Soc. Change 2021, 170, 120868. [Google Scholar] [CrossRef]
  9. Zhou, G.; Tong, Y.; Wang, H. Analysis of key features of high-value patent based on lasso-logit. Financ. Eng. Risk Manag. 2023, 6, 87–95. [Google Scholar]
  10. Miric, M.; Jia, N.; Huang, K.G. Using supervised machine learning for large-scale classification in management research: The case for identifying artificial intelligence patents. Strateg. Manag. J. 2023, 44, 491–519. [Google Scholar] [CrossRef]
  11. Zhou, Y.; Dong, F.; Liu, Y.; Ran, L. A deep learning framework to early identify emerging technologies in large-scale outlier patents: An empirical study of cnc machine tool. Scientometrics 2021, 126, 969–994. [Google Scholar] [CrossRef]
  12. Hu, Y.; Yang, S.; Shi, A. Research on transferable patent recognition based on machine learning. In Proceedings of the 2021 4th International Conference on Data Science and Information Technology, Shanghai, China, 23–25 July 2021; pp. 273–278. [Google Scholar]
  13. Hido, S.; Suzuki, S.; Nishiyama, R.; Imamichi, T.; Takahashi, R.; Nasukawa, T.; Id´e, T.; Kanehira, Y.; Yohda, R.; Ueno, T.; et al. Modeling patent quality: A system for large-scale patentability analysis using text mining. Inf. Media Technol. 2012, 7, 1180–1191. [Google Scholar] [CrossRef]
  14. Wu, J.-L.; Chang, P.-C.; Tsao, C.-C.; Fan, C.-Y. A patent quality analysis and classification system using self-organizing maps with support vector machine. Appl. Soft Comput. 2016, 41, 305–316. [Google Scholar] [CrossRef]
  15. Lee, C.; Kwon, O.; Kim, M.; Kwon, D. Early identification of emerging technologies: A machine learning approach using multiple patent indicators. Technol. Forecast. Soc. Change 2018, 127, 291–303. [Google Scholar] [CrossRef]
  16. Trappey, A.J.; Trappey, C.V.; Govindarajan, U.H.; Sun, J.J. Patent value analysis using deep learning models—The case of iot technology mining for the manufacturing industry. IEEE Trans. Eng. Manag. 2019, 68, 1334–1346. [Google Scholar] [CrossRef]
  17. Kwon, U.; Geum, Y. Identification of promising inventions considering the quality of knowledge accumulation: A machine learning approach. Scientometrics 2020, 125, 1877–1897. [Google Scholar] [CrossRef]
  18. Liu, W.; Li, S.; Cao, Y.; Wang, Y. Multi-task learning based high-value patent and standard-essential patent identification model. Inf. Process. Manag. 2023, 60, 103327. [Google Scholar] [CrossRef]
  19. Choi, J.; Jeong, B.; Yoon, J.; Coh, B.-Y.; Lee, J.-M. A novel approach to evaluating the business potential of intellectual properties: A machine learning-based predictive analysis of patent lifetime. Comput. Ind. Eng. 2020, 145, 106544. [Google Scholar] [CrossRef]
  20. Kumar, A.; Ranjan, P.; Koley, A.; Danish, S. A new hybrid machine learning model for predicting the renewal life of patents. PLoS ONE 2024, 19, 0306186. [Google Scholar] [CrossRef]
  21. Mohapatra, N.; Shreya, K.; Chinmay, A. Optimization of the random forest algorithm. In Advances in Data Science and Management: Proceedings of ICDSM 2019, Hunan, China, 22–23 February 2019; Springer: Berlin/Heidelberg, Germany, 2020; pp. 201–208. [Google Scholar]
  22. Yang, L.; Shami, A. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing 2020, 415, 295–316. [Google Scholar] [CrossRef]
  23. Luo, G. A review of automatic selection methods for machine learning algorithms and hyper-parameter values. Netw. Model. Anal. Health Inform. Bioinform. 2016, 5, 18. [Google Scholar] [CrossRef]
  24. Abdolrasol, M.G.; Hussain, S.S.; Ustun, T.S.; Sarker, M.R.; Hannan, M.A.; Mohamed, R.; Ali, J.A.; Mekhilef, S.; Milad, A. Artificial neural networks based optimization techniques: A review. Electronics 2021, 10, 2689. [Google Scholar] [CrossRef]
  25. Sarmah, D.K. A survey on the latest development of machine learning in genetic algorithm and particle swarm optimization. In Optimization in Machine Learning and Applications; Springer: Berlin/Heidelberg, Germany, 2020; pp. 91–112. [Google Scholar]
  26. Abd Elaziz, M.; Dahou, A.; Abualigah, L.; Yu, L.; Alshinwan, M.; Khasawneh, A.M.; Lu, S. Advanced metaheuristic optimization techniques in applications of deep neural networks: A review. Neural Comput. Appl. 2021, 33, 14079–14099. [Google Scholar] [CrossRef]
  27. Azizi, M.; Talatahari, S.; Gandomi, A.H. Fire hawk optimizer: A novel metaheuristic algorithm. Artif. Intell. Rev. 2023, 56, 287–363. [Google Scholar] [CrossRef]
  28. Shishehgarkhaneh, M.B.; Azizi, M.; Basiri, M.; Moehler, R.C. Bim-based resource tradeoff in project scheduling using fire hawk optimizer (FHO). Buildings 2022, 12, 1472. [Google Scholar] [CrossRef]
  29. Hosseinzadeh, M.; Yoo, J.; Ali, S.; Lansky, J.; Mildeova, S.; Yousefpoor, M.S.; Ahmed, O.H.; Rahmani, A.M.; Tightiz, L. A cluster-based trusted routing method using fire hawk optimizer (FHO) in wireless sensor networks (WSNs). Sci. Rep. 2023, 13, 13046. [Google Scholar] [CrossRef]
  30. Ashraf, A.; Anwaar, A.; Haider Bangyal, W.; Shakir, R.; Ur Rehman, N.; Qingjie, Z. An improved fire hawks optimizer for function optimization. In Advances in Swarm Intelligence, Proceedings of the 14th International Conference on Swarm Intelligence 2023, Shenzhen, China, 14–18 July 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 68–79. [Google Scholar]
  31. Baweja, D.; Jain, A.; Bhandari, A.; Bohat, V.K. Levy flight based fire hawk optimizer. In Proceedings of the 2024 IEEE Region 10 Symposium (TENSYMP), New Delhi, India, 27–29 September 2024; pp. 1–6. [Google Scholar]
  32. Alonazi, M.; Alnfiai, M.M. Fire Hawk Optimizer with Deep Learning Enabled Human Activity Recognition. Comput. Syst. Sci. Eng. 2023, 45, 3135–3150. [Google Scholar] [CrossRef]
  33. Said, M.; Ismaeel, A.A.; El-Rifaie, A.M.; Hashim, F.A.; Bouaouda, A.; Hassan, A.Y.; Abdelaziz, A.Y.; Houssein, E.H. Evaluation of modified fire hawk optimizer for new modification in double diode solar cell model. Sci. Rep. 2024, 14, 30079. [Google Scholar] [CrossRef]
  34. Khajuria, R.; Bukya, M.; Lamba, R.; Kumar, R. Optimal parameter identification of solid oxide fuel cell using modified fire hawk algorithm. Sci. Rep. 2024, 14, 22469. [Google Scholar] [CrossRef]
  35. Ma, J.; Hao, Z.; Sun, W. Enhancing sparrow search algorithm via multi-strategies for continuous optimization problems. Inf. Process. Manag. 2022, 59, 102854. [Google Scholar] [CrossRef]
  36. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  37. Chauhan, P.; Deep, K.; Pant, M. Novel inertia weight strategies for particle swarm optimization. Memetic Comput. 2013, 5, 229–251. [Google Scholar] [CrossRef]
  38. Taherkhani, M.; Safabakhsh, R. A novel stability-based adaptive inertia weight for particle swarm optimization. Appl. Soft Comput. 2016, 38, 281–295. [Google Scholar] [CrossRef]
  39. Gu, Y.; Lu, H.; Xiang, L.; Shen, W. Adaptive simplified chicken swarm optimization based on inverted s-shaped inertia weight. Chin. J. Electron. 2022, 31, 367–386. [Google Scholar] [CrossRef]
  40. Li, J.; An, Q.; Lei, H.; Deng, Q.; Wang, G.-G. Survey of Lévy flight-based metaheuristics for optimization. Mathematics 2022, 10, 2785. [Google Scholar] [CrossRef]
  41. Luo, W.; Wu, H.; Peng, J. Improvement of electric fish optimization algorithm for standstill label combined with Levy flight strategy. Biomimetics 2024, 9, 677. [Google Scholar] [CrossRef]
  42. Xiong, J.; Liang, W.; Liang, X.; Yao, J. Intelligent quantification of natural gas pipeline defects using improved sparrow search algorithm and deep extreme learning machine. Chem. Eng. Res. Des. 2022, 183, 567–579. [Google Scholar] [CrossRef]
  43. Osorio, F.; Galea, M.; Henr´ıquez, C.; Arellano-Valle, R. Addressing non-normality in multivariate analysis using the t-distribution. AStA Adv. Stat. Anal. 2023, 107, 785–813. [Google Scholar] [CrossRef]
  44. Suganthan, P.N.; Hansen, N.; Liang, J.J.; Deb, K.; Chen, Y.-P.; Auger, A.; Tiwari, S. Problem definitions and evaluation criteria for the cec 2005 special session on real-parameter optimization. KanGAL Rep. 2005, 2005005, 2005. [Google Scholar]
  45. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  46. Pang, H.; Lin, A.; Holford, M.; Enerson, B.E.; Lu, B.; Lawton, M.P.; Floyd, E.; Zhao, H. Pathway analysis using random forests classification and regression. Bioinformatics 2006, 22, 2028–2036. [Google Scholar] [CrossRef]
  47. Salman, H.A.; Kalakech, A.; Steiti, A. Random forest algorithm overview. Babylon. J. Mach. Learn. 2024, 2024, 69–79. [Google Scholar] [CrossRef]
  48. Naghibi, S.A.; Ahmadi, K.; Daneshi, A. Application of support vector machine, random forest, and genetic algorithm optimized random forest models in groundwater potential mapping. Water Resour. Manag. 2017, 31, 2761–2775. [Google Scholar] [CrossRef]
  49. Probst, P.; Wright, M.N.; Boulesteix, A.-L. Hyperparameters and tuning strategies for random forest. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2019, 9, 1301. [Google Scholar] [CrossRef]
  50. Zumbado-Corrales, M.; Esquivel-Rodríguez, J. Evoseg: Automated electron microscopy segmentation through random forests and evolutionary optimization. Biomimetics 2021, 6, 37. [Google Scholar] [CrossRef] [PubMed]
  51. Zhou, J.; Huang, S.; Qiu, Y. Optimization of random forest through the use of MVO, GWO and MFO in evaluating the stability of underground entry-type excavations. Tunn. Undergr. Space Technol. 2022, 124, 104494. [Google Scholar] [CrossRef]
  52. Wang, M.; Zhao, G.; Wang, S. Hybrid random forest models optimized by sparrow search algorithm (SSA) and harris hawk optimization algorithm (HHO) for slope stability prediction. Transp. Geotech. 2024, 48, 101305. [Google Scholar] [CrossRef]
  53. Grimaldi, M.; Cricelli, L. Indexes of patent value: A systematic literature review and classification. Knowl. Manag. Res. Pract. 2020, 18, 214–233. [Google Scholar] [CrossRef]
Figure 1. The function curve for inertia weight.
Figure 1. The function curve for inertia weight.
Biomimetics 10 00561 g001
Figure 2. The convergence diagram of five algorithms.
Figure 2. The convergence diagram of five algorithms.
Biomimetics 10 00561 g002aBiomimetics 10 00561 g002b
Figure 3. The ROC Curve before and after optimization by EFHO and the dashed line represents the random classifier baseline (AUC = 0.5).
Figure 3. The ROC Curve before and after optimization by EFHO and the dashed line represents the random classifier baseline (AUC = 0.5).
Biomimetics 10 00561 g003
Figure 4. Feature importance of patent indicators in the EFHO-RF model.
Figure 4. Feature importance of patent indicators in the EFHO-RF model.
Biomimetics 10 00561 g004
Table 1. The performance comparison results.
Table 1. The performance comparison results.
ResultsPSOFHOGWOWOAEFHO
F1Best5.86 × 10−11.4875 × 10−836.5908 × 10−291.0343 × 10−840
Mean2.17 × 1001.3582 × 10−691.1309 × 10−277.9071 × 10−730
Std8.76 × 10−17.0382 × 10−691.5143 × 10−274.1528 × 10−720
F2Best2.69 × 1003.0352 × 10−211.5109 × 10−175.31 × 10−580
Mean4.52 × 1002.4074 × 10−198.9556 × 10−171.0498 × 10−500
Std1.24 × 1003.3718 × 10−195.7898 × 10−174.4941 × 10−500
F3Best6.92 × 1015.4193 × 10−829.7792 × 10−91.59 × 1040
Mean1.83 × 1021.6619 × 10−694.6808 × 10−64.38 × 1040
Std5.94 × 1015.1581 × 10−697.6958 × 10−51.40 × 1040
F4Best1.48 × 1003.2607 × 10−351.2102 × 10−71.78 × 1000
Mean2.06 × 1001.0124 × 10−307.4167 × 10−75.33 × 1010
Std2.71 × 10−13.5189 × 10−306.7046 × 10−72.79 × 1010
F5Best2.79 × 1023.52 × 10−22.61 × 1012.71 × 1012.11 × 10−3
Mean1.10 × 1032.80 × 10−12.71 × 1012.80 × 1019.58 × 10−2
Std7.26 × 1022.09 × 10−17.51 × 10−14.92 × 10−19.21 × 10−2
F6Best9.05 × 10−17.85 × 10−32.50 × 10−16.70 × 10−25.21 × 10−3
Mean2.26 × 1008.94 × 10−18.35 × 10−13.75 × 10−11.48 × 10−2
Std8.27 × 10−11.66 × 1003.20 × 10−12.09 × 10−18.00 × 10−3
F7Best2.83 × 1002.30 × 10−42.43 × 10−46.7084 × 10−55.5551 × 10−7
Mean1.67 × 1019.06 × 10−41.70 × 10−32.14 × 10−33.6111 × 10−5
Std1.37 × 1014.29 × 10−49.03 × 10−41.99 × 10−33.6888 × 10−5
F8Best4.31 × 1033.61 × 1004.89 × 1036.76 × 10−11.34 × 10−2
Mean6.40 × 1033.93 × 1026.54 × 1032.32 × 1032.61 × 101
Std1.46 × 1037.02 × 1026.88 × 1021.73 × 1033.40 × 101
F9Best9.46 × 10105.6843 × 10−1400
Mean1.60 × 10201.93 × 10000
Std3.98 × 10103.31 × 10000
F10Best2.06 × 1004.4409 × 10−167.1498 × 10−144.4409 × 10−164.4409 × 10−16
Mean2.78 × 1004.4409 × 10−169.6959 × 10−143.4047 × 10−154.4409 × 10−16
Std3.75 × 10−10 × 1001.4564 × 10−142.2625 × 10−150
F11Best3.07 × 10−20000
Mean1.48 × 10−105.12 × 10−300
Std4.71 × 10−208.16 × 10−300
F12Best5.78 × 10−34.95 × 10−46.43 × 10−35.25 × 10−33.53 × 10−3
Mean3.31 × 10−21.47 × 10−34.08 × 10−22.31 × 10−22.35 × 10−2
Std2.57 × 10−26.48 × 10−42.23 × 10−21.13 × 10−26.51 × 10−3
F13Best2.26 × 10−13.96 × 10−34.08 × 10−11.19 × 10−17.9238 × 10−9
Mean5.11 × 10−11.08 × 10−26.87 × 10−15.22 × 10−15.9291 × 10−7
Std2.13 × 10−14.94 × 10−32.38 × 10−12.44 × 10−16.9933 × 10−7
F14Best0 × 1001.7555 × 10−81.8985 × 10−111.0154 × 10−111.98 × 100
Mean2.00 × 1005.71 × 10−13.46 × 1001.83 × 1006.12 × 100
Std2.70 × 1006.20 × 10−14.35 × 1002.96 × 1001.31 × 100
F15Best1.34 × 10−41.3838 × 10−56.009 × 10−101.667 × 10−61.06 × 10−4
Mean5.91 × 10−49.51 × 10−42.71 × 10−33.87 × 10−49.25 × 10−4
Std1.48 × 10−41.94 × 10−36.80 × 10−33.76 × 10−46.42 × 10−4
F16Best2.2204 × 10−152.1008 × 10−73.8243 × 10−106.9722 × 10−147.1983 × 10−6
Mean2.4351 × 10−159.9372 × 10−62.3649 × 10−81.1901 × 10−93.53 × 10−3
Std3.9858 × 10−171.0288 × 10−52.6884 × 10−84.2983 × 10−93.65 × 10−3
F17Best1.6653 × 10−162.4405 × 10−57.4074 × 10−81.7071 × 10−114.7274 × 10−5
Mean1.6653 × 10−163.77 × 10−42.1562 × 10−65.8586 × 10−66.94 × 10−3
Std0 × 1003.50 × 10−42.0217 × 10−61.1906 × 10−57.36 × 10−3
F18Best2.2204 × 10−151.2711 × 10−51.4177 × 10−77.6656 × 10−83.0105 × 10−6
Mean1.0214 × 10−141.23 × 10−33.0221 × 10−55.6182 × 10−51.03 × 10−1
Std4.704 × 10−151.08 × 10−34.3207 × 10−51.01 × 10−43.54 × 10−1
F19Best3.9968 × 10−152.34 × 10−41.9841 × 10−63.8805 × 10−76.17 × 10−3
Mean5.1514 × 10−151.59 × 10−21.09 × 10−35.90 × 10−31.23 × 10−1
Std1.907 × 10−155.06 × 10−22.22 × 10−37.07 × 10−31.32 × 10−1
F20Best7.0895 × 10−101.96 × 10−21.582 × 10−65.8291 × 10−53.13 × 10−1
Mean3.96 × 10−21.25 × 10−17.13 × 10−28.22 × 10−27.69 × 10−1
Std5.60 × 10−21.11 × 10−18.25 × 10−29.90 × 10−23.33 × 10−1
F21Best1.7764 × 10−151.07 × 10−14.06 × 10−44.41 × 10−41.77 × 10−4
Mean3.34 × 1003.78 × 10−11.10 × 1002.30 × 1008.31 × 10−1
Std3.44 × 1001.93 × 10−12.22 × 1002.86 × 1007.97 × 10−1
F22Best0 × 1007.39 × 10−24.29 × 10−43.47 × 10−41.98 × 10−2
Mean7.87 × 10−14.85 × 10−11.73 × 10−32.09 × 1008.65 × 10−1
Std2.03 × 1002.23 × 10−17.32 × 10−42.74 × 1007.89 × 10−1
F23Best8.3489 × 10−141.13 × 10−19.8691 × 10−51.20 × 10−31.08 × 10−2
Mean1.24 × 1005.42 × 10−11.82 × 10−12.81 × 1001.12 × 100
Std2.51 × 1002.95 × 10−19.70 × 10−13.25 × 1009.19 × 10−1
Table 2. Summary of Wilcoxon signed-rank test results.
Table 2. Summary of Wilcoxon signed-rank test results.
ComparisonFunctions with p < 0.05 (n/23)Proportion (%)
EFHO vs. FHO9/2339.13
EFHO vs. PSO14/2360.87
EFHO vs. GWO13/2356.52
EFHO vs. WOA12/2352.17
Table 3. The EFHO test results of high-dimensional problems.
Table 3. The EFHO test results of high-dimensional problems.
EFHO
500 Dimensions1000 Dimensions
MeanStdMeanStd
F10000
F20000
F30000
F40000
F57.76 × 1004.09 × 1019.48 × 10−21.28 × 10−1
F66.61 × 10−36.66 × 10−31.72 × 10−29.96 × 10−3
F73.3296 × 10−54.4944 × 10−52.9348 × 10−52.603 × 10−5
F81.65 × 1011.18 × 1012.55 × 1013.47 × 101
F90000
F104.4409 × 10−1604.4409 × 10−160
F110000
F121.62 × 10−24.78 × 10−32.64 × 10−25.58 × 10−3
F135.5754 × 10−61.2471 × 10−52.8439 × 10−74.072 × 10−7
Table 4. The selected patent value indicators.
Table 4. The selected patent value indicators.
No.Abbr.Description
1BCNThe backward citation number
2FCNThe forward citation number
3PANThe applicants number
4INNThe inventors number
5ASGNThe patent holders number
6CLMSNThe claims number
7FNThe patent family number
8NCSEThe country’s number belongs to the same family
9NRThe reexaminations number
10ITThe invalidation times
11LIQThe licensing frequency
12RAQThe assignments number
13PLQThe pledging frequency
14LTNThe lawsuits number
Table 5. Test performance comparison.
Table 5. Test performance comparison.
AlgorithmAccuracyRecallPrecisionF1 MeasureAUC
NB0.9180.9100.8740.8910.950
LR0.9500.8980.9530.9250.966
RF0.9580.9180.9670.9420.988
SVM0.9430.8800.9640.9200.977
BP0.9470.8830.9710.9250.966
EFHO-BP0.9590.9500.9630.9600.985
EFHO-RF0.9660.9580.9700.9630.990
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, X.; Li, H.; Wang, S. High-Value Patents Recognition with Random Forest and Enhanced Fire Hawk Optimization Algorithm. Biomimetics 2025, 10, 561. https://doi.org/10.3390/biomimetics10090561

AMA Style

Yao X, Li H, Wang S. High-Value Patents Recognition with Random Forest and Enhanced Fire Hawk Optimization Algorithm. Biomimetics. 2025; 10(9):561. https://doi.org/10.3390/biomimetics10090561

Chicago/Turabian Style

Yao, Xiaona, Huijia Li, and Sili Wang. 2025. "High-Value Patents Recognition with Random Forest and Enhanced Fire Hawk Optimization Algorithm" Biomimetics 10, no. 9: 561. https://doi.org/10.3390/biomimetics10090561

APA Style

Yao, X., Li, H., & Wang, S. (2025). High-Value Patents Recognition with Random Forest and Enhanced Fire Hawk Optimization Algorithm. Biomimetics, 10(9), 561. https://doi.org/10.3390/biomimetics10090561

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop