A Cross-Efficiency Evaluation Method Based on Evaluation Criteria Balanced on Interval Weights

: Cross-efficiency evaluation approaches and common set of weights (CSW) approaches have long been suggested as two of the more important and effective methods for the ranking of decision making units (DMUs) in data envelopment analysis (DEA). The former emphasizes the flexibility of evaluation and its weights are asymmetric, while the latter focuses on the standardization of evaluation and its weights are symmetrical. As a compromise between these two approaches, this paper proposes a cross-efficiency evaluation method that is based on two types of flexible evaluation criteria balanced on interval weights. The evaluation criteria can be regarded as macro policy—or means of regulation—according to the industry’s current situation. Unlike current cross-efficiency evaluation methods, which tend to choose the set of weights for peer evaluation based on certain preferences, the cross-efficiency evaluation method based on evaluation criterion determines one set of input and output weights for each DMU. This is done by minimizing the difference between the weights of the DMU and the evaluation criteria, thus ensuring that the cross-evaluation of all DMUs for evaluating peers is as consistent as possible. This method also eliminates prejudice and arbitrariness from peer evaluations. As a result, the proposed cross-efficiency evaluation method not only looks for non-zero weights, but also ranks efficient DMUs completely. The proposed DEA model can be further extended to seek a common set of weights for all DMUs. Numerical examples are provided to illustrate the applications of the cross-efficiency evaluation method based on evaluation criterion in DEA ranking.


Introduction
Data envelopment analysis (DEA) is a practical methodology originally proposed by Charnes et al. [1]. Since that time, DEA has been widely studied and applied all over the world, and the method has been further developed and expanded by many scholars. In recent years, the outstanding studies in DEA can be found in the literature [2][3][4][5][6][7][8]. DEA is used to evaluate the performance of a group of decision making units (DMUs) using multiple inputs to produce multiple outputs. DEA method requires each decision-making unit to evaluate its efficiency and assign the most favorable weight to itself. In addition, efficiency is the optimistic efficiency of DMU, which should not be greater than 1. If the efficiency value of the decision unit is equal to 1, the DMU is called a DEA efficient DMU. Otherwise, DMU is considered to be a non DEA efficient DMU.
The traditional DEA method has two major drawbacks. The first is a lack of discrimination, and the second is the existence of unrealistic weights. The DEA method allows each DMU to evaluate its efficiency with the most favorable weights. In this way, more than one DMU is often evaluated as DEA efficient, and these DMUs cannot be further distinguished. Therefore, the lack of discrimination is one of the main defects of the DEA method. This also leads to another important problem. The input and output that is beneficial to a particular DMU will be weighted heavily, while the input and output that is unfavorable to the DMU will be weighted lightly, or even ignored. As a result, weighting for self-assessment can sometimes be unrealistic.
Studies to overcome the weakness of DEA's discrimination power are grouped into two trends. One remedy is the cross-efficiency method suggested by Sexton et al. [9], which introduces a secondary goal. The most commonly used methods include the benevolent and aggressive cross-efficiency assessment proposed by Doyle and Green [10], both of which are calculated using the weights that are benevolent or aggressive to peers. Wang & Chin [11] proposed a neutral cross-efficiency evaluation method, in which the attitude of decision makers is neutral and there is no need to make a choice between benevolent and aggressive formulas. Liang et al. [12] put forward the game cross-efficiency evaluation method. Using the idea of game, each DMU is regarded as an independent player, and bargains between the optimistic efficiency. In addition, Jahanshahloo et al. [13] proposed symmetrical weight distribution technology. This technique rewards the DMU with symmetrical selection weights. Wu et al. [6] proposed a DEA model with balanced weight. The second goal is to reduce the number of zero weights and the large differences in weighted data. Ruiz [14] proposed the cross-efficiency evaluation of direction distance function for fractional programming. Cook and Zhu [15] proposed a unit invariant multiplicative DEA model, which can directly obtain the maximum and unique cross-efficiency scores of each DMU. Wu [7] proposed to use the target recognition model as a means to obtain the reachable targets of all DMUs. Several secondary objective models are proposed for weight selection. These models consider the expected and unexpected cross-efficiency goals of all DMUs. Other cross-efficiency evaluation methods are discussed in Wu and Chu [8], Oral et al. [16], Oukil [17], Carrillo [18], and Shi et al. [19].
Another remedy is the common set of weights (CSW) approach in DEA, which was first suggested by Cook et al. [20]. This method utilizes the idea of common weights to measure the relative efficiency of highway maintenance patrols. Years later, this study was further developed by Jahanshahloo et al. [21], Kao & Hung [22], and Liu & Peng [23]. In more recent studies, Amir et al. [24] proposes a novel TCO-based model in which a common set of weights imprecise DEA (CSW-IDEA) is used to address the managerial and technical issues of handling weighting schemes and imprecise data. Hossein et al. [25] suggests a novel method for determining the CSWs in a multi-period DEA. The CSWs problem is formulated as a multi-objective fractional programming problem. Then, a multi-period form of the problem is formulated, and the mean efficiency of the DMUs is maximized, while their efficiency variances are minimized. The CSW approaches have been developed to find a common set of weights for all DMUs, in order to overcome the shortcomings of the weights flexibility method, where each DMU can take its own most desirable weight.
From the literature review above, all the cross-efficiency evaluation methods are formulated so that each DMU chooses one set of weights determined by the CCR model(self-evaluation model proposed by Charnes, Cooper and Rhodes) that has alternate optima solutions. When a DMU evaluates its peers, the DMU selects one set of ideal weights from alternative weights by means of the quadratic optimization of the target function from various angles. In other words, cross-evaluation is an evaluation method, which determines a set of weights for each DMU, in order to rate itself and its peers in consideration of the diversity of DMUs. The common weight evaluation method, which determines a common set of weights as the common evaluation criteria, is used to evaluate each DMU, without considering the flexibility of the DMUs, and belongs to a non-differentiating evaluation method.
In terms of practical applications, differences between homogeneous DMUs still exist, such as scale, history, culture, and region. Common weight evaluation is obviously unfair, because this method does not take into account the differences of the DMUs. However, there are also shortcomings in cross-evaluation, such as the fact that each DMU excessively enlarges the weights of its own superiority indicators and ignores the importance of input-output indicators, thus forming unrealistic and subjective evaluation conclusions. The combination of the two methods discussed above is more meaningful. That is, each DMU can take its own desirable weight, which is obtained under the constraints of an objective criterion. Therefore, we propose the cross-evaluation method based on evaluation criteria in this paper.
In this paper, we propose a series of DEA models for cross-efficiency evaluation based on evaluation criteria. The evaluation criteria may be formed based on the overall situation of the industry, or the performance of some representative enterprises. Accordingly, two evaluation criteria are proposed, which are balanced on the interval weights of input-output variables. One is based on the eclectic decision-making method, which takes the aggregation of the minimum of upper limit of interval weights and maximum of lower limit of interval weights. The harmonic coefficient α is introduced into the eclectic decision-making method to increase the flexibility of evaluation criteria. The other is based on weighted mathematical expectation. Because the importance of each DMU is different in cross-evaluation, we introduce the parameter ( 1,2,..., ) j p j n = into the evaluation criteria as a weight that reflects the position of a DMU.
Mathematical expectations are weighted and added by j p to form the evaluation criteria. Then, the proposed method based on an evaluation criterion determines one set of input and output weights for each DMU. This is done by minimizing the deviations of input and output weights for peer evaluation from a standard criterion. In this way, aside from reducing zero weights, the weights for peer-evaluation (which are closer and more concentrated) are more realistic to peers.
The rest of the paper is organized as follows. Section 2 describes the cross-efficiency evaluations, mainly including aggressive and benevolent formulations. The evaluation criterion balanced on interval weights is developed in Section 3. The DEA models for cross-efficiency evaluation based on evaluation criteria are extended in Section 4. Numerical examples are demonstrated in Section 5. Conclusions are offered in Section 6.

The Efficiency Evaluation
Suppose there are n DMUs to be evaluated against m inputs and s outputs.
which aims to find a set of input and output weights that are most favourable to DMUk. The Charnes and Cooper transformation can be equivalently transformed into the linear program (LP) below for the solution: (2) subject to Model (2) is solved n times, with a different DMU being solved each time. As a result, n DMUs will have n sets of input and output weights, and each DMU will have (n − 1) cross-efficiency, plus a CCR-efficiency, which together form a cross-efficiency matrix. The average cross-efficiency for  Note that model (2) may have multiple optimal solutions. If the input and output weights are not unique, the use of cross efficiency evaluation will be destroyed. In order to solve this problem, a remedy proposed by Sexton et al. [9] is to introduce a secondary goal, one which optimizes the input and output weights while maintaining the CCR-efficiency determined by model (2). Doyle and green [10] proposed the most commonly used secondary goals, as follows: subject to Model (3) is called the aggressive formula of cross-efficiency evaluation, which aims to minimize the cross-efficiency of peers in some way. Instead, model (4) is called the benevolence formula for cross-efficiency evaluation. To some extent, the model improves the cross-efficiency of other EMUs. These two models optimize input and output weights in two different ways. Therefore, there is no guarantee that they can lead to the same efficiency ranking or conclusion of N DMUs.
In addition, other secondary goals or models are mentioned in the DEA literature. For interested readers, please refer to Sexton et al. [9], Liang et al. [12], Wang et al. [26,27], and Jahanshahloo et al. [13]. These secondary goals focus on how to uniquely determine input and output weights. In the next section, we focus on the diversity of input and output weights, and develop some alternative DEA models to minimize the weights differences used to evaluate peers. This enables cros-efficiency to be evaluated with more reasonable input and output weights.

Evaluation Criteria Balanced on Interval Weights of N DMUs
From our perspective, when a DMU is given an opportunity to unilaterally decide upon a set of input and output weights for evaluating peers, in addition to being as favourable as possible to itself, the DMU tends have specific preferences when choosing the set of weights. This preferential choice of weights leads to an unfair and arbitrary situation for peers. Therefore, we need to establish some evaluation criteria to ensure the relative consistency of cross-evaluation for peers by eliminating prejudice. That is, we not only look for non-zero weights, but we also seek to ensure that weights for evaluating peers are as close as possible by taking a certain evaluation criterion as a reference point. That means we seek to minimize the difference between the weights of each DMU and the evaluation criteria. Two evaluation criteria are proposed as follows:

The DEA Modes of Interval Weights
For a DMU, we get a set of maximum weights from among the alternate optima in the CCR model by maximizing the weight of each variable (including input and output variables). By contrast, a set of minimum weights is obtained by minimizing the weight of each variable. Consider an efficient DMU, say, DMUk. The maximum attainable value of rk u or ik v of the DMUk can be obtained by solving the following LP model: subject to  , and then n DMUs lead to n interval weights

Evaluation Criteria Based on the Eclectic Decision-Making Method
There are multiple methods in the criteria for evaluating peers based on the interval weights. The optimistic decision method would consider the upper limit of the interval weights as a peer evaluation criterion. The pessimistic decision method can also take the lower limit as a peer evaluation criterion. However, the decision makers cannot be absolutely optimistic or pessimistic; they are more likely to be between the two. Therefore, the eclectic decision-making method is introduced as peer evaluation criteria. We can use obtained as an evaluation criterion, which in turn is a balance between the maximum and the minimum attainable values of n DMUs in rj u and ij v . This ensures the evaluation criterion meets a variety of application requirements. This is the first evaluation criterion, which is based on the eclectic decision-making method, or ECED for short:

Evaluation Criterion Based on Weighted Mathematical Expectation
If we denote the optimal solution of the above model (5) by solving model (5) n times would lead to n sets of optimal solutions available for n DMUs, which in turn would form an interval weight matrix (IWM). An IWM is shown as follows: 11 In practical applications, it is unfair and subjective to use the upper or lower bounds of the interval weight by the criterion for peer evaluating. Rather, an application should consider the level that most DMUs can reach. However, we cannot obtain sufficient knowledge of the weight distribution information in the interval. The term probability distribution refers to the probability rule used to express the value of random variables. We may assume that the weights satisfy one or other forms of probability distribution (such as normal distribution or uniform distribution). The mathematical expectation of probability distribution can best represent the interval weight used as a criterion for peer evaluating when that weight obeys the probability distribution. According to the central limit theorem, the random variables approximately obey the normal distribution when the sample size is large enough. Supposing that weights satisfy the standard normal distribution in this paper, the mathematical expectation matrix-based interval weight matrix (IWM) is as follows:  For DMUj, rj μ and ij η may not be a real weight, and this represents the compromise decision for the decision-makers, as an objective evaluation criterion. There are n sets of mathematical expectations representing the evaluation criteria for n DMUs. As the importance of each DMU is different in cross-evaluation, let j p be the weight of DMUj, which embodies the position of DMUj.
Therefore, the evaluation criterion based on weighted mathematical expectation (ECWME) is calculated as follows:

DEA Models for Cross-Efficiency Evaluation Based on Evaluation Criteria
It is well known that each DMU personally chooses the profile of weights to be used in the cross-efficiency evaluation. Therefore, the DMU's choice is often prejudiced. One DMU's attitude towards its peers may be aggressive, benevolent, indifferent, or something else. Those prejudicial attitudes of DMUs to peers need to be avoided in many applications. This is why the weights that are chosen by each DMU for peer evaluations should be based on an evaluation criterion, as stated in part 3. We propose a method that makes a selection between alternate optima of CCR. This is done by, to the greatest extent possible, reducing the degree of deviation of weights for peer evaluation from the evaluation criteria. In other words, our purpose is to look for the profiles of DMU weights that are closest to the evaluation criterion. To do this, a nonlinear programming model is proposed as follows: subject to ) are the variables that need to be solved. The above model needs to be solved n times, one time for each DMU. The purpose of the model is to minimize the deviation of input and output weights from the evaluation criterion. In other words, each DMU obtains one set of weights that is favorable to that DMU and is also as close to the evaluation criterion of the evaluating peers as possible.
Model (8) is a form of nonlinear programming. To make this nonlinear model (8) to the set of constraints. Thus, we minimize the linear objective function rk rk r r rk rk rk r r rk u y u y u y u y If we denote the optimal solution of model (9)  Therefore, the cross-efficiency score of DMUj is the average of these cross-efficiencies: Besides cross-evaluation, the weights solved by model (9) are similar to their evaluation criteria derived from formulae (67). The weights are relatively concentrated, and their coefficient of variation is small, which can further be extended to be a common set of weights (CSW) based on evaluation criteria for all DMUs. The extended CSW deviates little from the cross weights solved by model (9), which are easily accepted by each DMU.
) , solved by model (2), be the optimal weights of inefficient DMUt. Suppose there are 1 n efficient DMUs and 2 n inefficient DMUs. The CSW based on evaluation criteria is obtained as follows: (10b) We consider the inefficient DMU and efficient DMU separately, because, in the case of low efficiency DMU, the weights' distribution provided by the proposed method is the only optimal solution to the weights in the CCR model. This same train of thought can be seen in Nuria Ramón (2011). This paper, however, differs from the idea proposed in Nuria Ramón (2011). The cross-evaluation proposed in this paper allows for the inefficient DMUs that keep the weights in the CCR model. For efficient DMUs, the weights for evaluating peers are obtained by minimizing deviation from the evaluation criterion, rather than reducing the differences between the weights of any two DMUs. In addition, the model proposed by Nuria Ramón needs to be solved 2 2 n times, and this approach is not suitable for a large number of DMUs. The measure of our approach is different. We focus attention on the deviation of input and output weights from the evaluation criterion. The model in this paper only needs to be solved n times and needs to be more practical.

Numerical Examples
Example 1. In this section, we provide one numerical example to illustrate the proposed methods detailed above. We consider the numerical examples with the data presented in Table 1. The case of seven academic departments in a university is presented in Table 1, with three inputs and three outputs.   DMU1  12  400  20  60  35  17  1  DMU2  19  750  70  139  41  40  1  DMU3  42  1500  70  225  68  75  1  DMU4  15  600  100  90  12  17  0.819  DMU5  45  2000  250  253  145  130  1  DMU6  19  730  50  132  45  45  1  DMU7  41  2350  600  305  159  97  1 For a DMU, the minimum output and input weights can be obtained by solving model (5). The result is taken as the upper bound of the interval weights (UBIW). The maximum output and input weights can also be obtained by solving model (5). This result is taken as the lower bound of the interval weights (LBIW). Therefore, each DMU will get a set of interval weights, as shown in Table 2. Firstly, the cross-evaluation efficiency is discussed, based on eclectic decision-making evaluation criterion (ECED). The UBIW and LBIW are obtained by Formulaes (6a) and (6b), as shown in the first and second row of Table 3. To make sure the evaluation criteria fall into the interval weights of all DMUs (as much as possible), the ECED is calculated with α = 0.5, as can be seen in the last row of Table 3. Each DMU attempts to obtain a set of weights that is as close (or as similar as possible) to the cross-evaluation criterion. This is done by minimizing deviation from the cross-evaluation criterion. When the evaluation criterion is the ECED, the cross-evaluation weights of each DMU solved by model (9) are shown in Table 4. Secondly, another criterion is proposed, which is based on weighted mathematical expectation. There is a mathematical expectation regarding the arbitrary weight interval. Then, n sets of mathematical expectations are computed from Table 2 for n DMUs. These are seen in the second to seventh lines in Table 5. In order to be comparable to the evaluation criterion based on eclectic decision-making (ECED), we assume that each DMU has the same status, that is, 1 2 7 ... p p p = = = . Then, the evaluation criterion based on weighted mathematical expectation is shown in the last line of Table 5. In Table 6, we show the weights for peer evaluation as solved by model (9) for the seven academic departments, which are based on the ECWME. When comparing Tables 4 and 6, we can see that the weights of DMU4 have not changed. This is because DMU4, which is an inefficient DMU, retains the weights in the CCR model in the cross-evaluation of this paper. In the coming section, we illustrate how the performance of the proposed approach is better than the classic cross-efficiency evaluation method in reducing zero weights. In Table 7, we show the weights solved by DEA model (2); that is, the CCR model. Cross-evaluation weights under the proposed method are limited by the constraints of the criteria, so the weights in Table 6 fluctuate around the evaluation criteria. The weights of the CCR model have no restriction of evaluation criteria. Each DMU only considers whether the weights are favourable or not, and does not need to consider the reality of the weights. So a large number of zero-value weights often appear in Table 7, which is easy to find by comparing Tables 6 and 7. In  Tables 8 and 9, we show the weights solved by aggressive and benevolent cross evaluation. It is particularly noticeable that the number of zero weights is sharply reduced in the proposed ECED and ECWME-based approach when comparing Tables 4, 6-9.  The cross-efficiency scores and the rankings based on all the methods mentioned in this paper are provided in Table 10. We find that the proposed ECED and ECWME-based method has more discrimination power than CCR-efficiency, and basically provides different ranking with benevolent and aggressive cross-evaluation. This indicates that the proposed approaches in this paper represent a new method, one which can achieve effective ranking for DMUs. In addition, cross-efficiency scores and the rankings of the proposed approach based on ECED are shown in Table 11 when harmonic coefficient α = 0.2, α = 0.5, and α = 0.8. The proposed approach's economic meaning is that each DMU has its own inputs and outputs, which are different from others. This leads to different DMU rankings under different evaluation criteria. This may become a macro policy whose value is high or low, according to the evaluation need, by adjusting α = 0.2, 0.5, or any value between 0 and 1.

CCR-Efficiency
Evaluation criteria are often formulated according to the present industry, which is not the integration of all enterprises, but the embodiment of some representative enterprises. That is, each DMU has a different role in the formation of evaluation criteria. Therefore, we consider two ECWME situations in this example, when p1 = p2 = p3 = p4 = p5 = p6 = p7 and 3p1 = 2 p2 = 6p3 = 6p4 = 6 p5 = 6 p6 = 6 p7. The cross-efficiency scores and rankings based on ECWME under the two combinations of p are shown in Table 12. The coefficient of variation (CV) of the weights of each variable is computed and compared with the CCR model. The proposed approach is based on ECED and ECWME, and the result can be seen in Table 13. As can be seen, the mean of CV of the weights provided by the proposed approach is smaller than that of the weights provided by the CCR model. Also, the proposed ECWME-based approach is smaller than the proposed approach based on ECED. The CV of weights based on ECED and ECWME changes in line with changes in coefficient α and combinations of p.
Obviously, the CV of the weights of the two methods is very small in any case. This means that the weights of each DMU are very similar. These similar weights are further extended to be CSW, under which the score and rank based on ECED and ECWME are shown in Table 14. From the results shown in Table 14, the CSW in any case can be used to achieve the effective ranking of DMUs, and the rankings are different from each other.  Table 14. Score and rank under the common set of weights (CSW) based on ECED and ECWME.  x2: Total expenditure on scientific and technological activities in the current year (in increments of $10,000); x3: Total expenditure on R&D of enterprises (in increments of $10,000); x4: Number of senior technicians and technicians at the end of the year (persons); y1: Sales revenue of new products (services or processes) of enterprises in this year (in increments of 10,000 yuan); y2: Added value of enterprises in this year (in increments of 10,000 yuan); y3: Total profits realized by enterprises in this year (in increments of 10,000 yuan); y4: Total labor productivity of enterprises in this year (in increments of 10,000 yuan/person).

Score and Rank Under ECWME
The letters in the enterprise number are the initials of the name of the area in which the enterprise is located. For example, "PT" in "PT 1" stands for Pu Tian City of Fujian Province.
The application example involves 27 innovative machinery manufacturing enterprises in Fujian Province, located in different cities in Fujian Province. The CCR-efficiency of each enterprise is shown in last column of Table 15. According to CCR-efficiency, 12 out of the 27 sampled enterprises are DEA effective. The interval weights of these 12 DEA-effective enterprises are obtained by solving model (5), as shown in Table A1 of the Appendix A. On the basis of the interval weights (Table A1 of Table A2 of the Appendix A. For the 27 enterprises, if they are evaluated in a non-differentiated way (such as using the common weight evaluation method), then it is not conducive to encouraging enterprises in different regions to carry out innovative activities according to local conditions. However, evaluating these enterprises in a differentiated way (such as using the traditional cross-efficiency evaluation method) allows each enterprise to avoid its own weaknesses and emphasize its own advantages. In this case, the evaluation results will inevitably be unfair. Therefore, differentiated evaluation under evaluation criteria is carried out for these 27 enterprises, that is, cross-efficiency evaluation under the 10 criteria (Table A2 of the Appendix A). The comparison of the evaluation results is shown in Figures 1 and 2.
Evaluation criteria We can see in Figure 1 that the degree of cross-efficiency of the 27 enterprises varies under different criteria, that is, the cross-efficiency of 27 enterprises is generally higher under the evaluation criterion (minimum criterion) when α = 1, and lower under the evaluation criterion (highest criterion) when α = 0.1. This result shows that the evaluation criteria have a supervisory and regulatory effect on the cross-efficiency of enterprises. When the evaluation criteria are lower, the weights space available for enterprises is larger, and the enterprises can select more favorable weights for themselves. Therefore, the cross-efficiency is generally higher. Conversely, when the weights space is smaller, the cross-efficiency is smaller. This conclusion is consistent with the application. The lower the evaluation criteria, the higher the performance of the enterprises will be.
According to Figure 2, the evaluation criteria also have an impact on the rankings of the 27 enterprises, and the rankings under almost every criterion are different. At the same time, the ranking of the 27 enterprises under different criteria has not changed to any significant degree. In particular, the ranking of the top enterprises and the bottom enterprises is relatively stable. This result shows that the rankings obtained under the proposed method can basically reflect the basic strength of enterprises.
By introducing the parameter p, the ECWME above takes into account the importance of DMUs in the criterion formulation. In order to highlight the changing regularity of cross-efficiency under different ECWME, we try to simplify the evaluation criteria. That is, one evaluation criterion only considers the importance of one DEA-effective enterprise. Therefore, the weighted mathematical expectation of 12 DEA-effective enterprises forms 12 evaluation criteria by adjusting the parameter p (see Table A3 of the Appendix A). Also, 12 ECWMEs are shown in Table A4 of the Appendix A. The cross-efficiency comparison chart of the 27 sampled machinery manufacturing enterprises based on 12 ECWME is shown in Figure 3, and the ranking comparison chart is shown in Figure 4. In Figure 3, the cross-efficiency of each DMU changes to different degrees under different standards, but the trend is basically the same. This shows that the evaluation criteria have a regulatory effect on the cross-efficiency. However, their decisive effect on the cross efficiency is the enterprise's own performance. In addition, in Figure 3, when DMU-14 is used as the evaluation criterion, the cross efficiency of each DMU is higher. Combined with DMU-14 (see Appendix A Table A4 for bid winning), its value is lower than other evaluation criteria; it can be seen that the lower the evaluation criterion, the higher the cross efficiency of the evaluated DMU. In Figure 4, we use red dots to mark the ranking of 12 DEA-effective enterprises under 12 ECWMEs (see Table A4 of the Appendix A). For example, a small red dot is used to mark the ranking of a DEA-effective enterprise denoted as DMU5 under ECWME, denoted as DMU-5. This only considers the importance of DMU5. From the red dot distribution in Figure 4, it is not difficult to find that the ranking of DMU5 under ECWME DMU-5 is higher than the ranking under other criteria. The conclusion is also valid for other DEA-efficient enterprises, such as DMU6, DMU7, DMU11, and so on. This conclusion is reached because the weights for the enterprises (DMU5, DMU6, DMU7, DMU11, and so on) to be selected are subject to less conditional constraints (under ECWME, such as DMU-5, DMU-6, DMU-7, DMU-11, and so on) in cross-evaluation. That is, there is more weights room for the enterprises to choose, and each enterprise can obtain more conducive weights for themselves. This result is also consistent with the actual application. That is, the enterprise that is used as the evaluation criterion has more advantages in the comprehensive evaluation. Of course, few evaluation criteria in practical application only consider the importance of a certain enterprise. Often, some mainstream enterprises or representative enterprises are taken as evaluation criteria, in order to enhance the objectivity of the evaluation results and avoid the issue of evaluation subjectivity.    Ranking   DMU1  DMU2  DMU3  DMU4  DMU5  DMU6  DMU7  DMU8  DMU9  DMU10  DMU11  DMU12  DMU13  DMU14  DMU16  DMU15  DMU17  DMU18  DMU19  DMU20  DMU21  DMU22  DMU23  DMU24  DMU25  DMU26  DMU27 Evaluation criteria From the analysis of the results of these examples, we can see that, firstly, the level of evaluation criteria will affect the ranking of an enterprise's performance. Low criteria will help to implement enterprise incentive policies, and high criteria will help to macro-control enterprises. Secondly, enterprises ranked higher when considered by evaluation criteria. Taking multiple enterprises as evaluation criteria is conducive to setting up market benchmarks, and taking the vast majority of enterprises as evaluation criteria is conducive to creating a fair and free competitive market environment.

Conclusions
Cross-efficiency evaluation is an important method for ranking DMUs. Existing DEA models for cross-efficiency evaluation tend to choose the set of weights for peer evaluating using subjective attitudes, without an objective evaluation criterion as a reference point for peer evaluation. This makes it difficult for the DM to make a subjective DEA ranking.
To resolve these problems, we have proposed in this paper a cross-efficiency evaluation method based on evaluation criterion balanced on interval weights. The DEA model determines one set of weights for each DMU to evaluate its peers. This is done by minimizing the distance from the CCR-weights of the DMU to the evaluation criteria, which is balanced on interval weights. The criteria standard is different from the original intention of the evaluation, and the corresponding ranking based on cross-efficiency evaluation is also different. On the basis of the interval weights, this paper proposes two types of flexible evaluation criteria. One is based on the evaluation criteria that, in turn, are based on the eclectic decision-making method. Also, changing the harmonic coefficient α can be done to adjust the evaluation criteria. The other criterion is based on the evaluation criteria of mathematical expectation, taking into account the importance of each DMU in the standard formulation by introducing the parameter P. As a result, the cross-efficiencies computed using this method are more objective and flexible, thus meeting the requirements of macro regulation. We have also extended the DEA model and proposed a cross-weight evaluation, which seeks a common set of weights for all DMUs. This method's usefulness has been illustrated with numerical examples.
From the results of the illustrative example, it is particularly noticeable that the number of zero weights is sharply reduced and using cross-evaluation weights is more objective. This method avoids the situation where decision-makers try, for some purpose, to choose weights that are too subjective. In addition, the proposed approach can lead to different DMU rankings under different evaluation criteria and has more discrimination power than the CCR-efficiency method. The cross-evaluation criteria in the manuscript can be regarded as a means of macro-control, which is derived from the real market situation, and which is applied to macro-control market trends at the same time. Under the market environment, whether an industry or a chain enterprise, there should be established industry evaluation criteria or enterprise management objectives, and performance evaluation can be carried out by reference to the evaluation criteria or management objectives. This paper effectively solves these two problems. Firstly, the two evaluation criteria proposed in this paper provide feasible means and methods for industry criteria or enterprise management objectives; then, the cross-evaluation method based on evaluation criteria provides methods and theoretical support for performance evaluation, taking industry criteria or enterprise management objectives as references. Therefore, the proposed approach can be effectively applied to different evaluation problems, such as enterprise performance evaluation, school management, and the macro-control of banks. This work has several limitations that should be improved in future research. Firstly, the proposed approach assumes that DMUs are homogeneous, which limits the method's application scope. Secondly, the criteria for macro-control should be based on a large sample, but this paper uses a small sample. Therefore, readers interested in this research can expand the approach by combining DEA and statistical methods that consider the heterogeneity of decision-making units (DMUs).