An Advanced Decision Making Framework via Joint Utilization of Context-Dependent Data Envelopment Analysis and Sentimental Messages

: Compared to widely examined topics in the related literature, such as ﬁnancial crises/ difﬁculties in accurate prediction, studies on corporate performance forecasting are quite scarce. To ﬁll the research gap, this study introduces an advanced decision making framework that incorporates context-dependent data envelopment analysis (CD-DEA), fuzzy robust principal component analysis (FRPCA), latent Dirichlet allocation (LDA), and stochastic gradient twin support vector machine (SGTSVM) for corporate performance forecasting. Ratio analysis with the merits of easy-to-use and intuitiveness plays an essential role in performance analysis, but it typically has one input variable and one output variable, which is unable to appropriately depict the inherent status of a corporate’s operations. To combat this, we consider CD-DEA as it can handle multiple input and multiple output variables simultaneously and yields an attainable target to analyze decision making units (DMUs) when the data present great variations. To strengthen the discriminant ability of CD-DEA, we also conduct FRPCA, and because numerical messages based on historical principles normally cannot transmit future corporate messages, we execute LDA to decompose the accounting narratives into many topics and preserve those topics that are relevant to corporate operations. Sequentially, the process matches the preserved topics with a sentimental dictionary to exploit the hidden sentiments in each topic. The analyzed data are then fed into SGTSVM to construct the forecasting model. The result herein reveals that the introduced decision making framework is a promising alternative for performance forecasting.


Introduction
The global financial tsunami that first erupted in the middle of 2007 has been widely acknowledged as the most serious financial crisis since the Great Depression [1][2][3]. Due to ever-increasing financial defaults, market participants have raised considerable attention on risk management in corporate finance and accounting studies. Peters et al. [4] pointed out that the main trigger for financial defaults/crises is bad operating performance-that is, a corporate with bad operating performance usually has insufficient resources and weak risk-absorbing ability to properly respond to financial troubles. This situation also explains why a firm with bad operating performance usually has a higher chance of being bought out or being liquidated. Thus, how to evaluate a corporate's operating performance turns out to be an essential and urgent requirement in today's highly turmoil economic atmosphere.
Ratio analysis with the benefits of easy-to-use and being highly intuitive is one of the most widely employed measures for evaluating corporate operations [5,6]. However, the structure of ratio analysis has one input variable and one output variable (such as return on assets (ROA) = net income/total assets), making it rather complicated to depict the real picture of a corporate's operations under today's highly uncertain and intense business atmosphere. Data envelopment analysis (DEA) has already proved itself as a very wellknown linear programming and management science method for handling performance measure tasks [7,8]. It not only can take multiple input and multiple output variables simultaneously without a presetting production function, but it also can be extended into dissimilar structures, thus increasing its application fields. This technique still has some weaknesses, such as too many variables inserted into a model will impede its discriminant ability. To combat this, Adler and Golany [9] applied principal component analysis (PCA) to identify an uncorrelated linear combination of inputs and outputs and then inserted the essential principal components (PCs) into DEA (PCA-DEA) to prevent arriving at an inappropriate judgment.
The challenge of PCA is that it suffers from sensitivity to extreme data and is limited to a linear structure [10][11][12]. In other words, if the data are contaminated by extreme values, then DEA cannot identify their real semantic structure. To solve this problem, one may consider fuzzy robust principal component analysis (FRPCA) with the function of modifying the covariance matrix calculation procedure, which can reduce the impacts caused by extreme values [13]. It is able to convert high-dimensional data into low-dimensional data, thus increasing comprehensibility as well as reducing the storage requirement. Dimensionality reduction is an essential data pre-processing technique in machine learning and pattern recognition. It helps analyze high-dimensional datasets by mapping data from high dimensional feature space to data in low dimensional feature space and tries to preserve as much as possible the relationships among data samples without distorting inherent data semantics in the traditional feature space [14]. Especially, with the rapid advancement and development of information technology and Internet, high dimension, distributed and dynamic complex data are promptly proliferated [15,16]. However, too many information surrounded by decision makers will confuse their decision making process and even worse lead to a biased judgment. To combat this, dimensionality reduction with the merits of increasing model's separation ability, removing irrelevant information, and eliminating computational burden is considered in this study [17]. Seiford and Zhu [18] indicated that adding or deleting an inefficient decision making unit (DMU) into the DEA model will not modify the position of the decision frontier/surface. The decision frontier/surface changes only if the structure of the decision frontier/surface is altered. Based on the theory of consumer choice, consumers' judgements are normally influenced by a certain context-e.g., a circle looks large when so many small circles surround it; or a circle looks small when so many large circles surround it [18,19]. In order to take this construct within the DEA model into consideration, one may employ context-dependent DEA (CD-DEA), which modifies the measuring context developed by alternative options so as to capture the relative attractiveness and progressiveness of each DMU. This method can overcome the obstacle of conventional DEA by generating a direction for an inefficient DMU to become an efficient DMU, but one often ends up with a target that is difficult or even impossible to achieve [20]. This method begins by a grouping strategy in order to establish many different evaluation contexts. For this purpose, this method performs a DEA to identify the best-practice frontier and removes all the efficient DMUs on this frontier (i.e., the first level frontier), then the remaining (i.e., inefficiency DMUs) will form a new second-level frontier. If the DMUs on the second-level frontier are removing, the third-level frontier can be developed, and so on, until no DMU is left [19,21]. By doing that it is able to group DMUs into different clusters via this non-parametric approach (i.e., CD-DEA) and then calculate the progressiveness and attractiveness score for each DMUs by adjusting the evaluation contexts [21]. Based on each DMU's cluster reference (i.e., progressiveness and attractiveness scores), CD-DEA enables assigning attainable goals for inefficient DMUs to become efficient gradually [20]. However, in most real-life situations, it is impossible for a DMU to improve all of the inputs and outputs proportionally at the same time. Based on this concept, the inefficient DMU can look for the most achievable target by performing a CD-DEA and compares its own resources with the benchmark so as to set up level-by-level improvement plans for the current and future. By joint utilization of FRPCA and CD-DEA, we are able to improve the model's discriminant ability as well as steer the users to a more suitable and accessible position.
Different from previous research works on setting up forecasting models that focused heavily on the adoption of numerical messages, this study provides users with an overarching consideration from different perspectives and further utilizes textual messages [22][23][24][25][26]. Rönnqvist and Sarlin [27] also indicated that the signs of a change in corporate operations very likely will appear in textual format that precedes users who try to find any subtle differences in various numerical ratios. People nowadays live in an era of an abundant amount of information that can be accessed easily without any hysteresis. However, too many messages surrounding decision makers can confuse them and even worse result in inappropriate judgments [28,29]. Some messages represented in narratives lack a structured format and sometimes simply turn out to be useless or an encumbrance [30]. How to digest a large amount of textual messages and convert them into manageable ones turns out to be an essential task in today's highly intertwined and connected economies.
To combat the circumstance above, one may use latent Dirichlet allocation (LDA), which is a sort of topic modelling technique. As Cambria and White [31] indicated, LDA is an endogenous Natural Language Processing (NLP) procedure that "involves the utilization of artificial intelligence (AI) methods to conduct a semantic extraction task by generating a knowledge hierarchy that conjecture the concepts from considerable amounts of narratives without a pre-determined knowledge foundation". The topic derived from LDA is illustrated by a distribution over words. The words' distribution for each topic and the topics' distribution for each document are unknown; they are learned from the data [32]. We are capable of discovering the essential topics hidden in large amounts of narratives and delete those topics that have no relation to corporate operations.
Tetlock [33] stated that the sentiment of messages (i.e., financial reports) issued by corporates has a strong impact on stock trading volume. Tetlock et al. [34] further identified that negative sentiment embedded into a corporate's issued documents implies a declining trend in the near future. In order to catch the sentiment hidden in such documents, a dictionary-based approach can be considered. By joint utilization of LDA and a dictionarybased approach, we can decompose financial reports into different topics, preserve those topics that highly relate to a corporate's operations, and match the preserved topics with dictionary-based word lists to exploit the hidden sentiments so as to make conjectures about the corporate's future prospects.
Adopting CD-DEA + FRPCA allows us to determine the performance rank for each corporate and convert the performance forecasting task into a binary classification problem. The sentimental messages hidden in annual reports can be exploited via LDA and a dictionary-based approach. All the analyzed outcomes are then injected into an AI algorithm, called stochastic gradient twin support vector machine (SGTSVM) [35] with superior learning capability, outstanding generalization ability, and a very small memory requirement, to establish a decision making framework for corporate performance forecasting [36]. Current and potential investors can view this model as a roadmap to adjust their investment portfolios so as to reach the maximized profits under an anticipated risk level [37,38]. From the perspective of managers, they can consider many different if-then scenarios to formulate future policies so as to obtain the goal of sustainable development.
Eventually, four main contributions of this study can be summarized as follows. First, this study displays the applicability of FRPCA and CD-DEA in a performance evaluation task so as to help inefficient DMUs to look for an attainable target to follow and plan for current performance improvement and future prosperity. Secondly, we add the stream of financial research that concentrates on risk management. In contrast to other studies (i.e., financial failure prediction, or credit rating forecasting), the works on performance prediction is not well-examined in depth. Third, we introduce a novel strategy to quantify large amount of textual information. By performing LDA (i.e., one of the topic modelling approaches), we can decompose considerable amount of textual messages into different categories. By removing the categories which are not relevant to business operation, and then matches the preserved categories with a sentimental dictionary in order to extract the essential messages hidden in texts. Finally, we insert the analyzed outcome into a kernelbased model (i.e., SGTSVM) to construct the performance prediction model. The model, tested by real cases, is a promising alterative for performance prediction and improvement.
The rest of this paper is organized as follows. Section 2 establishes the literature review. Section 3 presents the introduced decision making framework. Section 4 displays the empirical analysis and outcomes based on the adopted methodologies described herein. Section 5 discusses the conclusion and directions for future works.

Applications of Numerical Information
A corporate financial difficulty takes place when an enterprise suffers continuous and chronic losses or when it cannot pay obligations or liabilities that are disproportionately greater than its assets [39]. Bankers, creditors, stock shareholders, suppliers, and top managers are all interested in corporate financial difficulty assessment due to the considerable negative impacts imposed upon them [40]. The potential assumption underlying financial difficulty forecasting is that the financial ratios derived from financial reports, such as financial statements, income statements, etc., can be used to properly describe the aforementioned characteristics. In other words, financial ratio analysis can assess different facets of an enterprise, such as profitability, liquidity, long-term solvency, and market value. The effectiveness and usefulness of financial ratio analysis appear considerably in the finance, economic, and accounting domains [41].
Altman [5] and Beaver [6] sparked efforts towards the discernment of enterprises' financial difficulty through statistical techniques, namely discriminant analysis (DA), that are grounded on the usage of financial ratios. One of DA's critical drawbacks is its unrealistic assumptions that do not hold in numerous real-life applications [42]. To overcome this problem, some amended statistical approaches, such as linear conditional probability approach, logistic regression (LR), and logit model, were introduced. Martin [43] constructed a financial pre-warning model for banks by LR, gathering research samples from the Federal Reserve System. Ohlson [44] used a logit model to detect an enterprise's financial difficulty, collecting research samples and financial information from the Compustat database and 10-K financial reports. The forecasting performances reported by him were all above 90% for one year, two years, and three years. Kolari et al. [45] introduced a novel mechanism integrating the logit model and trait recognition technique (a non-parametric approach that relies on no statistical or distributional assumptions about the predictor variables) for financial difficulty prediction, achieving prediction accuracy of nearly 95%.

Applications of Textual Information
The narrative analysis of financial documents in recent years has become rather essential in the financial world. While the numerical part of a document only represents a corporate's past performance, the textual part holds information about its future development and performance [46,47]. The literature on accounting narratives can be briefly categorized into two main streams: (1) readability and (2) content analysis [17]. Works on readability have sought to evaluate how difficult textual contents are to comprehend, while those on content analysis seek to draw inferences from the data by "systematically identifying specified characteristics of [a] message" [48,49].
Most early studies placed a lot more emphasis on comprehending the readability of annual reports and their components. For example, Smith and Smith [50] evaluated the readability of financial reports' footnotes from Fortune 50 firms, while Healy [51] examined the readability of financial reports' footnotes from 50 New Zealand firms. Both studies concluded that the readability level of these footnotes is restrictive. Some works have examined the association between readability and other variables. For example, Li [52] found a positive relationship between readability and earnings persistence, while Goel et al. [53] identified that annual reports provided by fraudulent corporates are much harder to comprehend and read.
Another branch of research on accounting narratives covers content analysis. Bowman [54] implemented content analysis to examine the relationship between annual reports and corporate strategies. Bodnaruk et al. [55] stated that the frequency of negative words embedded in 10-K reports exhibits very low correlation with traditional measures of financial constraints and better forecasts impending cases of liquidity trouble.
Apart from the aforementioned studies that collected textual information from narrative documents, some studies focused on obtaining widely available and timely resources of information from news media. Because of the great improvement in information technology and Internet, we are now able to get instant global economic news on all the financial media in a very short time. Interpreting and extracting implicit information from financial news to assist the decision making process have become an attractive research issue. Tetlock [33] studied media (Wall Street Journal) influence on market practitioners and identified the strong influence of negative news on stock trading volume. Tetlok et al. [34] pointed out that negative wording implies a decrease in future corporate earnings. Grounded on previous works, we conclude that textual information derived from narrative documents provides great potential for pattern searching as well as offers a great contribution to the accounting and finance domains.
When compared to widely studied topics like financial difficulty prediction or credit risk prediction, the literature on performance forecasting that involves numerical and textual information is quite scant. In order to fill this gap, this study proposes a performance forecasting model with numerical and textual information. Furthermore, to extract the decisive messages from accounting narratives and to avoid considerable human judgments, a dictionary-based approach (that is, the Loughran-McDonald dictionary: it is a finance-specific dictionary) is adopted. To eliminate the problem of information overload and realize how a corporate's operating performance may change through its financial disclosures across different topics, this study also applies LDA. By joint utilization of a finance-specific dictionary and LDA, we are able to condense a large amount of textual information into holistic and synthesized readable data in a timely manner.

Fuzzy Robust Principal Component Analysis: FRPCA
By integrating Xu and Yuille's algorithm into principal component analysis (PCA), which is related to the energy function, Yang and Wang [56] introduced robust principal component analysis (RPCA) with higher tolerance to outliers and extreme values. The objective function of RPCA can be also extended to be fuzzy, hereafter fuzzy robust principal component analysis (FRPCA), so as to enhance its application fields. A brief illustration of FRPCA runs as follows.
The optimization function of Xu and Yuille's algorithm [57] subject to b i ∈ {0, 1} is expressed in Equation (1): where X = {x, . . . , x n } denotes the research targets, B = {b i |i = 1, . . . , n } expresses the membership function, and η represents the threshold. The purpose is to minimize G(B, w) with respect to b i and w. However, the problem of this method is that b i is a binary variable, and w is a continuous variable, making this optimization task hard to reach an optimal solution via a gradient descent algorithm. To combat this, the minimization task is transformed into a maximization task by performing the Gibbs distribution approach. The mathematical formulation is represented by: where Z denotes the partition function that ensures the summation of ∑ B w P(B, w) equals 1. The assessment criterion can be one of the following functions: The minimization rules of the gradient descent approach for G 1 = ∑ n i=1 g 1 (x i ) and G 2 = ∑ n i=1 g 2 (x i ) are displayed as follows: Here, α t depicts the learning rate, y = w T x i , b = yw, and v = w T b. The objective function done by Yang and Wang's [56] study can be expressed in Equation (7).
Here, b i is the membership of x i , which are assigned to a data cluster, and (1 − b i ) is the membership of x i , which are assigned to a noise cluster. Finally, g(x i ) is taken to measure the error/difference between x i and the cluster center.
The aforementioned idea is similar to the fuzzy c-means algorithm [58]. By performing this transformation, b i becomes a continuous variable and the aforementioned problem (that is, the optimization task that involves the continuous and discrete variables) can be eliminated and the gradient descent algorithm can be executed. By setting δG δb i = 0, we restructure Equation (7), displayed as follows.
By replacing the old membership function with the new membership function, Equation (9) can be obtained.
The gradient with respect to w is displayed in Equation (10): We now set: Here, m 1 denotes a fuzziness variable. If m 1 = 1, then fuzzy membership can be reduced to crisp membership and can be decided by the following rule: This study follows the work done by Yang and Wang [56] to determine the optimal solution. For a more detailed illustration of FRPCA, one can refer to Yang and Wang [56], Luukka [10], and Luukka [13].

Stochastic Gradient Twin Support Vector Machine: SGTSVM
We consider a classical two-class prediction task in n-dimentional real space n . The training instances are expressed as B ∈ n×m , where b ∈ n is the instance with the class label y = {±1}. We formulate the matrix B 1 ∈ n×m 1 and the matrix B 2 ∈ n×m 2 to represent the m 1 instances of class-lebel +1 and the m 2 instances of class-lebel −1, respectively. SGTSVM is based on TSVM that aims at solving the following primal problems.
For SGTSVM, the above quadratic programming problems (QPPs) can be viewed as unconstrained problems. min w 1 ,x 1 1 2 By developing a series of strictly convex function F 1,t (w 1 , x 1 ) and F 2,t (w 2 , x 2 ) with t ≥ 1, the above questions can be solved.
A more detained illustration of SGTSVM can be seen in Wang et al. [35] and Ren et al. [59].

An Advanced Decision Making Framework: FRPCA_CD-DEA + SGTSVM
This study proposes an advanced decision making architecture (see Figure 1) that consists of three procedures for performance ranking determination, essential messages extraction and prediction model construction. (1) Performance ranking determination: To fully capture the status of corporate's operation, CD-DEA with the merits of handing multiple inputs and outputs without predetermining a production function and yielding a reachable learning target to follow so as to setting up an appropriate business strategy for current and future development, is considered. In order to reach a robust result of CD-DEA, the appropriateness of input messages, and data clarification is a required and inevitable pre-process, (2) Essential message extraction: To ensure the inputs and outputs are representative for CD-DEA, the Pearson correlation is considered. CD-DEA involves with too many input and output variables will deteriorate its discriminant ability. To combat this, one of the dimensionality reduction techniques, namely FRPCA is taken to eliminate the effect of curse of dimensionality as well as increase its discriminant ability. Most proportion of the research works on pre-warning model constructions rely heavily on numerical information due to its characteristics of easy-to-use and intuitiveness. However, the numerical information represents merely historical messages, it contains less messages about corporate's future plans and prospects. To complement the time lag problem caused by considering numerical information only, the textual information which can be vised as a leading indicator is taken. How to handle large amount of textual information is an essential issue in today's information overload economy. LDA, one of the topic modelling algorithms, is performed to divide whole textual information into some topics and preserves the topics which are relevant to business operations. How to conjecture the managers' attitudes toward corporate's future development is an essential issue. To deal with this issue, the sentimental dictionary is considered. By joint utilization of LDA and sentimental dictionary, we can infer the managers' attitude toward company's future development. The market participants can view this as a guideline to modify their investment portfolios and allocate resources to suitable places to maximize their personnel wealth under anticipated risk levels, and (3) Prediction model construction: For an unfamiliar domain, the decision makers tend to collect all of the related information as much as possible to conjecture the inherent situation. However, too many information surround decision makers will easily confuse their judgments and then result in inappropriate decisions. Thus, essential message extraction turns out to be an important process before prediction model construction. Rough set theory (RST) has been demonstrated its usefulness and robustness in essential message extraction, knowledge generation and classification. Thus, the RST is executed to determine the essential messages [60,61]. The selected messages are then injected into SGTSVM to construct the prediction model. The data used to construct the prediction model is divided into two subsets: (1) training data subset and (2) testing data subset. The former is used to construct the prediction model, and the latter is applied to examine the effectiveness of the constructed model. Five-fold cross-validation is applied to deal with the over-fitting problem. The decision makers can take the introduced model as a roadmap/reference to adjust the corporate's capital structure, reduce the financing burden, and increasing profitability so as to reach the goal of sustainable development.

The Data and Decision Variables
The electronics industry in Taiwan has been widely admired as an economic miracle of the world, has considerable influences on global supply chains of personal computers (PCs), integrated circuits (ICs), and electronic devices, and is viewed as an essential capital market to worldwide investors [17]. In addition, this industry has received large amounts of financial incentives and policy support initiatives from governments and gathered funding from market participants at very low capital costs, turning it gradually into an economic backbone of Taiwan's stock market. The trading volumes of this specific industry typically exceed 60% of total trading volumes. As the electronics industry has a great impact on the society and economic development of Taiwan, this study takes it as a research target and aims to dig out more hidden useful insights for future policy formulations.
Before we convert the performance forecasting task into a classic binary classification problem, we need to determine the performance rank for each corporate in advance. DEA has demonstrated its usefulness and gained widespread success for performance evaluation among dissimilar research fields. The weakness of conventional DEA is that when the analyzed data have exogenous factors, the performance rank for inefficient corporates often ends up with a target that the model recommended, which is very complicated to reach or attain [14,16]. Thus, it is an urgent requirement to determine an achievable target for inefficient corporates. To reach this goal, CD-DEA is performed to determine a corporate's performance rank. In estimating the efficient frontier, there are two assumptions: constant return to scale (CRS): it assumes that any level of increase in inputs will proportionately increase the level of output, and variable return to scale (VRS): it assumes that any increase in the level of inputs will either increase or decrease the level of output [62]. There are two orientations in DEA model: input-oriented and output-oriented. The former seeks to minimize the inputs with the assumption of fixed level of outputs, and the latter aims to maximize the output with an assumption of fixed level of inputs, and thus maximize the efficiency. In this study, we apply CRS, output-oriented CD-DEA to reach the outcome. Based on the performance level the corporates are at, we divide all corporates into two categories: one for efficiency and the other one for inefficiency. In accordance to previous work done by Tone and Tsutsui [63] and Xu and Wang [64], we choose five input variables: total assets (TA), number of employees (NOE), fixed assets (FA), cost of goods sold (COGS), and total debt (TD); and four output variables: total sales (TS), gross margin (GM), net income (NI), and earnings per share (EPS). Table 1 displays the selected input and output variables. We can see that our selected input and output variables mixed use of raw data and ratio. Based on the research work done by Cook et al. [65] they indicated that a mixture of ratios/percentiles and raw data is permissible in DEA applications.

The Results
Because examining the selected variables for CD-DEA is reliable, we employ the Pearson correlation (see Table 2). We see that all the selected variables are positive and reach a significance level. In other words, all the selected variables are representative and appropriate.
For an unknown domain, users tend to collect as much information as possible to conjecture the inherent reality of this domain. However, too many messages to the endusers will deteriorate their judgmental ability and even worse may result in decision failure. To confront this, we utilize FRPCA, which is a dimensionality reduction technique. Table 3 displays the test results concerning the effectiveness of FRPCA. The results indicate that CD-DEA with FRPCA has a lower mean value and a higher variance value-that is, the model with dimensionality reduction has better discriminant ability than the model without dimensionality reduction. This finding is echoed with the work done by Lin [3].  In addition, the number of DMUs in comparison to the number of combined inputs and outputs has been a concern when it comes to applying DEA. That is, large numbers of combined inputs and outputs compared to the number of DMUs may deteriorate the DEA's discriminant ability [65]. A suggested "rule of thumb" is that the number of DMUs be at least twice the numbers of combined inputs and outputs [66]. To further examine the discriminant power of the introduced performance evaluation mechanism (i.e., FRPCA + CD-DEA), data eliminating strategy is executed. In this experimental test, we randomly eliminate the constant percentage (i.e., 20%) of the whole data and use the remaining data to get the results from two different evaluation models. The results are shown in Table 4. We can see that the introduced model still performs better than CD-DEA under all scenarios. Peters et al. [4] pointed out that the main trigger for financial difficulties/crises is bad operating performance. That is, the bad operating performance can be viewed as a prior stage before the financial difficulties/crises burst forth. Due to the nature of corporate financial difficulties/crises highly related to corporate's bad operating performance, the variables we chosen in financial difficulty forecasting can be taken as the surrogate for joining up with the predictors. According to the related works [67][68][69][70], the selected variables are represented in Table 5. Apart from prior works on building performance forecasting models that merely focus on numerical messages, this study aims at providing decision makers with an overarching decision making framework from different perspectives of corporate operations and equipping them with textual messages that can transmit future corporate performance without any time delay. How to digest large amounts of textual messages is an urgent requirement in today's information society. Thus, one may use a topic modelling technique, called Latent Dirichlet Allocation (LDA), to extract the essential topics embedded in a corporate's annual report. By implementing LDA, we can identify the essential topics and preserve those topics that are highly relevant to corporate operations. Perplexity is a commonly adopted criterion in information theory to measure how well a statistical approach describes a dataset, with lower perplexity denoting a superior probabilistic method [71]. The results show that five topics are preserved (see Figure 2) and the word lists for each topic is displayed in Table 6. To extract the inherent sentimental messages hidden in these topics, we utilize a finance-specific dictionary (i.e., Loughran-McDonald dictionary: LM dictionary). We count the number of positive words and the number of negative words in each topic. If the number of positive words exceeds the number of negative words in this topic, then it means this topic contain positive information and vice versa. Feature selection is an inevitable pre-process in data mining and machine learning. Rough set theory (RST) has proven its superior performance in the identification of essential features. The weakness of RST is to determine the minimal reduct that is viewed as a NP hard task-that is, the run time of generating all reducts is exponential [72,73]. To confront this, one type of heuristic algorithm, the fish swarm method, is employed.  Table 7 displays the selected features. We can see that the selected variables contain two topic-based messages. To further test the usefulness of the topic-based messages, we divide the experimental design into two conditions: one is the model with topic-based messages, and the other one is the model without topic-based messages. Apart from previous forecasting model establishment studies with findings that merely consider one assessment measure, such as total accuracy and total error rate, this study takes two assessment measures, called type I error and type II error, into consideration so as to reach a more reliable and trustworthy judgment. Five-fold cross-validation (CV) is adopted to eliminate the problem of over-fitting. We can see that the model with topic-based messages performs better than the model without topic-based messages. This finding (see Table 8) supports Huang et al. [74] who noted that sentimental messages from annual reports (i.e., financial narratives) provide much denser and detailed information beyond just financial messages [75][76][77][78].  To avoid the outcome just happening by coincidence, we take the introduced model as a benchmark and further compare it with the other three AI-based algorithms: relevant vector machine (RVM), extreme learning machine (ELM), and back propagation neural network (BPNN). Five-fold cross-validation is adopted to eliminate the problem of overfitting, and the Friedman test is conducted to examine the robustness of the introduced model. Table 9 displays the comparison outcomes. We see that the introduced model still outperforms the other three algorithms. The introduced model can therefore be used as an alternative for performance forecasting. Most related studies conclude their finding rely on a pre-determined dataset. To test the generalization ability of the proposed model, we consider the other four situations. The results are displayed in Table 10. We can see that the proposed model still reaches an outstanding performance under all situations. In summary, the introduced model is a promising alternative for performance prediction.

Conclusions and Future Works
Compared to the well-developed domain of financial crisis prediction and credit rating forecasting, there are scant works on corporate operating performance forecasting. To fill this research gap, our study aims at developing an effective decision making framework for this type of forecasting. Most studies on performance evaluation have focused considerably on ratio analysis, but this measuring framework generally has one input and one output, which cannot depict the real status of a corporate's operations, especially in today's highly intense and interconnected economy. To combat this, this study sets up CD-DEA, which can handle multiple input and output variables simultaneously without a pre-determined cost function and is able to yield an attainable target for analyzed corporates to initiate plans for short-and long-term development. In order to enhance the CD-DEA's discriminant ability, FRPCA is employed herein.
This research does not restrict itself to numerical messages, but further extends into including sentimental messages hidden in financial narratives so as to construct an overarching and comprehensive decision making framework. The findings echo those by Huang et al. [61] who stated that textual information can be used to complement the information lag concluded from numerical messages.
Future research works can consider several directions as follows.
(1) Adoption of advanced feature selection approaches: The original features of a model can be inserted into kernel-based methods, such as kernel principal component analysis (KPCA) and kernel independent component analysis (KICA), to extract much deeper insight through data semantics. (2) Advanced decision making framework: The CD-DEA can be integrated it with cross efficiency to prevent the obstacle caused by radial-based measure. That is, the problem of radial-based measure expresses how distinct the DMU under evaluation is from a single specific DMU in the evaluation context, not from the entire evaluation context overall. By considering cross efficiency, we can achieve a more precise conclusion and eliminate unrealistic weighting scheme so as to fit into practical applications.