Next Article in Journal
Fault Diagnosis Method of Planetary Gearboxes Based on Multi-Scale Wavelet Packet Energy Entropy and Extreme Learning Machine
Previous Article in Journal
Entropy and Stability: Reduced Hamiltonian Formalism of Non-Barotropic Flows and Instability Constraints
Previous Article in Special Issue
Entropy in the Assessment of the Labour Market Situation in the Context of the Survival Analysis Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Information-Theoretic Cost–Benefit Analysis of Hybrid Decision Workflows in Finance

1
Department of Engineering Science, University of Oxford, Oxford OX1 3QG, UK
2
Saïd Business School, University of Oxford, Oxford OX1 1HP, UK
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(8), 780; https://doi.org/10.3390/e27080780
Submission received: 3 June 2025 / Revised: 10 July 2025 / Accepted: 17 July 2025 / Published: 23 July 2025

Abstract

Analyzing and leveraging data effectively has been an advantageous strategy in the management workflows of many contemporary organizations. In business and finance, data-informed decision workflows are nowadays essential for enabling development and growth. However, there is yet a theoretical or quantitative approach for analyzing the cost–benefit of the processes in such workflows, e.g., in determining the trade-offs between machine- and human-centric processes and quantifying biases. The aim of this work is to translate an information-theoretic concept and measure for cost–benefit analysis to a methodology that is relevant to the analysis of hybrid decision workflows in business and finance. We propose to combine an information-theoretic approach (i.e., information-theoretic cost–benefit analysis) and an engineering approach (e.g., workflow decomposition), which enables us to utilize information-theoretic measures to estimate the cost–benefit of individual processes quantitatively. We provide three case studies to demonstrate the feasibility of the proposed methodology, including (i) the use of a statistical and computational algorithm, (ii) incomplete information and humans’ soft knowledge, and (iii) cognitive biases in a committee meeting. While this is an early application of information-theoretic cost–benefit analysis to business and financial workflows, it is a significant step towards the development of a systematic, quantitative, and computer-assisted approach for optimizing data-informed decision workflows.

1. Introduction

Data plays a significant role in modern decision workflows. In business, for example, the competitive advantage of harnessing and analyzing data efficiently can be a critical factor in determining market leaders and laggards. Companies are increasingly employing data-informed strategies to enhance customer experiences, improve operational capabilities, and automate certain decision-making processes. Because such workflows commonly consist of various human- and machine-centric processes (e.g., data collection, statistical inference, machine learning, data mining, data visualization, report reading, discussion meetings, etc.), we intentionally use the term “data-informed decision workflows”, as the widely used adjective data-driven commonly implies automated machine-centric processes. While it is vital for organizations to adopt data-informed decision workflows [1], there are many challenges in designing, managing, and optimizing such data-informed decision workflows. For instance, (a) data collection may not be as comprehensive as desired due to various restrictions, (b) the quality of statistical inference may be affected by the sparsity and skewness of data, (c) the reliability of machine-learned models may be undermined by dated training data, (d) interpretation of visualization may be subjective, (e) human participants may introduce biases in discussions, and so forth. At the same time, all these component processes in data-informed decision workflows incur costs. Hence, there is a serious need for a scientific methodology that can be used to analyze, measure, and predict the performance of the component processes in a data-informed decision workflow. In this work, we propose a new methodology for using information to analyze and explain the cost–benefit of different processes in hybrid decision workflows in business and finance, where decisions rely on information from data and humans as well as technologies used. Our aim is to translate an information-theoretic concept and measure for cost–benefit analysis to a methodology that is relevant to business and finance. As the original concept is highly abstract, it is necessary to equip the methodology with an engineering approach for modeling and decomposing complex decision workflows into processes to which the concept can be applied in a quantitative manner.
We have tested this methodology in conjunction with a number of workflows in business, finance, and industry. In this paper, we report three case studies in a business and financial workflow, which represent three types of commonly encountered phenomena: (i) uses of statistical and computational algorithm, (ii) incomplete information and uses of humans’ soft knowledge, and (iii) cognitive biases in committee meetings. These case studies demonstrate the feasibility of using the proposed methodology in analyzing and explaining the cost–benefit of different processes in data-informed decision workflows.
The remainder of the paper includes a brief overview of the relevant research in Section 2, a detailed description of the proposed methodology in Section 3, three case studies in Section 4, discussions on several related topics in Section 5, and our concluding remarks and suggestions for future work in Section 6.

2. Related Works

Data-informed decision-making in companies. In the foundational work “The Practice of Management”, Drucker [2] acknowledges the significance of data in influencing managerial decision-making to measure and enhance organizational performance. In a more recent study, Brynjolfsson et al. [3] put forth statistical evidence that data-informed decision-making significantly enhances productivity in modern firms. This finding is also confirmed by Provost and Fawcett [4], who strongly advocate for the positive effect of data-focused practices on organizational outcomes. Numerous scholars have concentrated on improving the utilization of data in enterprises over the last decades. Sun et al. [5] offer a data-flow framework, which focuses on data-flow specification and analysis, facilitating enhanced analytical rigor in business process management, particularly when encountering complex challenges. Further, Zhang and Perry [6] optimize time performance in business workflows using a data-centric approach to process modeling. Moreover, Kougka et al. [7] describe that data-centric workflow optimization undergoes a swift transformation and emphasize new capabilities of advanced data management for modern data-driven businesses, including big data analytics. Building on these studies, this work contributes to advancing data-informed decision-making in companies, illustrating a well-defined methodology to optimize business decision workflows.
Human-centric elements in business data workflows. Components in data intelligence processes can be machine- or human-centric [8]. Humans can enhance the scope of inference with “soft” information, which they can utilize in the form of pre-existing beliefs or knowledge. Tam et al. [9] conclude that human-centric designs of decision-tree models can outperform machine-centric approaches due to soft alphabets available to humans that are not present in data. Kijmongkolchai et al. [10] substantiate this finding with empirical evidence that humans use soft knowledge to reduce potential distortion. Our research incorporates the phenomena discussed in these previous studies into one information-theoretic framework. We utilize information theory to demonstrate that humans can mitigate potential distortion in a data-driven workflow through soft knowledge. At the same time, we demonstrate that human biases in decision-making can be effectively remedied through technology.
Economic CBA methods. Monetary cost–benefit analysis can be traced back a century ago. The net present value, as one of the most widely applied investment appraisal methods, was presented by Irving Fisher [11]. The Internal Rate of Return computes the discount rate that leads to a net present value of 0 of any given project and is widely used in private equity [12,13]. The Value of Information approximates the expected economic value of acquiring more information before a decision point [14]. Furthermore, the benefit–cost ratio, enables project evaluation by following the mathematical foundation of NPV to compare a project’s benefits and costs, albeit in a normalized form [15]. Whereas CAPEX and OPEX modeling, prevalent in project finance analysis, differentiate upfront investment from ongoing operational costs [16]. We will discuss the difference between economic CBA methods and the information-theoretic cost–benefit analysis in Section 5.
Workflow optimization and decision process modeling. The seminal work from Bellman (1957) introduces dynamic programming and Markovian decision processes (MDPs) to achieve optimal outcomes in multi-stage decision processes [17]. Kelley and Walker (1959) present the critical path method as a systematic technique to optimize a project schedule within a workflow [18]. Subsequently, Hammer and Champy (1993) put forth business process reengineering as a radical method for redesigning workflows [19]. Howard and Matheson (1984) utilize influence diagrams to represent the structure of complex decision problems under uncertainty [20]. In more contemporary work, Belousov and Peters (2019) extend MDP optimization by using information-theoretic regulation in the context of reinforcement learning [21]. Habermehl et al. (2022) connect model-based systems engineering with multidisciplinary design analysis and optimization to improve decision workflows in engineering [22]. This work builds upon the findings of these previous studies to develop an approach for analyzing data-informed decision workflows.
Modern information-theoretical advances in explainable AI (XAI), bias control, and industrial digitization. Multiple studies put forth evidence for the effective utilization of information-theoretical optimization in the context of explainable AI, bias control, and industrial digitization. Barbiero et al. (2022) extract explanations from neural networks, while optimizing an entropy-based criterion that determines the most pertinent concepts [23]. Sanneman (2024) describes an information-theoretic method, the Information Bottleneck approach, that optimizes for maximum informativeness and minimum complexity, thereby facilitating the design of user-specific XAI [24]. Mehdi and Tiwary (2024) present interpretation entropy as a general approach to assessing the human ability to interpret linear models [25]. Furthermore, Bang et al. (2021) present VIBI, the variational information bottleneck for interpretation, which utilizes the information bottleneck principle to identify concise yet extensive elaborations [26]. The contemporary review by Dutta and Haman (2023) emphasizes partial information decomposition as a strong information-theoretical tool to investigate explainability [27]. Kim and Cho (2022) present an information-theoretical method to alleviate bias in machine learning to achieve fairness [28]. Li et al. (2023) derive a non-vacuous upper bound based on information theory on the bias removal strength of any approach [29]. Similarly, substantial advances in entropy optimization are achieved in the field of industrial digitization. Omar and Plapper (2022) utilize path-transfer flow entropy to assess manufacturing flexibility quantitatively [30]. Herrera-Vidal et al. (2025) present an entropic metric of complexity for production systems developed based on simulation and programming methods [31]. Based on these previous studies, we present an operational method that utilizes decomposition and applies information-theoretic cost–benefit analysis for the optimization of data-informed decision-making workflows.
Cost–benefit analysis of data intelligence workflows. Chen and Jänicke [32] propose an information-theoretic framework of visualization, addressing the need to measure information content, loss and flow within pipelines and workflows. Chen and Golan [33] describe that processes within a data intelligence workflow can be subdivided into multiple steps, referred to as transformations. Further, building on literature on general information processing [34] and information-theoretical quantification in mathematics, economics, and statistics [35,36,37], Chen and Golan [33] present a cost–benefit measure, which can be utilized as a cost function to optimize data intelligence workflows. In other words, they provide the first measure that represents the trade-off between alphabet compression, potential distortion, and costs within a pipeline or data intelligence workflow. Subsequently, Chen and Ebert [38] outline the first systematic process for optimizing visual analytics workflows. Thereby, they analyze the weaknesses (symptoms) of the processes in a workflow and identified technical solutions (remedies) to improve the workflow. As the cost–benefit measure proposed by Chen and Golan [33] makes use of the Kullback–Leibler (KL) divergence [39], its potential distortion term is unbounded even when the total Shannon entropy [40] of the data space of a process is finite. Chen and Sbert [41] propose to use a bounded divergence measure for the potential distortion term. They studied several candidate measures conceptually, followed by several experiments in practical scenarios [42]. Furthermore, Chen [43] suggests applying the information-theoretic cost–benefit analysis to many data intelligence workflows in other fields, including machine learning, psychology, language development, and media communication.

3. Methodology

In this section, we first describe information-theoretic cost–benefit analysis, followed by a description of the workflow modeling and decomposition method. As mentioned in Section 1, using the latter (an engineering approach) to build a bridge from the former (a highly abstract concept) to hybrid decision workflows in business and finance is the aim and novel contribution of this work.

3.1. Information-Theoretic Cost–Benefit Analysis

A data-informed workflow consists of both human- and machine-centric processes that transform data into one or more decisions. In information theory, each process is represented by a function F referred to as a transformation, all possible inputs to F form an input space, and all possible outputs from F form an output space. The input and output spaces are represented by two alphabets, Z in and Z out , respectively. All possible different inputs are the letters of Z in , which is associated with a probability mass function (PMF). Similarly, all possible different outputs are the letters of Z out . In general, a decision alphabet at the end of a data-informed workflow has much less entropy than the total entropy of all data alphabets that have contributed to the decision. Hence, there is usually a large entropy loss from data to decision.
Chen and Golan [33] first noticed the need to explain this ubiquitous phenomenon mathematically. They recognized that entropy loss, which is colloquially referred to as information loss, should not be regarded as a negative factor, since it is ubiquitous in data-informed workflows. They used two information-theoretic measures to represent the positive and negative aspects of entropy loss as Alphabet Compression and Potential Distortion. Together with a measure of cost, this gives the following conceptual formula:
Benefit Cost = Alphabet Compression ( AC ) Potential Distortion ( PD ) Cos t ( Ct )
Given a transformation F i : Z i 1 Z i (among a series of transformations), the term AC measures the entropy reduction achieved by a transformation F i , i.e., H Z i 1 H Z i , where H is the measure of Shannon entropy [40]:
H ( Z ) = H P ( Z ) = H ( P ) = i = 1 n p ( z i ) log 2 p ( z i )
where the PMF of Z is P ( Z i ) = { p ( z 1 ) , p ( z 2 ) , , p ( z n ) } . Since an input alphabet is more likely to have more entropy than an output alphabet, AC measures entropy loss positively. Note that with the log 2 measure, H ( Z ) also has the unit “bit”.
The term PD measures the negative impact caused by the entropy loss due to the transformation F i , such as errors of a recommendation algorithm or a human’s decision. Chen and Golan [33] proposed to use the KL divergence [39] for the PD term. Later, Chen and Sbert [42] proposed to use a bounded divergence measure instead, to make the AC and PD terms more comparable in scale. In this work, we adopt the bounded measure. Let P be the PMF of a reference alphabet Z ref and Q be the PMF of a modeled or believed alphabet Z bel , the divergence measure proposed by Chen and Sbert is as follows:
D = D cs ( P Q ) = 1 2 i = 1 n p i + q i log 2 p i q i 2 + 1
with the log 2 measure, D cs ( P Q ) has the unit “bit”.
Chen and Sbert also considered that Jensen–Shannon divergence D js could be a suitable bounded measure and discussed the relative merits with the aid of numerical experiments [42]. For consistency, we use a generic symbol D for divergence measure in the next section, while using D cs in all calculations. Readers can replace D cs with D js to observe slightly different behaviors of the D js measure. As D cs (as well as D js ) is bounded by [0, 1], it is necessary to define PD in conjunction with a maximum entropy bound of the alphabet concerned. Since Z ref and Z bel share the same set of letters, we can use the maximum entropy of either.
Chen and Golan [33] considered that the ideal measure for the term cost would be energy, but recognized that, in practice, cost would typically be approximated by time, monetary cost, or another more attainable measure.
With the above definitions, the conceptual cost–benefit ratio in Equation (1) can be expressed quantitatively as follows:
Benefit Cost = H ( P ) H ( Q ) H max ( Z ref ) D ( P Q ) Cost
where P is the PMF of a reference alphabet Z ref and Q is the PMF of a modeled or believed alphabet Z bel .
It is useful to emphasize that the cost–benefit measure is reference-dependent. When we measure the cost–benefit ratio, alphabet Z ref can be a confirmed ground truth, which leads to the actual distortion incurred with Z bel . However, in practice, we may use an estimated or presumed ground truth, and in such a case, when D ( P Q ) > 0 , it does not necessarily imply Z bel is actually incorrect. Z bel is seen to cause a distortion only from the reference point of Z ref . Hence, the terms “reference alphabet” and “potential distortion” are here to imply that Z ref is not necessarily a ground truth and D ( P Q ) measures only the potential distortion. In Section 4, we will use Z ref in different contexts. Because of the reference point, the cost–benefit measure is also reference-dependent.
As this work focuses on the translation of the abstract information-theoretic concept and the cost–benefit measure in Equation (4), we assume that the letters of each alphabet have already been defined and all PMFs have already been obtained. There is no constraint as to the distribution that a PMF may take. In Section 5, we will discuss the relevance of the existing and future research on PMF estimation, uncertainty analysis, and sensitivity analysis to further advancement in transforming this work to real-world applications.

3.2. Workflow Decomposition

Although the information-theoretic cost–benefit analysis has been used to explain why data visualization is a cost-beneficial tool in general [33] and to measure simple visualization processes [41,42], it has not yet been used to analyze any slightly complex workflow quantitatively. There were suggestions that such applications may be possible [43], but the attempts reported in the literature are largely qualitative methods based on Equation (1) (e.g., [38]). The main reason behind the lack of reported quantitative attempts is that for any slightly complex data-informed decision process (i.e., a transformation in terms of information theory), the input data space is usually large and very complex, which translates to a situation where the corresponding input alphabet has numerous letters and the PMF for such an alphabet is usually very difficult to estimate.
To address this challenge, we adopt the common engineering approach for decomposing a complex data-informed decision workflow into a series of processes. If any processes are still too complex to be analyzed using the information-theoretic cost–benefit analysis, we can further decompose these processes iteratively, until the input and output alphabets for each process are reasonably easy to define and the corresponding PMFs are relatively easy to estimate. Of course, one could end up with dozens or even hundreds of processes, which might sound unmanageable. However, in many engineering disciplines (e.g., the design and management of power or chemical plants), engineers have commonly dealt with workflows involving hundreds or thousands of processes, even before computers were used to aid such tasks. Hence, it will be feasible to analyze many processes in a complex data-informed decision workflow if we can develop a generalizable method to analyze some individual processes. This is precisely the focus of this work. We believe that as this generalizable method matures, computer-aided tools will emerge to help human analysts manage hundreds or thousands of processes, as is common in many engineering disciplines today.
For example, consider a complex data-informed decision workflow at the top of Figure 1. This is a representative workflow for making discretionary long/short equity decisions in many financial organizations. At the top level, there are nine processes, each of which is still too complex to be analyzed using the information-theoretic cost–benefit analysis. We can decompose any of these processes further. For instance, if we are interested in process “4. Data Analysis”, we can consider three design options in the middle of Figure 1, labeled with black-circled numbers 1, 2, and 3, respectively. (1) A human-centric sub-system represents an old practice without using many data science tools. (2) A machine-centric sub-system represents a series of processes for making decisions automatically, which many consider to be optimal in the future. (3) A sub-system, where human- and machine-centric processes are integrated together and make decisions, represents the most common practice today.
Let us focus on the sub-system (3), where we can identify process blocks that are small enough for information-theoretic cost–benefit analysis (as opposed to others that will need further decomposition). The decomposition of one such block is shown at the bottom part of Figure 1. In the next section, we show how the processes in three case studies can be analyzed using the method described in Section 3.1. These three case studies are marked in Figure 1 with red boxes.

4. Case Studies

In the following section, we illustrate the feasibility of using cost–benefit analysis for machine-centric and human-centric decision processes through three case studies, each demonstrating selective facets of the method. The abstraction describes a data-informed decision workflow with machine- and human-centric components, examining an exemplary long/short equity strategy. We compare various process types, such as parallel and sequential processes, as well as the level of requested information. Furthermore, we focus on the human-centric components of the investment committee and suggest improvements to ensure the effective execution of the meeting process. Similarly, we estimate the amount of alphabet compression, potential distortion, and cost of selected components. We consider some typical phenomena in such a workflow, including consistent data analysis or erroneous recommendations by algorithms, appropriate or inappropriate use of visualization, expert knowledge or personal biases of committee members, missing data, and so forth. Finally, we analyze the internal policy and cost considerations of involving an investment committee or executing a trade without further external scrutiny. In the first case study, we demonstrate the analysis of machine-centric processes, illustrated by examining an analytical moving average (MA) crossover algorithm. In the second and third case studies, we examine the influence of incomplete information, soft knowledge, human bias and cost on the workflow structure and decision outcomes.

4.1. Case Study 1: An Analytical Algorithm Scenario

In business and financial decision workflows, there are many computational processes. Intuitively, we all know that these computational processes can significantly reduce the cost. Information-theoretically, they also typically reduce a significant amount of entropy from input to output, thereby facilitating alphabet compression. For example, in a financial decision-making workflow of a long/short equity hedge fund, many analytical algorithms facilitate trading decisions by utilizing quantitative signals derived from a multitude of data sources, such as industry index data, stock price data, and proprietary alternative data. One algorithm is frequently used for monitoring moving average crossovers—a well-established technical momentum indicator. In Figure 1, the block marked as “Case Study 1” commonly includes the use of the moving average crossovers method.
Consider an analyst who utilize a crossover algorithm to compare 50-day and 200-day moving averages for M equities, over a period T. When the 50-day average rises above (or falls below) the 200-day average for any single stock, the algorithm triggers a bullish (or bearish) signal and recommends a long (or short) position due to positive (or negative) momentum. While a systematic trading algorithm could execute a long (or short) position based on this signal alone, its utilization varies among hedge funds. Nevertheless, one can reasonably assume that these crossover indicators allow the analyst to identify potentially interesting trading opportunities for further analysis and enable more efficient monitoring of stocks. Let the period T be sampled as t 1 , t 2 , , t n . Let s k be one of the M stocks covered by the analyst. The moving average crossover signal is a ternary time series for the stock σ 1 ( s k ) , σ 2 ( s k ) , , σ n ( s k ) , such that
σ t ( s k ) = + 1 , if MA 50 ( t 1 , s k ) MA 200 ( t 1 , s k ) and MA 50 ( t , s k ) > MA 200 ( t , s k ) 1 , if MA 50 ( t 1 , s k ) MA 200 ( t 1 , s k ) and MA 50 ( t , s k ) < MA 200 ( t , s k ) 0 , otherwise
where MA 50 ( t , s k ) and MA 200 ( t , s k ) are, respectively, the 50-day and 200-day moving averages of stock s k at time t [44].
In the context of an individual temporal point t, the original stock value is represented by a 32-bit integer [0, 4,294,967,295]. (To date, the highest share value recorded is less than 100,000,000 cents.) Let Z sv be the alphabet for representing share values in a 32-bit integer, and the maximum entropy of Z sv is thus 32 bits. Even if we consider only the value range [0, 100,000,000], the maximum entropy would still be about 26.575 bits. As very high share values are rare, the typical entropy of Z sv is much lower than its maximum entropy. Meanwhile, to encode σ t ( s k ) , one needs an alphabet, Z σ , with only three letters ( + 1 , 1 , 0 ), and its maximum entropy is about 1.585 bits. Likely, the probability mass function (PMF) is in favor of 0 at most time steps, and the entropy of Z σ is much lower than 1.585 bits.
To simplify the discussions below, we assume that the time steps in the period T are the intervals between two consecutive trading days. To derive a single σ t ( s k ) indicator, the algorithm needs 200 days of stock values of s k . To derive n σ t ( s k ) indicators for t 1 , t 2 , , t n , the algorithm needs n + 199 days of stock values of s k . If we estimate the alphabet compression of this transformation using maximum entropy, we have the following:
MA   Crossover   Transformation : ( n + 199 ) H max ( Z sv ) n H max ( Z σ )
For n = 100 days, this amounts to 299 × 32 100 × 1.585 = 9409.5 bits per stock. We can consider this as an upper bound. To estimate a lower bound, we may assume that the entropy of an individual share is about 6 bits (up to 64 different values uniformly distributed). With such an assumption, for n = 100 days, it amounts to 299 × 6 100 × 1.585 = 1635.5 bits per stock. We, thus, have the lower and upper bounds for alphabet compression per stock as [1635.5, 9409.5] bits. When we consider the analyst has to cover M equities, the amount of alphabet compression is even more impressive.
With such a considerable amount of alphabet compression, the transformation will no doubt incur a large amount of potential distortion if one is required to use a single indicator σ t ( s k ) to reconstruct a 200-point time series of the share values of s k , or the time series representing the long- and short-term moving averages. The task is unattainable due to the unusable and erroneous outcome, as well as the unaffordable cognitive effort. However, this is hardly ever necessary. The time series plot for the original data and moving averages is usually readily available, and the analyst can simply glance at the plot to regain information about the original time series and the moving average lines. In this way, the reconstruction accuracy is reasonably high, and the cost is very low. Let us consider the following:
  • AC—The amount of Alphabet Compression achieved by the MA Crossover Algorithm;
  • PD-noVis—The amount of potential distortion if one does not visualize any of the time series (i.e., the original and moving averages);
  • PD-Vis—The amount of potential distortion if one does visualize the time series (i.e., the original and moving averages);
  • Cost-noVis—The amount of cognitive effort for reconstructing any of the time series from the MA Crossover indicators (or the cost of decision errors due to not being able to view any of the time series);
  • Cost-Vis—The amount of cognitive effort for viewing the time series (or the cost of decision errors after viewing the time series);
We can easily conclude:
benefit cost : AC PD-noVis Cost-noVis AC PD-Vis Cost-Vis
The information-theoretic analysis of this scenario applies to many uses of statistical and computational methods in business and financial workflows and can explain why financial analysts and managers frequently glance at time series plots and other data visualization plots.

4.2. Case Study 2: Two Analyst–Portfolio Manager Scenarios

In business and financial decision workflows, it is common to have a process where an analyst proposes a decision and a manager confirms, rejects, or modifies the decision. Figure 1 shows such an example process (marked as “Case Study 2”), where an analyst obtains conventional financial data, alternative data, internal data, third-party data, inter alia, to assess equity investment opportunities. A portfolio manager receives a recommendation from the analyst and makes the final decision. In such processes, critical information may often not be captured by the available data. When an analyst presents a trade idea and the corresponding investment thesis, the recommendation may or may not be appropriate, depending on how the analyst addresses the missing information. Meanwhile, a diligent portfolio manager may pay attention to the risk due to missing information, discussing how the analyst’s knowledge was used to make assumptions or derive inferences from other information. In contrast, an inattentive portfolio manager may simply accept the analyst’s recommendation at face value. Let us consider two scenarios:
  • CS2x—Inappropriate Recommendation, Diligent Manager
  • CS2y—Inappropriate Recommendation, Inattentive Manager
Consider the detailed diagram illustrating this process in Figure 2, where alphabet Z data consists of all possible variants of the data that the analyst has processed and alphabet Z miss consists of all possible variants of the missing information. Z miss a is the alphabet of missing information after the analyst has reasoned with assumptions and inference, while Z miss b is the alphabet of missing information after the manager has considered the analyst’s reasoning together with the manager’s own assumptions and inference. Along the recommendation flow, alphabet Z rcmd consists of all possible recommendations that the analyst could make before processing the data, Z rcmd a be the alphabet after a recommendation has been formulated, and Z rcmd b be the alphabet representing the final decision on the recommendation by the portfolio manager.
Information-theoretically, we can first recognize the amount of alphabet compression that has been achieved by the analyst who transforms the three alphabets on the left to the alphabet of the analyst’s recommendation, since Z data , Z miss , and Z rcmd should have much more entropy than Z rcmd a . Typically, the recommendation alphabet Z rcmd consists of two sub-alphabets. One has 75 letters for encoding different summary views on stock direction, catalyst existence, and conviction, i.e.,
Z rcmd 1 = strong rise rise neutral decline strong decline × only one short-term catalyst only one medium-term catalyst only one-long term catalyst multiple catalysts no catalysts × high conviction medium conviction low conviction
The other consists of four letters representing possible recommended decisions on a summary view:
Z rcmd 2 = { accept , reject , request   more   information , wait }
Its maximum entropy is H max ( Z rcmd ) 8.2 bits. As the PMF of Z rcmd (before processing) is expected to be less certain than the PMF of Z rcmd a (after processing), we have H ( Z rcmd ) > H ( Z rcmd a ) . Meanwhile, alphabets Z data and Z miss are defined upon many complex variables, including time series variables and multivariate factual data, their maximum entropy is likely to be of hundreds of bits or more. The alphabet compression achieved by the analyst can reduce the cost of subsequent processes, including but not limited to the manager’s effort for processing data.
There are many factors that may affect the quality of the analyst’s recommendation Z rcmd a , including the analyst’s knowledge and experience, the critical nature of the missing information, and the access to data analysis and visualization tools. An inattentive portfolio manager may simply adopt the analyst’s recommendation. For example, an analyst may present a summary view (strong rise, multiple catalysts, high conviction) and then express “I’m 90% sure that the summary view is correct, though I am also OK to reject.” An inattentive portfolio manager makes a quick decision to accept the summary view. In terms of the alphabets in Figure 2, we have the PMFs:
P ( Z rcmd 1 a ) = 1 letter ( strong rise , multiple catalysts , high conviction ) 0 other 74 letters P ( Z rcmd 2 a ) = 0.9 letter accept 0.1 letter reject 0 other 2 letters P ( Z rcmd 2 b ) = 1 letter accept 0 letter reject 0 other 2 letters
A diligent portfolio manager evaluates the analyst’s recommendation more carefully, reasons about some or all missing information independently, and makes the final decision that may differ from the analyst’s recommendation, e.g.,
P ( Z rcmd 1 b ) = 0.4 letter ( strong rise , multiple catalysts , medium conviction ) 0.6 letter ( rise , multiple catalysts , medium conviction ) 0 other 73 letters P ( Z rcmd 2 b ) = 1 letter accept 0 other 3 letters
Assume that a post hoc analysis shows an ideal recommendation Z rcmd * such that
P ( Z rcmd 1 * ) = 1 letter ( rise , multiple catalysts , medium conviction ) 0 other 74 letters P ( Z rcmd 2 * ) = 1 letter accept 0 other 3 letters
We can consider that Z rcmd * represents the ground truth, and it could be discovered by an ideal process (e.g., without missing information). We can also consider that Z rcmd is a masked version Z rcmd * . The goal of the process in Figure 2 is thus to reconstruct Z rcmd * , i.e.,
Z rcmd * disguising Z rcmd processing Z rcmd a evaluating Z rcmd b
Note that this sequence of transformations is very similar to those found in typical empirical studies in psychology. For such a sequence, we can evaluate the potential distortion of the analyst’s recommendation and the diligent portfolio manager’s final decision. For example, using the above Z rcmd * , Z rcmd a , and Z rcmd b , we have the following:
analyst : H max ( Z rcmd * ) D Z rcmd * Z rcmd a 8.229 × 0.886 7.290 bits inattentive portfolio manager : H max ( Z rcmd * ) D Z rcmd * Z rcmd b 8.229 × 1.000 8.229 bits diligent portfolio manager : H max ( Z rcmd * ) D Z rcmd * Z rcmd b 8.229 × 0.214 1.762 bits
Let the PMF of the disguised alphabet Z rcmd be uniformly distributed. We can also calculate the alphabet compression achieved by the analyst, and the inattentive and the diligent portfolio managers:
analyst : H ( Z rcmd ) H ( Z rcmd a ) 8.229 0.469 7.760 bits inattentive portfolio manager : H ( Z rcmd ) H ( Z rcmd b ) 8.229 0.000 8.229 bits diligent portfolio manager : H ( Z rcmd ) H ( Z rcmd b ) 8.229 0.971 7.258 bits
We can calculate the informative benefit as follows:
analyst : 7.760 7.290 0.470 bits inattentive portfolio manager : 8.229 8.229 = 0 bits diligent portfolio manager : 7.258 1.762 5.496 bits
We can also estimate the costs of an analyst and a portfolio manager in terms of the amount of time they would be expected to perform the tasks concerned. For example, one may expect an analyst to spend 5 h and a portfolio manager to spend 1 h on the left and right parts of the tasks in Figure 2. We can estimate the cost benefit as follows:
AnalystInattentive PMDiligent PM
Benefit Cost 0.470 5 = 0.094 bits hour 0 1 = 0 bits hour 5.496 1 = 5.496 bits hour

4.3. Case Study 3: An Investment Committee Scenario

Following the portfolio manager’s decision to accept an investment thesis, it is frequently presented to an investment committee for approval. In Figure 1, this would be conducted as part of the Investment Thesis Development and Evaluation component. While members of the investment committee are typically experienced portfolio managers, they may have different specialities. Some may focus on equities within industries less relevant to the presented investment thesis. In Figure 1, this case study is labeled as “Case Study 3”. At this point in the pipeline, most funds have a policy regarding the mandatory involvement of an investment committee, which provides an additional governance layer and offers domain-specific expertise, soft knowledge, and risk evaluation.
Let us consider a scenario in which an investment committee is tasked with evaluating a recommendation by a portfolio manager in Case Study 2. The investment committee needs to make a decision, the alphabet of which is Z cmdc . Similar to Z rcmd , Z cmdc consists of two sub-alphabets, Z cmdc 1 and Z cmdc 2 . Z cmdc 1 has the same letters as Z rcmd 1 , while Z cmdc 2 has three of the four letters in Z rcmd 2 , i.e.,
Z cmdc 2 = { accept , reject , request   more   information }
The maximum entropy is H max ( Z cmdc ) 7.8 bits. Consider a case where a portfolio manager, PM, presents a proposal, Z cmdc PM , to the committee before the committee meeting:
P ( Z cmdc 1 PM ) = 0.8 letter ( strong decline , one short-term catalyst , high conviction ) 0.2 letter ( strong decline , only one medium-term catalyst , high conviction ) 0 other 73 letters P ( Z cmdc 2 PM ) = 1 letter accept 0 other 2 letters
At the beginning of the committee meeting, we have
Z cmdc 0 = Z cmdc 1 0 Z cmdc 2 0 = Z cmdc 1 PM { accept ( p = 0.6 ) , reject ( p = 0.3 ) , request more information ( p = 0.1 ) }
where the PMF (0.6, 0.3, 0.1) reflects the historical statistics of the committee’s decisions. Note that P ( Z cmdc 1 PM ) = P ( Z cmdc 1 0 ) , but P ( Z cmdc 2 PM ) P ( Z cmdc 2 0 ) .
Consider a committee that has four members:
  • PM—The portfolio manager who proposed the recommendation.
  • MS—A member of the committee with useful soft knowledge.
  • MP—A member of the committee with positive bias.
  • MN—A member of the committee with negative bias.
We also assume that the ground truth of Z cmdc 1 * and the ideal decision P ( Z cmdc 2 * ) are as follows:
P ( Z cmdc 1 * ) = 1 letter ( rise , multiple catalysts , medium conviction ) 0 other 74 letters P ( Z cmdc 2 * ) = 1 letter reject 0 other 2 letters
The member of the committee, MS, with a strong informational network, pattern recognition capability, and routine (i.e., soft knowledge), expresses a view that is represented by the alphabet Z cmdc MS :
P ( Z cmdc 1 MS ) = 0.9 letter ( rise , multiple catalysts , medium conviction ) 0.1 letter ( rise , multiple catalysts , low conviction ) 0 other 73 letters P ( Z cmdc 2 MS ) = 1 letter reject 0 other 2 letters
This is fairly close to Z cmdc * .
One member, MP, who has advised on the investment thesis before the committee and who holds an overall negative view of the industry, exhibits a strongly positive bias in favor of the trade with Z MP cmdc :
P ( Z cmdc 1 MP ) = 1 letter ( strong decline , one short-term catalyst , high conviction ) 0 other 74 letters P ( Z cmdc 2 MP ) = 1 letter accept 0 other 2 letters
Another member, MN, offers a view at a later stage, exhibiting a strong negative bias against the trade. Such cognitive bias may stem from the explicit competition for capital allocation or the need to enhance the relative merit of their trade, resulting in Z cmdc MN :
P ( Z cmdc 1 MN ) = 0.5 letter ( rise , multiple catalysts , high conviction ) 0.5 letter ( rise , multiple catalysts , medium conviction ) 0 other 73 letters P ( Z cmdc 2 MN ) = 1 letter reject 0 other 2 letters
We can compute the alphabet compression achieved by each committee member as if they do not know Z cmdc 0 (including PM):
PM ( the proposer ) : H max ( Z cmdc ) H ( Z cmdc PM ) 7.814 0.722 7.092 bits MS ( with soft knowledge ) : H max ( Z cmdc ) H ( Z cmdc MS ) 7.814 0.469 7.345 bits MP ( with positive bias ) : H max ( Z cmdc ) H ( Z cmdc MP ) 7.814 0.000 7.814 bits MN ( with negative bias ) : H max ( Z cmdc ) H ( Z cmdc MN ) 7.814 1.000 6.814 bits
However, if the views of MS, MP, and MN are formulated after seeing the proposal of PM, we compute the alphabet compression as follows:
PM ( the proposer ) : H ( Z cmdc PM ) H ( Z cmdc PM ) 0.722 0.722 = 0 bits MS ( with soft knowledge ) : H ( Z cmdc PM ) H ( Z cmdc MS ) 0.722 0.469 0.253 bits MP ( with positive bias ) : H ( Z cmdc PM ) H ( Z cmdc MP ) 0.722 0.000 0.722 bits MN ( with negative bias ) : H ( Z cmdc PM ) H ( Z cmdc MN ) 0.722 1.000 0.278 bits
Here, it is useful for us to observe the potential distortion (i) between each committee member’s view and the ground truth, and (ii) between each committee member’s view and the proposal.
PM ( the proposer ) : H max ( Z cmdc ) D Z cmdc * Z cmdc PM 7.814 × 0.791 6.182 bits MS ( with soft knowledge ) : H max ( Z cmdc ) D Z cmdc * Z cmdc MS 7.814 × 0.014 0.112 bits MP ( with positive bias ) : H max ( Z cmdc ) D Z cmdc * Z cmdc MP 7.814 × 1.000 7.814 bits MN ( with negative bias ) : H max ( Z cmdc ) D Z cmdc * Z cmdc MN 7.814 × 0.322 2.515 bits PM ( the proposer ) : H max ( Z cmdc ) D Z cmdc PM Z cmdc PM 7.814 × 0 = 0 bits MS ( with soft knowledge ) : H max ( Z cmdc ) D Z cmdc PM Z cmdc MS 7.814 × 0.677 5.290 bits MP ( with positive bias ) : H max ( Z cmdc ) D Z cmdc PM Z cmdc MP 7.814 × 0.0566 0.442 bits MN ( with negative bias ) : H max ( Z cmdc ) D Z cmdc PM Z cmdc MN 7.814 × 0.452 3.533 bits
It is in fact more important for us to observe the overall view of the committee, which is denoted as Z cmdc 1 and is a probabilistic combination of Z cmdc PM , Z cmdc MS , Z cmdc MP , and Z cmdc MN :
P ( Z cmdc 1 1 ) = 0.450 letter ( strong decline , one short-term catalyst , high conviction ) 0.050 letter ( strong decline , only one medium-term catalyst , high conviction ) 0.125 letter ( rise , multiple catalysts , high conviction ) 0.350 letter ( rise , multiple catalysts , medium conviction ) 0.025 letter ( rise , multiple catalysts , low conviction ) 0 other 70 letters P ( Z cmdc 2 1 ) = 0.5 letter accept 0.5 letter reject 0 letter request more information
The committee view at this stage, i.e., Z cmdc 1 , is less certain (i.e., less alphabet compression) but may be closer or further away from the ground truth Z cmdc * than individual views, i.e., Z cmdc PM , Z cmdc MS , Z cmdc MP , and Z cmdc MN :
alphabet compression ( committee ) : H max ( Z cmdc ) H ( Z cmdc 1 ) 7.814 1.773 6.041 bits potential distortion ( committee ) : H max ( Z cmdc ) D Z cmdc * Z cmdc 1 7.814 × 0.405 3.161 bits
Like most committee-based decision processes, the members of the committee spend time to deliberate on their respective views. This can be modeled as the progression of the committee alphabet from Z cmdc 0 to Z cmdc 1 , then Z cmdc 1.1 , Z cmdc 1.2 , Z cmdc 1.3 , …. Ultimately, the committee reaches a decision. In this case study, it is reasonable to assume that the members of the committee gradually are largely convinced by the arguments of MS (the member with soft knowledge), leading to the final committee decision alphabet, Z cmdc 2 , as follows:
P ( Z cmdc 1 2 ) = 0.8 letter ( rise , multiple catalysts , medium conviction ) 0.2 letter ( rise , multiple catalysts , low conviction ) 0 other 73 letters P ( Z cmdc 2 2 ) = 1 letter reject 0 letter request more information
The final decision alphabet of the committee, Z cmdc 2 , is more certain (that is, more alphabet compression) and is closer to the ground truth Z cmdc * in comparison with Z cmdc 0 and Z cmdc 1 :
alphabet compression ( committee ) : H max ( Z cmdc ) H ( Z cmdc 2 ) 7.814 0.722 7.092 bits potential distortion ( committee ) : H max ( Z cmdc ) D Z cmdc * Z cmdc 2 7.814 × 0.057 0.442 bits
Therefore, the informative benefit for the individual views, i.e., Z cmdc PM , Z cmdc MS , Z cmdc MP , and Z cmdc MN , and committee views, i.e., Z cmdc 0 , Z cmdc 1 , and Z cmdc 2 are as follows:
PM   ( the proposer ) : 7.092 6.182 0.910 bits MS   ( with soft knowledge ) : 7.345 0.112 7.233 bits MP   ( with positive bias ) : 7.814 7.814 = 0 bits MN   ( with negative bias ) : 6.814 2.515 4.299 bits CMDC   0   ( at beginning ) : 7.092 6.182 0.910 bits CMDC   1   ( before discussions ) : 6.041 3.161 2.880 bits CMDC   2   ( final decision ) : 7.092 0.442 6.650 bits
We estimate the costs of the committee in terms of the amount of time that they would be expected to spend during the committee meeting to make their decision. Assuming the investment committee meeting took 1.5 h, comprised 15 min for proposal presentation, 30 min for each member to share their initial views, and 45 min for discussions and final decisions, we estimate the cost–benefit ratios of the three committee stages as:
CMDC 0CMDC 1CMDC 2
Benefit Cost 0.910 0.250 = 3.640 bits hour 2.880 0.500 = 5.760 bits hour 6.650 0.750 = 8.867 bits hour
as well as the gradual reduction in potential distortion of the committee alphabet as follows:
D Z cmdc * Z cmdc 0 > D Z cmdc * Z cmdc 1 > D Z cmdc * Z cmdc 2
This highlights the merits of such a committee, although it also had biased committee members and an initially inappropriate proposal. We conclude that the investment committee process is cost-beneficial based on the information-theoretic analysis. As the investment committee meeting progresses from Z cmdc 0 to Z cmdc 2 , the informative benefit increases and the potential distortion decreases at every stage. As long as the discussions remain fact-based among the committee members, they can result in a committee PMF convergence of Z cmdc towards the ground truth PMF of Z cmdc * due to the soft knowledge and a mitigation of positive and negative biases.

5. Discussions

In this section, we first summarize and discuss the findings of the three case studies. We then consider the difference between economic cost–benefit analysis and information-theoretic cost–benefit analysis. This is followed by discussions on several related research topics, which present both challenges and opportunities in further development of the work reported in this paper.

5.1. The Findings of Three Case Studies

As this work aims to translate an information-theoretic concept and measure for cost–benefit analysis to a methodology that is relevant to business and finance, the findings of the three case studies are indicative within the scope of conceptual and methodological research. We present the general findings first, followed by the findings of each case study.
  • General:
  • The three case studies confirmed that the information-theoretic concept can be used to analyze the effectiveness of information transformations in various component processes in decision workflows in business and finance. The effectiveness, referred to as “benefit”, has two measures, alphabet compression and potential distortion. On the one hand, such transformations should ideally achieve as much alphabet compression as possible. On the other hand, they should ideally have as little potential distortion as possible. Hence, optimizing a process involves finding the optimal trade-off between these two measures.
  • As shown in Figure 1, the processes in the top-level workflow are too complex (i.e., too many PMFs to consider) or too vague (the PMFs are not well-defined) for us to conduct information-theoretic analysis. Therefore, it is necessary to decompose each process into several low-level processes. The decomposition procedure can be recursively invoked until each process is defined with a small number of PMFs. This is a typical engineering approach for workflow modeling and decomposition. The three case studies confirmed that this approach can also be applied to workflows in business and finance.
  • The three case studies feature one machine-centric process (i.e., Case Study 1) and two human-centric processes (i.e., Case Studies 2 and 3). They confirmed that the methodology can be applied to both machine- and human-centric processes.
  • The computation reported in the three case studies was carried out using tools such as spreadsheets. While this approach is effective in conceptual and methodological studies, it does not scale up to real-world workflows, where there are numerous processes and many PMFs. It is thus highly desirable to develop software tools for supporting such computation.
  • Case Study 1:
5.
This case study indicates that machine-centric processes usually produce very high alphabet compression at a low cost (excluding the development cost of the process).
6.
Some machine-centric processes have the risk of incurring high potential distortion.
7.
Data visualization, commonly used in business and finance, provides an effective means for alleviating the potential distortion. Data visualization usually reduces entropy less than statistics and algorithms, and the cost of using data visualization is relatively low.
  • Case Study 2:
8.
This case study demonstrates that information-theoretic cost–benefit analysis can be used to examine scenarios of missing information in data and additional information provided through human knowledge (i.e., the instance of a diligent portfolio manager).
9.
Information-theoretic analysis can quantitatively demonstrate that missing information may lead to higher potential distortion, thereby reducing the overall cost–benefit measure.
10.
Information-theoretic analysis can quantitatively show that soft knowledge can reduce potential distortion, increasing the informative benefit.
  • Case Study 3:
11.
This case study demonstrates that information-theoretic cost–benefit analysis can be used to examine scenarios that feature soft knowledge, as well as positive and negative biases.
12.
Information-theoretic analysis can quantitatively show that both positive and negative bias increase potential distortion, thereby reducing the cost–benefit of the process.
13.
Information-theoretic analysis can quantitatively demonstrate that committee meetings can be effective if they include members with useful soft knowledge.
14.
Information-theoretic analysis can quantitatively show that committee meetings can be effective if there is a reasonable balance between members with positive and those with negative bias.

5.2. Comparison with Traditional Economic CBA Methods

Traditional economic CBA methods (e.g., [12,15,45]) focus on monetary outcomes over time and seek beneficial risk-adjusted returns. Many of these methods have been extensively studied and are now widely used in business and finance. The work reported in this paper is not intended to replace any of these methods, but to address a gap that these methods cannot cover.
Today, many decision workflows involve both machine- and human-centric processes. At a higher level, one may use “monetary benefit” and “monetary cost” to characterize these processes and economic CBA methods are still applicable. At a lower level, each process receives raw data, processed data from preceding processes, and/or knowledge from humans, while generating data to be processed by the succeeding process. Using “monetary benefit” to characterize the benefit of such information transformations is rather coarse, to say the least.
Information-theoretic cost–benefit analysis offers a potential to analyze machine- and human-centric processes in hybrid decision workflows at such a lower level. Some may argue that such low-level analysis is unnecessary. This reminds us of the time when the concept of entropy was first proposed by Ludwig Boltzmann in 1877. Many scientists then considered that the modeling of thermodynamic phenomena using a concept based on microscopic behaviors was unnecessary, and the macroscopic measures (e.g., temperature, volume, density, etc.) were adequate. Today, the concept of entropy underpins not only thermodynamics but also information theory, data communication, and many computational subjects.
Meanwhile, we must recognize that economic CBA methods have been around for decades while Information-theoretic cost–benefit analysis was first proposed in 2016 [33]. There is no doubt a long road ahead to reach its maturity theoretically and to be developed into a technology that can be deployed in practical applications for optimizing hybrid decision workflows in business and finance.

5.3. Estimation of Probability Distribution, Uncertainty Analysis, and Sensitivity Analysis

This work focuses on the translation from the information-theoretic cost–benefit analysis to a methodology for analyzing hybrid decision workflows in business and finance. For this work, we assumed that all probability mass functions (PMFs) in the analysis have already been obtained. In practice, obtaining accurate PMFs is a significant challenge, and there is a large volume of previous work on PMF and entropy estimation (e.g., [46,47,48]), uncertainty analysis (e.g., [49,50,51,52]), sensitivity analysis (e.g., [53,54,55]), and interval computation (e.g., [56,57,58]). The conceptual work reported in this paper should be considered as an early step towards deployable software tools and real-world applications. Like many previous developments that transitioned from theoretical concepts to practical applications in business and finance, there will be numerous steps of technical advancement by researchers in the field. Such technical advancement will be necessary to address several challenges, which include, but are not limited to, the following:
  • PMFs Estimation—The credibility of information-theoretical analysis depends critically on the accuracy of the PMFs available at every stage of the analytical process. PMFs can be estimated from various types of data, including measured real-world data, survey data, and experimental data. To analyze a hybrid decision workflow, one has to consider a number of challenges, including frequent temporal changes to PMFs, global PMFs vs. local PMFs, sparsity samples of human decision patterns, and so on.
  • Uncertainty Analysis—The uncertainty of a uni-variable is often modeled as a distribution of the possible values of the variable. One can easily anticipate the complexity of modeling the uncertainty of a PMF, since there would be distributions of distributions. The complexity increases rapidly as the number of letters in a PMF and/or the number of PMFs whose interactions need to be analyzed grows.
  • Sensitivity Analysis—With information-theoretic analysis, changes to some PMFs often have more impact than changes to other PMFs. Sensitivity analysis allows one to identify important variables that contribute critically to the outcome (e.g., used in COVID-19 modeling [59]). Similar to uncertainty analysis, the sensitivity analysis of n PMFs is, in general, much more complex than that of n variables.
  • Interval Computation—Instead of specifying the distributions of each probability element p i in a PMF, one may use a simplified way to select the error bounds of each p i . Interval computation is a family of methods for computing with error bounds. The application of Interval computation to PMFs usually is more complex than that with multiple variables, as each PMF may have different error bounds for each p i in a PMF and the computation needs to maintain the validity of every PMF (i.e., p i [ 0 , 1 ] and p i = 1 ).
  • Software Tools—With the aforementioned challenges, one can anticipate that future research will be able to formulate mathematical solutions. To make these mathematical solutions practically useful, the key will be to develop software tools that enable the automation of such computations.
These challenges cannot be addressed by any single piece of research, but require much extensive research effort in the future.

6. Conclusions and Future Directions

In this paper, we presented a methodology for analyzing the cost–benefit of processes in data-informed decision workflows. The methodology is based on the theoretical development of information-theoretic cost–benefit analysis, as well as a practical approach to workflow decomposition in engineering. Using three case studies, we demonstrate the feasibility of conducting such an analysis quantitatively. This confirms the main research question about whether we can translate the information-theoretic concept and measure for cost–benefit analysis to a methodology for analyzing hybrid decision workflows in business and finance. It also confirms that the engineering approach for workflow modeling and decomposition is a valid and useful instrument for enabling the translation.
As three case studies feature machine- and human-centric processes (including a statistical and computational algorithm, incomplete information and human soft knowledge, and human biases in a committee meeting), they also demonstrate the general applicability of the proposed methodology.
We recognize that the method is currently not automated or supported by any computer-assisted tool. This is a typical limitation of any quantitative method that has not yet been supported by a computer-assisted tool. We strongly believe that future research will lead to the development of computer-assisted tools by building on past, current, and future theoretical, conceptual, and methodological research on information-theoretic cost–benefit analysis, including this work. Meanwhile, we will continue our methodological research by applying information-theoretic cost–benefit analysis to processes in machine learning workflows.

Author Contributions

Conceptualization, P.B. and M.C.; methodology, P.B., M.C. and H.M.; validation, H.M. and P.B.; formal analysis, P.B. and M.C.; investigation, P.B.; resources, P.B.; data curation, P.B.; writing—original draft preparation, P.B.; writing—review and editing, M.C. and H.M.; supervision, M.C. and H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This is not applicable to this research.

Data Availability Statement

The data and calculations for the case studies are presented in the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACAlphabet Compression
BelBelieved
CBACost–Benefit Analysis
CmdcCommittee Decision
CSCase Study
CtCost
JSJenson-Shannon
KLKullback–Leibler
MASimple Moving Average
MDPsMarkovian Decision Processes
MissMissing Information
MNCommittee Member with Negative Bias
MPCommittee Member with Positive Bias
MSCommittee Member with Soft Knowledge
PDPotential Distortion
PMPortfolio Manager
PMFProbability Mass Function
RefReference
RcmdRecommendations
SvShare Value
VisVisualization
XAIExplainable Artificial Intelligence

References

  1. Ajah, I.A.; Nweke, H.F. Big Data and Business Analytics: Trends, Platforms, Success Factors and Applications. Big Data Cogn. Comput. 2019, 3, 32. [Google Scholar] [CrossRef]
  2. Drucker, P.F. The Practice of Management; Harper & Row: New York, NY, USA, 1954. [Google Scholar]
  3. Brynjolfsson, E.; Hitt, L.M.; Kim, H.H. Strength in Numbers: How Does Data-Driven Decisionmaking Affect Firm Performance? SSRN Electron. J. 2011, 1, 1–33. [Google Scholar] [CrossRef]
  4. Provost, F.; Fawcett, T. Data Science and its Relationship to Big Data and Data-Driven Decision Making. Big Data 2013, 1, 51–59. [Google Scholar] [CrossRef] [PubMed]
  5. Sun, S.X.; Zhao, J.L.; Nunamaker, J.F.; Sheng, O.R.L. Formulating the Data-Flow Perspective for Business Process Management. Inf. Syst. Res. 2006, 17, 374–391. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Perry, D.E. A Data-Centric Approach to Optimize Time in Workflow-Based Business Process. In Proceedings of the 2014 IEEE International Conference on Services Computing, Anchorage, AK, USA, 27 June–2 July 2014; pp. 709–716. [Google Scholar] [CrossRef]
  7. Kougka, G.; Gounaris, A.; Simitsis, A. The many faces of data-centric workflow optimization: A survey. Int. J. Data Sci. Anal. 2018, 6, 81–107. [Google Scholar] [CrossRef]
  8. Hansen, K.A.; Hillenbrand, S.F.; Ungerleider, L.G. Human Brain Activity Predicts Individual Differences in Prior Knowledge Use during Decisions. J. Cogn. Neurosci. 2012, 24, 1462–1475. [Google Scholar] [CrossRef] [PubMed]
  9. Tam, G.K.L.; Kothari, V.; Chen, M. An Analysis of Machine- and Human-Analytics in Classification. IEEE Trans. Vis. Comput. Graph. 2017, 23, 71–80. [Google Scholar] [CrossRef] [PubMed]
  10. Kijmongkolchai, N.; Abdul-Rahman, A.; Chen, M. Empirically Measuring Soft Knowledge in Visualization. Comput. Graph. Forum 2017, 36, 73–85. [Google Scholar] [CrossRef]
  11. Fisher, I. The Rate of Interest: Its Nature, Determination and Relation to Economic Phenomena; The Macmillan Company: New York, NY, USA, 1907. [Google Scholar]
  12. Lorie, J.H.; Savage, L.J. Three Problems in Rationing Capital. J. Bus. 1955, 28, 229. [Google Scholar] [CrossRef]
  13. Fisher, I. The Theory of Interest: As Determined by Impatience to Spend Income and Opportunity to Invest It; The Macmillan Company: New York, NY, USA, 1930. [Google Scholar]
  14. Howard, R. Information Value Theory. IEEE Trans. Syst. Sci. Cybern. 1966, 2, 22–26. [Google Scholar] [CrossRef]
  15. Boardman, A.E.; Greenberg, D.H.; Vining, A.R.; Weimer, D.L. Cost-Benefit Analysis; Cambridge University Press: Cambridge, UK, 2018. [Google Scholar] [CrossRef]
  16. Ganat, T.A.A.O. CAPEX and OPEX Expenditures. In Technical Guidance for Petroleum Exploration and Production Plans; Springer Nature: Cham, Switzerland, 2020; pp. 53–56. [Google Scholar] [CrossRef]
  17. Bellman, R. Dynamic Programming; Princeton University Press: Princeton, NJ, USA, 1957. [Google Scholar]
  18. Kelley, J.E.; Walker, M.R. Critical-path planning and scheduling. In Proceedings of the IRE-AIEE-ACM ’59 (Eastern): Eastern Joint IRE-AIEE-ACM Computer Conference, Boston, MA, USA, 1–3 December 1959; pp. 160–173. [Google Scholar] [CrossRef]
  19. Hammer, M.; Champy, J. Reengineering the corporation: A manifesto for business revolution. Bus. Horizons 1993, 36, 90–91. [Google Scholar] [CrossRef]
  20. Howard, R.A.; Matheson, J.E. Influence Diagrams. Decis. Anal. 2005, 2, 127–143. [Google Scholar] [CrossRef]
  21. Belousov, B.; Peters, J. Entropic Regularization of Markov Decision Processes. Entropy 2019, 21, 674. [Google Scholar] [CrossRef] [PubMed]
  22. Habermehl, C.; Höpfner, G.; Berroth, J.; Neumann, S.; Jacobs, G. Optimization Workflows for Linking Model-Based Systems Engineering (MBSE) and Multidisciplinary Analysis and Optimization (MDAO). Appl. Sci. 2022, 12, 5316. [Google Scholar] [CrossRef]
  23. Barbiero, P.; Ciravegna, G.; Giannini, F.; Lió, P.; Gori, M.; Melacci, S. Entropy-Based Logic Explanations of Neural Networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 6046–6054. [Google Scholar] [CrossRef]
  24. Sanneman, L.; Tucker, M.; Shah, J.A. An Information Bottleneck Characterization of the Understanding-Workload Tradeoff in Human-Centered Explainable AI. In Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency, Rio de Janeiro, Brazil, 3–6 June 2024; pp. 2175–2198. [Google Scholar] [CrossRef]
  25. Mehdi, S.; Tiwary, P. Thermodynamics-inspired explanations of artificial intelligence. Nat. Commun. 2024, 15, 7859. [Google Scholar] [CrossRef] [PubMed]
  26. Bang, S.; Xie, P.; Lee, H.; Wu, W.; Xing, E. Explaining a Black-box By Using a Deep Variational Information Bottleneck Approach. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021. [Google Scholar]
  27. Dutta, S.; Hamman, F. A Review of Partial Information Decomposition in Algorithmic Fairness and Explainability. Entropy 2023, 25, 795. [Google Scholar] [CrossRef] [PubMed]
  28. Kim, J.Y.; Cho, S.B. An information theoretic approach to reducing algorithmic bias for machine learning. Neurocomputing 2022, 500, 26–38. [Google Scholar] [CrossRef]
  29. Li, J.; Khayatkhoei, M.; Zhu, J.; Xie, H.; Hussein, M.E.; Abdalmageed, W. Information-Theoretic Bounds on The Removal of Attribute-Specific Bias From Neural Networks. arXiv 2023, arXiv:2310.04955. [Google Scholar]
  30. Omar, Y.M.; Plapper, P. Entropy of complex manufacturing networks as a metric of flexibility. J. Ind. Inf. Integr. 2022, 27, 100285. [Google Scholar] [CrossRef]
  31. Herrera-Vidal, G.; Coronado-Hernández, J.R.; Derpich-Contreras, I.; Paredes, B.P.M.; Gatica, G. Measuring Complexity in Manufacturing: Integrating Entropic Methods, Programming and Simulation. Entropy 2025, 27, 50. [Google Scholar] [CrossRef] [PubMed]
  32. Chen, M.; Jänicke, H. An Information-theoretic Framework for Visualization. IEEE Trans. Vis. Comput. Graph. 2010, 16, 1206–1215. [Google Scholar] [CrossRef] [PubMed]
  33. Chen, M.; Golan, A. What May Visualization Processes Optimize? IEEE Trans. Vis. Comput. Graph. 2016, 22, 2619–2632. [Google Scholar] [CrossRef] [PubMed]
  34. Golan, A. Information and Entropy Econometrics—A Review and Synthesis. Found. Trends® Econom. 2008, 2, 1–145. [Google Scholar] [CrossRef]
  35. Hansen, L.P. Large Sample Properties of Generalized Method of Moments Estimators. Econometrica 1982, 50, 1029–1054. [Google Scholar] [CrossRef]
  36. Golan, A.; Judge, G.; Miller, D. Maximum Entropy Econometrics: Robust Estimation with Limited Data; Wiley: Hoboken, NJ, USA, 1996. [Google Scholar]
  37. Golan, A.; Karp, L.S.; Perloff, J.M. Estimating Coke’s and Pepsi’s Price and Advertising Strategies. J. Bus. Econ. Stat. 2000, 18, 398–409. [Google Scholar] [CrossRef]
  38. Chen, M.; Ebert, D.S. An Ontological Framework for Supporting the Design and Evaluation of Visual Analytics Systems. Comput. Graph. Forum 2019, 38, 131–144. [Google Scholar] [CrossRef]
  39. Kullback, S.; Leibler, R.A. On Information and Sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  40. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  41. Chen, M.; Sbert, M. A Bounded Measure for Estimating the Benefit of Visualization (Part I): Theoretical Discourse and Conceptual Evaluation. Entropy 2022, 24, 228. [Google Scholar] [CrossRef] [PubMed]
  42. Chen, M.; Abdul-Rahman, A.; Silver, D.; Sbert, M. A Bounded Measure for Estimating the Benefit of Visualization (Part II): Case Studies and Empirical Evaluation. Entropy 2022, 24, 282. [Google Scholar] [CrossRef] [PubMed]
  43. Chen, M. Cost-Benefit Analysis of Data Intelligence—Its Broader Interpretations. In Advances in Info-Metrics; Oxford University Press: New York, NY, USA, 2020; pp. 433–463. [Google Scholar] [CrossRef]
  44. Gurrib, I. The Moving Average Crossover Strategy: Does it Work for the S&P500 Market Index? SSRN Electron. J. 2014, 1–24. [Google Scholar] [CrossRef]
  45. Mishan, E.; Quah, E. Cost-Benefit Analysis; Routledge: Abingdon, UK, 2020. [Google Scholar] [CrossRef]
  46. Kang, H.J.; Kim, K.; Kim, J.H. Approximating Optimally Discrete Probability Distribution with kth-order Dependency for Combining Multiple Decisions. Inf. Process. Lett. 1997, 62, 67–75. [Google Scholar] [CrossRef]
  47. Gamboa, F.; Loubes, J.M.; Rochet, P. Maximum Entropy Estimation for Survey Sampling. J. Stat. Plan. Inference 2011, 141, 305–317. [Google Scholar] [CrossRef]
  48. Melucci, M. A Brief Survey on Probability Distribution Approximation. Comput. Sci. Rev. 2019, 33, 91–97. [Google Scholar] [CrossRef]
  49. Hughes, I. Measurements And Their Uncertainties: A Practical Guide to Modern Error Analysis; Oxford University Press: New York, NY, USA, 2009. [Google Scholar]
  50. Hofer, E. The Uncertainty Analysis of Model Results; Springer International Publishing: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  51. Kadane, J.B. Principles of Uncertainty; Chapman and Hall/CRC: Boca Raton, FL, USA, 2020. [Google Scholar] [CrossRef]
  52. Morrison, F. Uncertainty Analysis for Engineers and Scientists: A Practical Guide; Cambridge University Press: Cambridge, UK, 2021; pp. 1–370. [Google Scholar]
  53. Saltelli, A.; Chan, K.; Scott, E. Sensitivity Analysis: Gauging the Worth of Scientific Models; Wiley Series in Probability Statistics; Wiley: Hoboken, NJ, USA, 2000; Volume 535, pp. 1–512. [Google Scholar]
  54. Saltelli, A.; Tarantola, S.; Campolongo, F.; Ratto, M. Sensitivity Analysis in Practice: A Guide to Assessing Scientific Models; John Wiley & Sons INC International Concepts: Hoboken, NJ, USA, 2004; pp. 1–232. [Google Scholar]
  55. Kozlova, M.; Yeomans, J.S. Sensitivity Analysis for Business, Technology, and Policymaking; Routledge: London, UK, 2024. [Google Scholar] [CrossRef]
  56. Dawood, H. Theories of Interval Arithmetic: Mathematical Foundations and Applications; Lambert Academic Publishing: Saarbrücken, Germany, 2011; pp. 1–128. [Google Scholar]
  57. Bridger, M. Real Analysi: A Constructive Approach Through Interval Arithmetic; American Mathematical Society: Providence, RI, USA, 2019; p. 302. [Google Scholar]
  58. Kreinovich, V.; Lauter, C. Books on Interval Computations. Available online: https://www.reliable-computing.org/books.html (accessed on 9 July 2025).
  59. Rydow, E.; Borgo, R.; Fang, H.; Torsney-Weir, T.; Swallow, B.; Porphyre, T.; Turkay, C.; Chen, M. Development and evaluation of two approaches of visual sensitivity analysis to support epidemiological modelling. IEEE Trans. Vis. Comput. Graph. 2023, 29, 1255–1265. [Google Scholar] [PubMed]
Figure 1. Decomposition of a data-informed decision workflow for discretionary long/short equity trading.
Figure 1. Decomposition of a data-informed decision workflow for discretionary long/short equity trading.
Entropy 27 00780 g001
Figure 2. The detailed alphabets and transformations in the “Case Study 2” part of the workflow in Figure 1.
Figure 2. The detailed alphabets and transformations in the “Case Study 2” part of the workflow in Figure 1.
Entropy 27 00780 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Beaucamp, P.; Maylor, H.; Chen, M. Information-Theoretic Cost–Benefit Analysis of Hybrid Decision Workflows in Finance. Entropy 2025, 27, 780. https://doi.org/10.3390/e27080780

AMA Style

Beaucamp P, Maylor H, Chen M. Information-Theoretic Cost–Benefit Analysis of Hybrid Decision Workflows in Finance. Entropy. 2025; 27(8):780. https://doi.org/10.3390/e27080780

Chicago/Turabian Style

Beaucamp, Philip, Harvey Maylor, and Min Chen. 2025. "Information-Theoretic Cost–Benefit Analysis of Hybrid Decision Workflows in Finance" Entropy 27, no. 8: 780. https://doi.org/10.3390/e27080780

APA Style

Beaucamp, P., Maylor, H., & Chen, M. (2025). Information-Theoretic Cost–Benefit Analysis of Hybrid Decision Workflows in Finance. Entropy, 27(8), 780. https://doi.org/10.3390/e27080780

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop