Next Article in Journal
Reinforcement Learning Approach for Adaptive C-V2X Resource Management
Previous Article in Journal
Edge-Computing-Based People-Counting System for Elevators Using MobileNet–Single-Stage Object Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Financial Data Quality Evaluation Method Based on Multiple Linear Regression

School of Software Engineering, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Future Internet 2023, 15(10), 338; https://doi.org/10.3390/fi15100338
Submission received: 20 August 2023 / Revised: 24 September 2023 / Accepted: 28 September 2023 / Published: 14 October 2023
(This article belongs to the Section Big Data and Augmented Intelligence)

Abstract

:
With the rapid growth of customer data in financial institutions, such as trusts, issues of data quality have become increasingly prominent. The main challenge lies in constructing an effective evaluation method that ensures accurate and efficient assessment of customer data quality when dealing with massive customer data. In this paper, we construct a data quality evaluation index system based on the analytic hierarchy process through a comprehensive investigation of existing research on data quality. Then, redundant features are filtered based on the Shapley value, and the multiple linear regression model is employed to adjust the weight of different indices. Finally, a case study of the customer and institution information of a trust institution is conducted. The results demonstrate that the utilization of completeness, accuracy, timeliness, consistency, uniqueness, and compliance to establish a quality evaluation index system proves instrumental in conducting extensive and in-depth research on data quality measurement dimensions. Additionally, the data quality evaluation approach based on multiple linear regression facilitates the batch scoring of data, and the incorporation of the Shapley value facilitates the elimination of invalid features. This enables the intelligent evaluation of large-scale data quality for financial data.

1. Introduction

Quality evaluation commonly pertains to the methodologies that organizations employ to ensure the attainment of strategic objectives, establish regulatory compliance, and monitor data integrity [1]. In the past, data collected from various transaction systems within companies or institutions were often regarded as a mere by-product of business operations, possessing limited value beyond the transactions [2]. Today, as the complexity and volume of data continue to escalate, numerous companies and businesses are compelled to adapt their business models accordingly. Customers need personalized products, and service products must be industrialized. These factors inevitably influence business processes and organizational strategies, and high-quality data are a prerequisite for meeting these changing business demands and achieving corporate agility objectives [3]. A single data management solution is insufficient to meet the demands of the business. Therefore, different approaches to integrating solutions to data problems must be employed [4].
Financial institutions must ensure the completeness and accuracy of their data quality so that regulatory bodies can better understand the operation of the institutions, predict risks, and take corresponding measures [5]. In the era of rapid big data technology advancement, financial institutions have been compelled to enhance their management, processing, and analysis of vast volumes of personal customer information. This imperative arises from the goals to enhance business efficiency and strengthen risk management capabilities. Institutions are increasingly utilizing machine learning technologies to analyze customer behavior, predict market trends, and identify fraudulent behaviors. These technologies also must be able to support a large amount of high-quality data; therefore, financial institutions must establish effective data governance and evaluation mechanisms to ensure data quality.
Mcgilvray [6] introduced a methodical yet adaptable framework consisting of ten steps to address an organization’s business needs and data quality issues. Within this framework, data quality dimensions are employed to precisely define, measure, and effectively manage the quality of data. Omara et al. [7] presented a comprehensive compilation of prevalent data quality dimensions, encompassing accuracy, completeness, consistency, and timeliness. Furthermore, they proposed a novel approach to gauge row completeness utilizing a data mining model developed on neural networks. Peltier et al. [8] measured data quality from a systems perspective with more dimensions and considered customer data quality to facilitate the development of personalized interactive marketing initiatives. Taleb et al. [9] developed a scheme to evaluate data quality on large datasets using sampling strategies, which adopted some data quality measurement metrics tailored to specific data quality dimensions. Juddoo [10] conducted an investigation on a variety of data quality metrics, with a specific focus on the measurement and quantification of particular dimensions like completeness and consistency. Recently, the analytic hierarchy process (AHP) was used to integrate various data quality dimensions as well as expert preferences for quantifying a comprehensive score of data quality [11], and has been applied to healthcare [12], financial statements [13] and cloud computing [14].
For the data quality evaluation system, most experts only consider qualitative or quantitative situations, citing a single data quality evaluation method, such as fault tree analysis, AHP, gray relational degree analysis, etc. Some scholars combine commonly used engineering evaluation methods to build evaluation models on the basis of combining qualitative and quantitative methods. Through research and combining, this paper finds that many scholars have not unified the selection of data quality indicators, and most of them use more traditional methods, which lack certain innovation. This leads to the main contributions of this paper as listed below:
  • We proposed a new system for constructing quality evaluation indicators by surveying the current status of data quality evaluation research. This system forms primary indicators according to completeness, accuracy, timeliness, consistency, uniqueness, and compliance. We used AHP to determine the weight among these indicators.
  • We proposed a data quality evaluation method based on multiple linear regression (MLP) to objectively evaluate data quality. In addition, we employed et al. Shapley value for feature selection to reduce the complexity of the model and shorten the training time.
  • By using customer-related table information from a specific financial institution as an example, we verified the feasibility and accuracy of the quality evaluation system and the quality evaluation model. Additionally, we found that the evaluation and processing of incremental customer form information could be automated.
The remainder of this paper is organization as follows: In Section 2, we formulate the primary indicators and adopt the AHP to determine the weights of the indicators. In Section 3, we construct a quality evaluation model using MLP algorithms. We next utilize a concept of the Shapley value to evaluate the contribution of each indicator to the scoring model, facilitating feature selection. In Section 4, we use the customer information system of a trust company as an example for empirical analysis. In Section 5, we summarize the results of this study and discuss its contributions.

2. Construction of Quality Evaluation Indicator System

The construction of the evaluation system should consider the following points data [15]:
  • Quality evaluation focuses on each dataset as the target. Every department is responsible for managing multiple datasets [16]. These datasets are typically represented as information source tables.
  • Data usability is the fundamental prerequisite for effective data utilization. The data resource centralization department can contribute by identifying factors that impede the basic usability of the data. These factors include empty data items, irregular formats, incomplete data, and other similar issues (Table 1).
  • Data flexibility takes center stage as a crucial consideration. When evaluating the quality of customer data types, it is essential to consider its unique characteristics while maintaining a certain level of flexibility. For instance, empty fields may be the result of non-required fields [17].
Because of multiple systems and vast amounts of data, it is crucial to strike a balance between quality and efficiency of the evaluation process [18]. To address this, a quality evaluation index framework for trustworthy data should be established, consisting of six dimensions: completeness, accuracy, timeliness, consistency, uniqueness, and compliance.

2.1. Basic Attributes of Quality Evaluation

A completeness evaluation encompasses three secondary indicators: record completeness, attribute completeness, and value completeness. The following fields should be assigned:
  • Required fields: These fields must be filled in accordance with the business rules of each department or according to the specifications outlined in the data dictionary.
  • Key fields: These fields serve as unique primary keys or facilitate associations with related data tables. Examples of such key fields include the identity card number, the unified social credit identifier, and other identification numbers or codes that act as primary keys in related associations.
Data accuracy encompasses three aspects: logical accuracy, value accuracy, and conceptual accuracy (Table 2). If the identity card number for a piece of data is not accurate or does not correspond to the name, this identity card number is considered to be one piece of erroneous data.
The timeliness requirement differs depending on the update frequencies of various information resource tables. As a result, rules can be established differently [19].
A consistency evaluation measures the level of data association, assessing the logical accuracy and completeness of the same information subject across different datasets [20].
A uniqueness evaluation gauges the extent of data duplication. For a dataset that consists of four fields, customer number, customer name, customer status, and gender, if N data records have the same values for all four fields, it would produce N-1 duplicate data records. As presented in Table 3, we identified a total of two duplicate data records.
Compliance ensures that the data types and precision of each data field fall within the specified range. For those fields governed by industry unified standards, adherence to the relevant standards is required. In the absence of unified standards, compliance is based on the data dictionary or relevant regulations provided by each department.

2.2. Determination of Indicator Weights Based on the AHP

The AHP is characterized by its ability to convert judgments into comparisons of importance between pairs of several factors [21]. The complete process involves decomposition, judgment, synthesis, and other steps, effectively addressing the limitations of alternative methods that aim to minimize subjective judgment from decision makers.

2.2.1. Single Scoring

To score a dataset from an information resource table, the scores of each secondary indicator can be computed using the previous scoring rules. The total score of the evaluated dataset can then be calculated by taking the weighted average.
Y = n λ n X n
where Y is the total score of the evaluated data set, λ n is the indicator weight, X n is the indicator score, and n is the secondary indicator number.
In the evaluation of a data set from an information resource table, the scoring range for each indicator ranges from 0 to 100 points. The scoring rules and formula are detailed in Table 4.

2.2.2. Comprehensive Scoring

Trusts continuously update the collected data. Therefore, if it is necessary to showcase the overall quality of a data set from a specific information resource table over a certain time frame, the numerous evaluation outcomes within this period can be comprehensively processed:
S i = k Y k K i
where Y k represents the single data quality evaluation score of the information resource data table, k is the order of evaluation, K i represents the total number of times the data table is evaluated during the evaluation period, and i is the number of each information resource table.

3. Construction of Quality Evaluation Model Based on Multiple Liner Regression

In this study, we present a machine learning-based data quality evaluation method. The method revolves around the creation of a data quality evaluation indicator system. The key steps of this approach are as follows:
  • Select relevant customer information data from the customer information database, and preprocess the data based on the proposed label generation algorithm.
  • Extract quality evaluation indicators and associated data, constructing a quality evaluation indicator system suitable for machine learning models.
  • Train and test the machine learning model to complete its construction.
  • Evaluate the results using performance indicators.

3.1. Data Acquisition and Indicator Extraction

As shown in Figure 1, we extracted the table being evaluated from the ECIF (Enterprise Customer Information Facility) database by employing the data quality evaluation indicators. Subsequently, to facilitate the implementation of the machine learning model, we had to transform the actual customer data into a trainable dataset. Therefore, in this study, by leveraging the constructed data quality evaluation indicator system, we introduced a related label generation algorithm. Subsequently, we used the overall score of the single-table quality evaluation as the output indicator, which represents the data quality. Meanwhile, we used the scores of the six primary indicators as the input indicators for the machine learning-based quality evaluation model.

3.2. Label Generation Algorithm

Suppose there are p primary indicators ( i = 1 , 2 , , p ), and each primary indicator i has q i secondary indicators. A given data table is scored based on the data quality evaluation system established. Specifically, Equation (3) is utilized to generate the corresponding labels for this data table:
y = i = 1 p α i j = 1 q i β j x j
where α i represents the weight corresponding to each primary indicator, and β j represents the weight corresponding to the secondary indicator. In the end, each table is associated with a quality evaluation score, which serves as the label for that table.
On the basis of the existing large data table, we randomly sampled different numbers of data entries and generated data subtables. We then applied the label generation algorithm to produce labels y n for each table. Next, we compiled a trainable assessment table X = { X i } , X i = { x j } based on each table’s primary indicators and their respective final quality evaluation scores. This process led to the formation of a dataset that includes n samples.
To examine the relationships among the dependent variables and between the dependent and independent variables, we conducted a Pearson correlation analysis of the data. Once the correlations among the indicators were determined, we were able to identify and remove the correlated indicators to streamline the model and reduce the training time.

3.3. Feature Selection Based on Shapley Value

Feature selection and hyperparameter tuning are two important steps in every machine learning task. They help improve performance most of the time but have the disadvantage of being expensive in time. The more parameter combinations, or the more precise the selection process, the longer the duration. The usual approach is to combine tuning and feature selection. For feature selection, we adopt a ranking-based selection algorithm. Rank selection involves iteratively removing less important features while retraining the model until convergence is reached. SHAP helps when we perform feature selection using ranking-based algorithms. Instead of using the default variable importances generated by gradient boosting, we choose the best features, e.g., the one with the highest Shapley value.
The Shapley value originates from cooperative game theory and is a distribution method based on contribution. It requires that profits and costs be fairly shared among each agent in an alliance [22]. The Shapley value has numerous applications in machine learning, including data pricing, federated learning, interpretability, reinforcement learning, and feature selection [23].
The equation for the Shapley value of n members is as follows [24]:
ϕ i ( v ) = S N , i S ( | S | 1 ) ! ( n | S | ) ! n ! [ v ( S ) v ( S { i } ) ] , i = 1 , , n ,
where ( | S | 1 ) ! ( n | S | ) ! n ! represents the probability of member i joining the alliance ( | S | i ) . The denominator represents the number of permutations of n members. The numerator represents the number of permutations of the first ( | S | 1 ) members entering the alliance ( | S | i ) . Upon joining the alliance, member i acquires a certain value, denoted as [ v ( S ) v ( S { i } ) ] . This value represents the contribution that member i makes to the alliance S.
In numerous scenarios involving data mining and machine learning, we frequently encounter challenges, such as the high dimensionality of features [25] and unknown relationships. In this study, we adopted the Powershap algorithm to select and eliminate indicators [26].
The following steps outline the automated feature selection methodology implemented through Powershap, which is constructed based on the foundation of the genetic algorithm:
  • Population initialization: A set of feature subsets is randomly produced, marking the formation of the initial population.
  • Fitness evaluation: A fitness function is implemented to assess the aptness of each entity, employing metrics like accuracy, F1 score, and so forth.
  • Selection process: Individuals exhibiting higher fitness are chosen to serve as the base for the next-generation population as determined by the fitness function assessment results.
  • Crossover procedure: Using a crossover operation, select individuals with superior fitness are randomly amalgamated to generate new offspring.
  • Mutation procedure: To enhance the population’s diversity, random alterations are introduced to the new offspring through a mutation operation.
  • Iteration: Steps 2–5 are reiterated until the predetermined stop conditions are satisfied.

3.4. Regression Model Building

Linear regression is a linear model that assumes a linear relationship between an input variable and a single output variable.Specifically, using the linear regression model, the output variable y can be calculated from the linear combination of a set of input variables x, that is, y = a x + b . If there are two or more independent variables, such linear regression analysis is called multiple linear regression (MLP). The multiple linear regression model is a relatively simple regression model in machine learning. After using the Shapley value to eliminate model features, the weights calculated based on AHP will no longer be accurate. Therefore, consider using MLP to recalculate the weights.
In the MLP model, it assumes that if the secondary indicators are reasonably set, the quality of the data table is considered only from the dimensions of the primary indicators [27]. If the primary indicator value and the quality of the data table are related linearly, a multivariate regression model can be established [28]: y = i p a i x i + b . The most suitable { a i } can be solved by solving the MLP problem. We employed the MLP solution method, specifically gradient descent, to determine the optimal parameters for the indicators. This was achieved after eliminating irrelevant features from the model.

4. Results

Financial institutions, such as trusts, have a large number of customers. Analyzing the current status of customers is extremely important for enhancing the accuracy of customer marketing and improving the user experience of customer services [29]. This section takes the personal customer table and institutional customer table in the customer information integration system of a trust institution as examples to illustrate the practical application of the data quality evaluation model.

4.1. Quality Evaluation Indicator System

By following the AHP procedure and the selected primary and secondary indicators in Figure 2, the decision makers have to indicate preferences or priority for each decision alternative in terms of how it contributes to each criterion as shown in Table 5.
Then, the following can be performed manually or automatically by the AHP software Expert Choice:
  • Synthesizing the pair-wise comparison matrix (example: Table 5).
  • Calculating the priority vector for a criterion such as experience (example: Table 5).
  • Calculating the consistency ratio.
  • Calculating λ m a x .
  • Calculating the consistency index, C I .
  • Selecting appropriate value of the random consistency ratio from Table 5.
  • Checking the consistency of the pair-wise comparison matrix to check whether the decision maker’s comparisons were consistent or not.
w is the eigenvector corresponding to the largest eigenvalue.
Now, we find the consistency index, CI, as follows:
C I = λ m a x n n 1 = 6.068 6 6 1 = 0.0136
Selecting an appropriate value of random consistency ratio R I for a matrix size of five using Table 5, we find R I = 1.24 . We then calculate the consistency ratio C R as follows:
C R = C I R I = 0.0136 1.24 = 0.011
As the value of C R is less than 0.1, the judgments are acceptable. Similarly, the pair-wise comparison matrices and priority vectors for the remaining criteria can be found as shown in Table 6, Table 7, Table 8 and Table 9, respectively.
Regarding completeness, out of the 264,762 records, including 14,709 data items and 18 attributes, we obtained 67,839 records with null values. This accounted for approximately 25.62% of the data, resulting in a completeness score of 81.70.
Concerning accuracy, among the 14,709 data records, we did not have any instances of data being updated or created later than the current time. Additionally, we had 5191 records with expired certificates. Roughly 35.29% of the data exhibited accuracy issues, resulting in an accuracy score of 91.58.
In terms of timeliness, we did not identify any problems within the 14,709 data records analyzed, warranting a perfect timeliness score of 100.
In terms of consistency, out of the 14,709 data records, 13,869 records belonging to the same information subject within the dataset did not meet the criteria for logical accuracy. This suggested that approximately 94.29% of the data exhibited consistency issues, resulting in a consistency score of 5.71.
Regarding uniqueness, we did not find any instances of noncompliant record uniqueness among the 14,709 data records, resulting in a uniqueness score of 100.
In terms of compliance, among the 14,709 data records, 169 records did not conform to the requirement of 18-digit identity card numbers. Further scrutiny of the birth year, month, day, and last digit did not reveal any records that violated the year and month requirements. However, 1380 records did not comply with the day requirement. Additionally, for contact information, 4908 records had noncompliant mobile phone numbers. Overall, the number of noncompliant data items reached 6457, accounting for approximately 21.95% of the total data. Consequently, the compliance score was calculated as 58.43.
Considering the six dimensions of data quality and the data quality evaluation model, the individual quality evaluation score for the consignment customer form was determined to be 82.39 as shown in Table 10 and Figure 3.

4.2. Multiple Linear Regression Quality Evaluation Model

We created the evaluation dataset using the label generation algorithm and the quality evaluation indicator system. The descriptive statistical results of various indicators of the trust agency’s individual customer table and institutional customer table are shown in Table 11 and Table 12.
To examine the presence of correlations among dependent variables and between the dependent and independent variables, we performed a Pearson correlation analysis of the data. The results were visualized through a heatmap as depicted in Figure 4, showcasing the correlations between the variables. In the heatmap, black represents a negative correlation, white represents a positive correlation, and the diagonal line represents the correlation of each independent variable with itself, resulting in a correlation coefficient of 1.
We observed that the correlation between the various indicator variables was not notably high, suggesting their independence from each other and the absence of multicollinearity. Moreover, the correlation between the independent variables and dependent variables was quite strong. Among these correlations, the compliance_score exhibited the highest correlation, with a coefficient of 0.93. This result indicated that compliance, as one of the first-layer indicators, exerted the most significant influence on the final quality score. Additionally, the integrityạccuracy_score also demonstrated notable correlations with the quality score, with coefficients of 0.64 and 0.59, respectively.
The modeling approach for the MLP model in this study had the following steps: (1) The indicator data and quality score data were normalized. (2) The dataset was divided into a training set (70% of the data) and a testing set (remaining 30% of the data). (3) The MLP model was constructed using the gradient descent algorithm to derive the optimal weight values for the model. The experimental results are shown in Figure 5 and Figure 6.
The Figure 5 shows the results of the real value and predicted value of the test set. Among them, the blue is the predicted value, and the red is the real value. It is evident from the figure that a discernible variance exists between the predicted and actual values. The magnitude of this difference is relatively small, however, which suggests a high level of accuracy in the prediction model.
In real-world applications, each sample represents a distinct customer table. Table 4, Table 5, Table 6 and Table 7 demonstrate the evaluation of the consignment personal customer information table and the general personal customer information table from the trust institution using two different methods: the conventional scoring technique and a machine learning-based method. The time invested and the quality assessment scores obtained from the AHP are approximate representations of the outcomes achieved by the manual scoring method currently employed by the trust.
According to the incremental experimental results, we observed that the time required by both methods increased as the sample size grew. When the number of customer tables doubled, the time spent on the traditional method also doubled, whereas the time expenditure for the machine learning-based approach experienced only a slight increase. Moreover, the quality assessment scores generated through the MLP method were relatively higher than the traditional quality evaluation techniques. By employing Powershap for feature selection, we excluded redundant data, thus reducing the time expenditure and improving score precision (Table 13).

4.3. Feature Selection Based on Shapley Value

We initially split the data into a 3:7 ratio for testing and training purposes. Subsequently, we categorized the data into six classes, representing the six primary indicators. Then, we constructed a Powershap function, utilizing the chosen classification model, which, in this study, was the logistic regression CV algorithm.
When employing the logistic regression CV algorithm model for 200 iterations, we observed that significant features accounted for five-sixths of the total (Figure 7). By considering impact scores and other statistical results, the timeliness factor could be excluded from the primary indicators. This exclusion not only reduced the required effort, but also improved the precision in determining the weight values for subsequent analysis.

5. Conclusions

This article organizes and analyzes the existing quality evaluation indicators, and comprehensively uses completeness, accuracy, timeliness, consistency, uniqueness and compliance to establish a quality evaluation indicator system for trusts and other financial institutions. Preliminary weight determination is made based on AHP. Since there are many indicators, the Shapley value is introduced for feature selection. A multiple linear regression model is used to determine new weights for the screened indicators. The model is verified using customer data and institutional data from a trust company’s customer information integration system.
Through data quality scoring, trust data can be graded for quality, thereby improving data quality and promoting the transfer and transformation of data results from trusts and other financial institutions. Using machine learning regression algorithms to evaluate data quality has the advantages of quantifiable evaluation results, more intuitive results, and higher accuracy. In addition, in the face of massive customer data, the intelligent data quality evaluation model based on machine learning is faster and more convenient, and the evaluation results are more accurate and objective.
Therefore, the data quality evaluation model based on the regression algorithm model can not only be used to evaluate the quality of trust customer data but can also be used in other industries and government departments to evaluate the quality of other types of data, including their own, as well as the benefits and value that the data can bring.

Author Contributions

Conceptualization, M.L. and Y.Y.; methodology, M.L.; software, M.L.; validation, M.L., J.L. and Y.Y.; formal analysis, M.L.; investigation, J.L.; resources, Y.Y.; data curation, M.L.; writing—original draft preparation, M.L.; writing—review and editing, J.L.; visualization, M.L.; supervision, J.L.; project administration, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Weber, K.; Otto, B.; Osterle, H. One Size Does Not Fit All—A Contingency Approach to Data Governance. J. Data Inf. Qual. 2009, 1, 4. [Google Scholar] [CrossRef]
  2. Begg, C.; Cairat, T. Exploring the SME quandary:data governance in practise in the small to medium-sized enterprise sector. Electron. J. Inf. Syst. Eval. 2012, 15, 3–13. [Google Scholar]
  3. Newman, D.; Logan, D. Governance is an essential building block for enterprise information management. Gart. Res. Stamford 2006, 13, 4. [Google Scholar]
  4. Niemi, E. Designing a data governance framework. In Proceedings of the IRIS Conference, Turku, Finland, 16–19 August 2011. [Google Scholar]
  5. Karkošková, S. Data governance model to enhance data quality in financial institutions. Inf. Syst. Manag. 2023, 40, 90–110. [Google Scholar] [CrossRef]
  6. Mcgilvray, D. Executing Data Quality Projects Ten Steps to Quality Data and Trusted Information; Elsevier: Amsterdam, The Netherlands, 2008; Volume 12, pp. 2–4. [Google Scholar]
  7. Omara, E.; Said, T.E.; Mousa, M. Employing neural networks for assessment of data quality with emphasis on data completeness. Int. J. Artif. Intell. Mach. Learn. 2011, 11, 21. [Google Scholar]
  8. Peltier, J.W.; Zahay, D.; Lehmann, D.R. Organizational learning and CRM success: A model for linking organizational practices, customer data quality, and performance. J. Interact. Mark. 2013, 27, 1–13. [Google Scholar] [CrossRef]
  9. Taleb, I.; Kassabi, H.; Serhani, M.A.; Dssouli, R.; Bouhaddioui, C. Big Data Quality: A Quality Dimensions Evaluation. Ubiquitous Intelligence and Computing. In Proceedings of the Advanced and Trusted Computing, Scalable Computing and Communications, Cloud and Big Data Computing, Internet of People, and Smart World Congress, Toulouse, France, 18–21 July 2016. [Google Scholar]
  10. Juddoo, S. Overview of data quality challenges in the context of Big Data. In Proceedings of the 2015 International Conference on Computing, Communication and Security (ICCCS), Pointe aux Piments, Mauritius, 4–5 December 2015; pp. 1–9. [Google Scholar]
  11. Madhikermi, M.; Kubler, S.; Robert, J.; Buda, A.; Främling, K. Data quality assessment of maintenance reporting procedures. Expert Syst. Appl. 2016, 63, 145–164. [Google Scholar] [CrossRef]
  12. Mashoufi, M.; Ayatollahi, H.; Khorasani-Zavareh, D.; Talebi Azad Boni, T. Data quality in health care: Main concepts and assessment methodologies. Methods Inf. Med. 2023, 62, 5–18. [Google Scholar] [CrossRef]
  13. Uzoka, F.M. AHP-based system for strategic evaluation of financial information. Inf. Knowl. Syst. Manag. 2005, 5, 49–61. [Google Scholar]
  14. Khan, A.W.; Khan, M.U.; Khan, J.A.; Ahmad, A.; Khan, K.; Zamir, M.; Kim, W.; Ijaz, M.F. Analyzing and evaluating critical challenges and practices for software vendor organizations to secure big data on cloud computing: An AHP-based systematic approach. IEEE Access 2021, 9, 107309–107332. [Google Scholar] [CrossRef]
  15. Alam, M.K. A systematic qualitative case study: Questions, data collection, NVivo analysis and saturation. Qual. Res. Organ. Manag. Int. J. 2021, 16, 1–31. [Google Scholar] [CrossRef]
  16. Gomes, V.C.F.; Queiroz, G.R.; Ferreira, K.R. An overview of platforms for big earth observation data management and analysis. Remote Sens. 2020, 12, 1253. [Google Scholar] [CrossRef]
  17. Liu, Y.; Wang, Y.; Xue, F.; Chen, J. A hybrid approach for supplier selection based on quality management system evaluation and grey relational analysis. J. Intell. Fuzzy Syst. 2021, 41, 1149–1159. [Google Scholar]
  18. Malik, S.; Tahir, M.; Sardaraz, M.; Alourani, A. A resource utilization prediction model for cloud data centers using evolutionary algorithms and machine learning techniques. Appl. Sci. 2022, 12, 2160. [Google Scholar] [CrossRef]
  19. Titus, B.D.; Brown, K.; Helmisaari, H.S.; Vanguelova, E.; Stupak, I.; Evans, A.; Clarke, N.; Guidi, C.; Bruckman, V.J.; Varnagiryte-Kabasinskiene, I.; et al. Sustainable forest biomass: A review of current residue harvesting guidelines. Energy Sustain. Soc. 2021, 11, 10. [Google Scholar] [CrossRef]
  20. Hou, Y.; Biljecki, F. A comprehensive framework for evaluating the quality of street view imagery. Int. J. Appl. Earth Obs. Geoinf. 2022, 115, 103094. [Google Scholar] [CrossRef]
  21. Sun, H.; Chen, Z. Interval neutrosophic hesitant fuzzy AHP method based on combined weights. J. Intell. Fuzzy Syst. 2021, 41, 8015–8028. [Google Scholar]
  22. Wang, Y.; Liu, Z.; Cai, C.; Xue, L.; Ma, Y.; Shen, H.; Chen, X.; Liu, L. Research on the optimization method of integrated energy system operation with multi-subject game. Energy 2022, 21, 123305. [Google Scholar] [CrossRef]
  23. Liu, J.; Huang, J.; Zhou, Y.; Li, X.; Ji, S.; Xiong, H.; Dou, D. From distributed machine learning to federated learning: A survey. Knowl. Inf. Syst. 2022, 64, 885–917. [Google Scholar] [CrossRef]
  24. Chen, H.; Covert, I.C.; Lundberg, S.M.; Lee, S.I. Algorithms to estimate Shapley value feature attributions. Nat. Mach. Intell. 2023, 5, 590–601. [Google Scholar] [CrossRef]
  25. Liu, L.; Li, Z.; Yang, J. An emergency plan evaluation model based on combined DEA and TOPSIS methods. J. Clean. Prod. 2021, 315, 62. [Google Scholar]
  26. Kitiyodom, P.; Chindapa, P. Development of an emergency response plan assessment model for hazardous chemical accidents in Thailand. J. Loss Prev. Process. Ind. 2021, 70, 307. [Google Scholar]
  27. Wen, C.; Yang, J.; Gan, L.; Pan, Y. Big data driven Internet of Things for credit evaluation and early warning in finance. Future Gener. Comput. Syst. 2021, 34, 295–307. [Google Scholar] [CrossRef]
  28. Liapis, C.M.; Kotsiantis, S. Energy Balance Forecasting: An Extensive Multivariate Regression Models Comparison. In Proceedings of the 12th Hellenic Conference on Artificial Intelligence, Corfu, Greece, 7–9 September 2022; pp. 1–7. [Google Scholar]
  29. Tiwari, P. Bank affection and customer retention: An empirical investigation of customer trust, satisfaction, loyalty. SN Bus. Econ. 2022, 2, 54. [Google Scholar] [CrossRef]
Figure 1. Quality evaluation model based on multiple linear regression.
Figure 1. Quality evaluation model based on multiple linear regression.
Futureinternet 15 00338 g001
Figure 2. Dimensions of quality evaluation system.
Figure 2. Dimensions of quality evaluation system.
Futureinternet 15 00338 g002
Figure 3. Integrity statistics results.
Figure 3. Integrity statistics results.
Futureinternet 15 00338 g003
Figure 4. Heat map of correlation between variables.
Figure 4. Heat map of correlation between variables.
Futureinternet 15 00338 g004
Figure 5. True value and predicted value of quality evaluation score.
Figure 5. True value and predicted value of quality evaluation score.
Futureinternet 15 00338 g005
Figure 6. Test set loss function graph.
Figure 6. Test set loss function graph.
Futureinternet 15 00338 g006
Figure 7. The importance of each feature.
Figure 7. The importance of each feature.
Futureinternet 15 00338 g007
Table 1. Problem situations and examples of some data.
Table 1. Problem situations and examples of some data.
Problem SituationDescription and Examples
field emptyEmpty values appear in required fields such as gender, nationality, address, contact information, and certificate validity period
irregular formatThe 18-digit format that does not meet the standard, such as ID number, etc.
consistency errorFormats that do not meet the standard, such as the ID number does not match the gender or the date of birth, etc.
timeliness errorThe validity period of the certificate has “valid until less than the current date”, etc.
data record duplicationThe same data are repeatedly recorded, and there is a situation of “multiple compilations for one household”
Table 2. Some types of data accuracy and related descriptions.
Table 2. Some types of data accuracy and related descriptions.
Question TypeProblem Statement
FalseThe data are not true, such as the ID number being non-existent
inconsistent informationThe related information does not correspond, such as the ID number not corresponding to the name, or there being a situation where one ID number corresponds to multiple names
timing relationship errorThe relative timing between data is not correct, such as the creation time of individual customers being later than the modification time
invalid dataData records should not be valid relative to a certain period of time or a certain point in time; for example, the establishment time of the enterprise is later than the current time
Table 3. Example of duplicate data records.
Table 3. Example of duplicate data records.
Data RecordNO.Customer NumberCustomer NameGender
11000103445112803He Zhixiong2
21000103445112803He Zhixiong2
31000103445112803He Zhixiong2
41000083575112803Zhu Hongyan1
Table 4. Scoring rules and scoring formulas for data quality evaluation indicators.
Table 4. Scoring rules and scoring formulas for data quality evaluation indicators.
Indicator NumberIndicator NameScoring Formula and Instructions
B 1 integrity X B 1 = λ C 11 X C 11 + λ C 12 X C 12 + λ C 13 X C 13
C 11 record integrity X C 11 = 100 × ( 1 N C 11 / W C 11 )
N C 11 : the number of missing data items in the data table
W C 11 : the total number of data items that should be recorded in the data table
C 12 attribute integrity X C 12 = 100 × ( 1 N C 12 / W C 12 )
N C 12 : the number of missing attributes in the data table
W C 12 : the total number of attributes that should be recorded in the data table
C 13 value integrity X C 13 = 100 × ( 1 N C 13 / W C 13 )
N C 13 : missing required fields in the data table
W C 13 : the total number of data items that should be required in the data table
B 2 accuracy X B 2 = λ C 21 X C 21 + λ C 22 X C 22 + λ C 23 X C 23
C 21 record accuracy X C 21 = 100 × ( 1 N C 21 / W C 21 )
N C 21 : the number of inaccurate data items recorded in the data table
W C 21 : the total number of assigned data items in the data table
C 22 value accuracy X C 22 = 100 × ( 1 N C 22 / W C 22 )
N C 22 : number of data items in the data table with inaccurate values
W C 22 : the total number of data items that should be recorded in the data table
C 23 conceptual accuracy X C 23 = 100 × ( 1 N C 23 / W C 23 )
N C 23 : the number of conceptually inaccurate data items in the data table
W C 23 : the total number of data items that should be recorded in the data table
B 3 timeliness X B 3 = X C 31
C 31 data update timeliness X C 31 = 100 × ( 1 N C 31 / W C 31 )
N C 31 : the number of data items that are not updated in time in the data table
W C 31 : the total number of data items that should be recorded in the data table
B 4 consistency X B 4 = X C 41
C 41 logical consistency X C 41 = 100 × ( 1 N C 41 / W C 41 )
N C 41 : the number of data items in the data table that do not satisfy the logic
W C 41 : the total number of data items that should be recorded in the data table
B 5 uniqueness X B 5 = X C 51
C 51 record uniqueness X C 51 = 100 × ( 1 N C 51 / W C 51 )
N C 51 : the number of data items repeated in the data table
W C 51 : the total number of assigned data items in the data table
B 6 compliance X B 6 = X C 61
C 61 data format compliance X C 61 = 100 × ( 1 N C 61 / W C 61 )
N C 61 : number of data items with non-compliant format in the data table
W C 61 : the total number of assigned data items in the data table
Table 5. Average random consistency (RI).
Table 5. Average random consistency (RI).
Size of Matrix 12345678910
Random consistency000.580.91.121.241.321.411.451.49
Table 6. Quality evaluation system evaluation matrix A 1 .
Table 6. Quality evaluation system evaluation matrix A 1 .
B 1 Integrity B 2 Accuracy B 3 Timeliness B 4 Consistency B 5 Uniqueness B 6 Compliance w i
B 1 integrity121/23320.229
B 2 accuracy1/211/32210.133
B 3 timeliness2314430.364
B 4 consistency1/31/21/41110.084
B 5 uniqueness1/31/21/41110.084
B 6 compliance1/211/31110.106
λ m a x = 6.068 , C R = 0.011 < 0.1
Table 7. Integrity assessment matrix A 2 .
Table 7. Integrity assessment matrix A 2 .
C 11 Record Integrity C 12 Attribute Integrity C 13 Value Integrity w i
C 11 record integrity1550.714
C 12 attribute integrity1/5110.143
C 13 value integrity1/5110.143
λ m a x = 3.000 , C R = 0.000 < 0.1
Table 8. Accuracy evaluation matrix A 3 .
Table 8. Accuracy evaluation matrix A 3 .
C 21 Logical Accuracy C 22 Value Accuracy C 23 Conceptual Accuracy w i
C 21 logical accuracy121/30.238
C 22 value accuracy1/211/40.137
C 23 conceptual accuracy3410.625
λ m a x = 3.018 , C R = 0.000 < 0.1
Table 9. Weights of quality bid evaluation indicators.
Table 9. Weights of quality bid evaluation indicators.
Level 1 IndicatorsLevel 2 Indicators
indicator numberindicator nameweights ( μ m )indicator numberindicator nameweights ( μ n )
B1integrity0.229C11record integrity0.714
C12attribute integrity0.143
C13value integrity0.143
B2accuracy0.133C21logical accuracy0.238
C22value accuracy0.137
C23conceptual accuracy0.625
B3timeliness0.364C31data update timeliness1
B4consistency0.084C41logical consistency1
B5uniqueness0.084C51record uniqueness1
B6compliance0.106C61format compliance1
Table 10. Individual customer completeness statistical results.
Table 10. Individual customer completeness statistical results.
Field Chinese NameField English NameIs EmptyNull RatioField Chinese NameField English NameIs EmptyNull Ratio
nameCST_NM00type of certificateCRDT_TP00
ID numberCRDT_NO00genderGND422528.72%
nationalityNAT514534.98%professionOCP11,54978.52%
addressPHY_ADR_INF285919.44%contact informationCTC_PER385026.17%
certificate validity periodCRDT_EFDT515635.05%customer numberCST_ID00
contact typeCTC_PER_TP14,7091system sourceSYS_SRC00
address typePHY_ADR_TP285819.43%sales channelsSALE_SRC00
updaterUPT_NM874459.45%update timeUPT_TM874459.45%
founderCRT_NM00creation timeCRT_TM00
Table 11. Descriptive statistics of indicators in the personal customer table.
Table 11. Descriptive statistics of indicators in the personal customer table.
IndexMaximum ValueMinimum ValueAverageStandard Deviation
integrity87.5974.480.90.25
accuracy98.380.2691.560.38
timeliness1001001000
consistency9.0901.150.23
uniqueness10052.6199.941.26
Compliance74.126.6767.531.71
Table 12. Descriptive statistical table of indicators in the institutional client table.
Table 12. Descriptive statistical table of indicators in the institutional client table.
IndexMaximum ValueMinimum ValueAverageStandard Deviation
integrity84.7578.8180.120.3
accuracy97.8382.192.330.28
timeliness1001001000
consistency9.31.53.850.29
uniqueness10089.6698.10.86
Compliance50.6611.543.212.03
Table 13. The results for samples of different orders of magnitude were calculated based on the AHP and the method proposed in this paper.
Table 13. The results for samples of different orders of magnitude were calculated based on the AHP and the method proposed in this paper.
Table NameNumber of SamplesAHP Time SpentQuality Review ScoreMachine Learning Time SpentQuality Review Score
MO_CI_PER_CONSIGNMENT10,000600.02 s81.39613.89 s83.55
20,0001192.28 s83.21650.25 s85.89
30,0001532.43 s80.91690.11 s83.22
MO_CI_PER_CUSTOMER10,000576.22 s84.43623.14 s85.99
20,0001312.34 s79.98652.25 s82.01
30,0001701.36 s80.12661.21 s80.56
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, M.; Liu, J.; Yang, Y. Financial Data Quality Evaluation Method Based on Multiple Linear Regression. Future Internet 2023, 15, 338. https://doi.org/10.3390/fi15100338

AMA Style

Li M, Liu J, Yang Y. Financial Data Quality Evaluation Method Based on Multiple Linear Regression. Future Internet. 2023; 15(10):338. https://doi.org/10.3390/fi15100338

Chicago/Turabian Style

Li, Meng, Jiqiang Liu, and Yeping Yang. 2023. "Financial Data Quality Evaluation Method Based on Multiple Linear Regression" Future Internet 15, no. 10: 338. https://doi.org/10.3390/fi15100338

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop