1. Introduction
The concept of supply chain (SC) has arisen with the comprehension that the achievement of companies depends on the interplay between raw materials, orders, information flow, workforce, money, and existing equipment and machinery [
1], and the transition from individual organization-based management to an integrated management approach being actualized [
2]. In today’s world, competition takes place between SCs [
3,
4] and the effective supply chain management (SCM), which comprises whole activities relevant to the flow and conversion of goods from the raw material phase to the final user [
5], makes many remarkable contributions to SCs, especially enabling a sustainable competitive advantage [
6] and creating value for SC stakeholders [
7], and has a critical significance in the tremendously competitive global business world. Effective SCM enables strategic and operational benefits to countries, regions, and undoubtedly companies [
8,
9]. In the light of these developments, its analysis and improvement have become vitally significant and inevitable [
10,
11].
The foundation of analysis and improvement of the SC is based on performance measurement [
12].Performance measurement is described as “
the process of quantifying effectiveness and efficiency of actions” [
13]. Supply chain performance measurement (SCPM) is a set of measures used for measuring the efficiency and effectiveness of relationships and processes of SC, comprising multiple organizational functions and companies, and providing the regulation of the SC. It is an activity that measures the efficiency of SC processes, enables information about the relationships in the SC, contributes to the effective usage of resources [
14,
15], enables information about the success and health of the SC by identifying potential problems and improvement fields [
16], and facilitates decision-making by ensuring information to senior management and operations managers. It also enables effective planning and control in the SC [
17] and contributes to operational excellence [
18]. Moreover, the customer identifies whether the needs are met, and ensures the identification of wasteful issues and bottlenecks. It offers a more transparent cooperation and communication and provides a better comprehension of processes in the SC [
19]. Thanks to these contributions, it is an integral part of SCM [
20].
The main aim of SCPM is to comprehend how the SC system works and then try to determine occasions for supply chain performance (SCP) improvement. It is the foundation for improving SCP [
21,
22]. Although it is a critical issue for improving SCP and enhancing productivity and operational efficiency for SC stakeholders [
23], the extant studies have highlighted that many companies fail to tap into the potential of the SC, that is, to maximize SCP [
24,
25]. There are two main aspects to consider in SCPM. They are the determination of appropriate SCP measures and an appropriate SCP measurement system [
26,
27]. SCM is an integrative philosophy that manages the whole flow in the SC [
28]. SCP measures should be comprehensive, measurable, not far from the real world, and universal, and consider the SC as a whole [
29,
30]. Moreover, they should be consistent with organizational goals and allow comparison under different conditions [
31]. The growing complexity and magnitude of the SC often has the effect of complicating coordination and undermining performance in the SC. For this reason, it is vital to use a systematic approach to analyzing performance [
32] and to determine performance measures accurately [
10].
The Supply Chain Operations Reference (SCOR) model is the well-known systemic and balanced performance measurement system among the prevailing SCPM systems. It was introduced by the Supply Chain Council (SCC) in 1996 and is defined as “
systematic approach for identifying, evaluating and monitoring supply chain performance” [
33]. It is accepted as the basis of performance measurement [
34,
35], and has been adopted by many companies since its launch. Additionally, this model is critical in enabling universally accepted standard performance metrics because it is difficult for senior management to agree on SCP metrics, unlike SC strategies [
36]. Consequently, this model ensures a common language for implementing, organizing, and deciding the procedures of SC [
37,
38]. Its implementation contributes to quicker assessment and improvement of performance in SC, identification of performance gaps clearly, analysis of the competitive basis, implementation of processes and systems, and structuring of the SC [
39,
40]. It is also critically important as it assists managers in making strategic decisions [
41].
There are some difficulties encountered in practice, in contrast to contributions of the SCOR model [
42]. When investigating company-level challenges, one of the challenges is the need for automated data collection in the company. Likewise, other companies in the industry must have automated data collection to be able to compare performance [
43]. On the other hand, the researchers in this field cannot investigate the performance of companies, and cannot make inter-sectoral comparisons by this model’s measures. While investigating SCM operations and their interrelationships, they face some difficulties, such as time constraints, access to administrators, and data privacy. This situation is a limitation for future research in this field [
44,
45]. Measurement scales have been utilized as a data collection tool in the literature to measure SCP for reasons such as time constraints, easier access to data, and faster measurement and comparison. Statistical analysis is used for data analysis [
46].
SCs have looked for ways to overcome the insurmountable difficulties associated with SCs and carry out the highest possible potential of them by implementing innovative notions, policies, and strategies today [
47,
48]. The theoretical aim of this study is to develop and validate an SCP scale based on the standardized SC performance structure and measures of the SCOR model, and which deals with SCP more comprehensively than the extant scales in the literature. It also aims to adapt the SCP structure to the requirements of the era by performing the measurement on the usage levels of disruptive technologies (DTs) and to determine the contributions of the technologies. The practical aim of this study is to prevent the calculation of performance measures of the SCOR model separately, to enable a faster and more practical performance measurement. The performance attributes of the SCOR model version 13.0 (v13.0) [
49,
50] were defined as factor structures (or sub-dimensions) in this scale study, and the performance measures of SCOR v13.0 were expanded by an exhaustive literature review, expert group interviews, and effects of disruptive technologies within the scope of the study. Consequently, the effect of DTs was considered both when generating the item pool and taking measurements.
The remainder of this paper is designated as follows:
Section 2 illustrates the theoretical background of this study with its widely utilized performance measurement systems and scales developed in the SCP field.
Section 3 describes the research methodology and quantitative findings on which the study is based. This scale was developed and validated as “
Supply Chain Performance Scale as SCOR-Based (SCPSS)”.
Section 4 discusses the findings, illustrates the theoretical contributions, managerial implications, limitations of the study, and suggestions for further research.
Section 5 outlines the most significant insights of this study.
3. Materials and Methods
This study designed a seven-stage model to measure, validate, and constitute the structure and predictability of SCP. Stage 1 focused on item generation and purification. In Stage 2, a draft questionnaire was made up, the pilot study was performed, and the item pool was finalized for the main study. In Stage 3, exploratory factor analysis (EFA) was executed by compressing five factors to perform SCOR v13.0-based dimensionality, and the initial determination of dimensionality was performed. Confirmatory factor analysis (CFA) was carried out performing confirmation of dimensionality in Stage 4. While convergent and discriminant validity appraisal was executed in Stage 5, nomological validity was evaluated in Stage 6. Bias effect was investigated in Stage 7. The macro view of the research process is illustrated in
Figure 1 below:
In this study, data were collected face-to-face and by e-mail, after necessary ethics committee permissions were acquired. The sample sizes of the study for the pilot study and main study comprise 227 companies and 452 companies, respectively. The industrial sectors included six groups: retail and FMCG, transportation, distribution and warehousing, e-trade, service, manufacturing, and import-export—so the study aimed at a large population size. Accordingly, the sample size for the main study is more than the 384 sample size for the 1,000,000-population size recommended by Sekaran and Bougie (2016) [
96]. The measurement is carried out according to the difference created by the use of DTs in a year. SPSS V.21 for EFA, LISREL 8.51 for CFA, nomological validity, and investigation of bias effect. Office 365 Excel for convergent and discriminant validity were used. The questionnaire was constituted using a five-point Likert scale ranging from 1 (I totally disagree) to 5 (I totally agree).
3.1. Stage 1: Item Generation and Purification
Item generation was performed by considering the Level-1 and Level-2 performance metrics of the SCOR model v13.0, the current academic literature, and the effects of DTs on SCP. The most current version of the SCOR model was version 13.0 (v13.0) at the time of this study. For this reason, this version was considered in the scope of this study. A list of expert groups to be interviewed was identified to discuss the clarity of the items and whether there were any missing or added SCP measures. Accordingly, a total of nine experts were identified, five of which are industry professionals working in this field, and four of them are previous academicians from Turkey who maintain their academic studies in this field. Industry professionals were selected from companies’ SCM and digital transformation departments. Academicians who are experts in the fields of SCM and management information systems have continued to work as faculty members in these fields. The interviewed expert group list is included in
Appendix A of this study. The interviews were conducted face-to-face or online through the Zoom platform. In this way, the content validity of the study was realized with expert group interviews.
The SCOR model enables standard and up-to-date performance metrics and defines the SC structure from all aspects. The sub-dimensions of the scale are based on SCOR v13.0 performance attributes, and scale items are based on SCOR v13.0 performance metrics, the extant literature, and the effects of DTs on SCP, as previously stated. This version includes five performance attributes: reliability, responsiveness, agility, cost, and asset management [
47,
48]. First, performance metrics are generated. Afterwards, these metrics are grouped according to these attributes, and the item pool for the pilot study is constituted. Level-1 and Level-2 measures of SCOR model v13.0 are given in
Appendix A of this study, and the generated items are given in
Appendix B of this study. In the next step of this study, the findings of the pilot study performed before reaching the findings of the main study are given.
3.2. Stage 2: Exploratory Factor Analysis for Pilot Study
Pilot studies allow for the research subject to be addressed, to obtain information about the feasibility of the study, and to determine a roadmap before launching the main study [
97]. There are different values in the literature for the number of samples required for the pilot study. Treece and Treece (1982) emphasized that a sample corresponding to 10% of the required sample in the main study is adequate [
98]. In this study, a pilot study was conducted with the data of 227 different companies, which exceeds approximately 10% of the targeted main sample. The findings of the reliability analysis of the alternative models for the pilot study and the findings of the sub-scales and factor analysis reached by the EFA are illustrated in
Table 2 and
Table 3, respectively:
The Cronbach’s alpha value allows testing the internal consistency of the scale and its sub-scales. The minimum value of it should be at least 0.70 for existing scales and at least 0.60 for newly developed scales [
99,
100]. These values are at a high level for the pilot study, as seen
Table 2. Before evaluating the findings of the EFA for the pilot study, KMO and Bartlett’s Test of Sphericity values are investigated, as is whether the data are convenient for factor analysis. It is expected that the KMO value is above 0.60 and the Bartlett’s Test value is statistically significant (
p < 0.05) [
101,
102]. The data are appropriate for factor analysis as they enable relevant threshold values for the pilot study.
EFA is a type of factor analysis that allows the correlations between observed variables to be explained by fewer latent variables or factors [
103]. It ensures the initial definition of the sub-dimensions of the structure, or in other words, the factors [
104,
105]. The type of rotation and extraction are determined by the researcher while performing the analysis. Oblique rotation is used if sub-dimensions are highly correlated to each other [
106,
107]. Structures are usually highly correlated to each other in social sciences, so this rotation is commonly used [
108]. In this study, the correlations between the items were investigated to understand the relationships between the structures, and since most of the correlations were above 0.30, direct oblimin, a type of oblique rotation, was utilized. Principal axis factoring, which is an effective method to reveal weak factors, was used as an extraction method [
109].
The eigenvalue indicates the amount of information obtained from a factor. According to the eigenvalue rule, factors with an eigenvalue lower than 1.0 should not be considered [
110]. The scree plot is based on eigenvalues and supports deciding the correct number of factors. The change in eigenvalue for successive factors is taken into consideration, while investigating this graph [
111].
Figure 2 gives information about the eigenvalues for the factor structures. This value is 27.00 for a single factor structure and 3.00 for a three-factor structure approximately. Since the study is based on SCOR v13.0, it was compressed into five factors due to its theoretical background. The eigenvalue is 2.50 for a five-factor structure approximately. The structures compressed into the five factors given in
Table 2 reflect 58.436%, 60.732%, and 61.137% of the total variances, respectively. Since a flattening is seen in the graph after the fifth factor in the scree plot, it is possible to say that the data point to the five-factor structure.
The factor loadings of 0.50 or higher are considered as practically significant [
112]. Standardized loadings should have a lower limit of at least 0.50 or 0.60 for newly developed scales [
113]. As a finding of EFA, six items that encountered cross loading problems were excluded from the item pool and the structure in
Table 4 was obtained. The loadings for the five-factor model consisting of sixty-eight items altered between 0.324 and 0.843. Subsequently, the items with the loadings below 0.50 were removed from the item pool and the structure given in
Table 5 was obtained. The loadings for the five-factor model consisting of fifty-seven items altered between 0.468 and 0.843. The same process was performed to this structure, and the structure presented in
Table 6 was obtained by removing the items with the loadings below the threshold value. The loadings for the five-factor model consisting of fifty-five items altered between 0.505 and 0.841.
3.3. Stage 3: Exploratory Factor Analysis for Main Study
The demographic characteristics of the 452 samples collected for the main study were investigated, before the EFA for the main study was performed. While creating these questions, the opportunity to leave them blank was enabled due to the participants’ concerns about data privacy. Demographic characteristics of the participants are illustrated in
Table 7. The unanswered questions are indicated in the table as no response.
A preliminary version of the study was generated, and the alternative factor structures were examined in the pilot study [
114]. Items with cross loading problems and factor loading values below 0.50 [
113] were removed and the relevant items were renumbered. Accordingly, the item pool was updated, and the items were renumbered. The item pool for the main study is presented in
Table 8 below:
All alternative models obtained for the main study are structures obtained by compressing the structure to five factors on the theoretical grounds that the scale is based on SCOR v13.0. When the conceptual structures that make up the sub-dimensions (or factors) of the scale are investigated, reliability (RL) is the ability to carry out tasks as expected and focuses on the predictability of the outcome of a process. Responsiveness (RV) is defined as the velocity at which tasks are accomplished and the velocity at which an SC delivers product to a customer. Agility (AG) is the ability to respond to external influences and market changes. Cost (C) is the performance dimension that consists of the costs associated with operating SC processes. Asset management (ASM) refers to the ability to use assets efficiently [
49].
When the EFA was performed for the item pool presented in
Table 6, the RV12, RV13, C1, and C2 items were removed from the analysis due to the cross-loading problem, and Model 1 was obtained. Afterwards, a total of six items with factor loading values below 0.50 (RL1, RL12, AS3, AS1, RV1, RV11) were excluded from this structure, and Model 2 was obtained. One item with a loading value lower than 0.50 (RV2) was encountered in Model 2, and the analysis was repeated by excluding them from the model, and Model 3 was attained. The findings of the reliability analysis of the structures and sub-scales and factor analysis are given in
Table 9 and
Table 10, and the factor loadings of them are presented in
Table 11,
Table 12 and
Table 13:
The Cronbach’s alpha value is above the threshold value (
p > 0.60) and the sub-scales have a high internal consistency (
Table 9). Before evaluating the findings of the EFA for the main study, KMO and Bartlett’s Test of Sphericity values are investigated, as is whether the data are convenient for factor analysis. It is expected that the KMO value is above 0.60 and the Bartlett’s Test value is statistically significant (
p < 0.05) [
101,
102]. The data is appropriate for factor analysis as it enables relevant threshold values for the main study.
The three structures, compressed into five factors, explain 58.538%, 61.084%, and 61.487% of the total variance, respectively (
Table 10). According to the scree plot chart for the main study illustrated in
Figure 3 below, the eigenvalue for the single factor structure is approximately 19.00. The eigenvalue is approximately 3.00 for the three-factor structure, while it is approximately 2.00 for the five-factor structure. Since there is a flattening in the graph after the fifth factor in the scree plot, it is possible to say that the data point to the five-factor structure and support the theoretical background of the study.
3.4. Stage 4: Confirmatory Factor Analysis
CFA enables confirmation of the construct revealed by the EFA or provides confirmation of previously determined structures on theoretical grounds [
155,
156]. It is a statistical confirmation [
157]. Since this scale is based on the SCOR v13.0 model, its sub-dimensions and items belonging to sub-dimensions are certain. Nevertheless, the EFA was carried out initially and then this structure was confirmed to carry out a more rigorous study and to investigate the compatibility of the data with the theoretical structure.
After the alternative structures reached in the EFA were verified, the model obtained by removing the items with factor loading below 0.70 was confirmed to avoid problems in convergent validity, which will be evaluated in the next stage. As in the other models, compression was applied to five factors and this structure was named Alternative Model 4. The CFA offers several suggestions to improve the model’s fit indices values. One of them is to correlate items belonging to the same factor that demonstrate high correlation with each other in the model. It must be based on theoretical grounds for the suggestion to be performed. Accordingly, the following items are correlated in alternative models. The related items and the chi-square reductions that this correlation will enable are given in
Table 14 below:
Cost of goods sold (COGS) is the total cost related to purchasing raw materials and producing goods, and includes both direct costs (e.g., materials, labor) and indirect costs [
158,
159]. Distribution of goods or services among company activities is significant [
160]. Distribution costs include the costs related with receiving, storing, and transporting goods [
161]. C7 and C8 were correlated in the indicated models, as the distribution of raw materials and materials to warehouses or factories for production has a direct effect on the COGS. The negotiation process with suppliers is part of the purchasing function [
162]. The cycle time for selecting and negotiating suppliers affects the purchasing cycle time. RV1 and RV2 were correlated with this theoretical reason. Since the purchasing materials cost will affect the unit production cost, C3 and C5 were correlated in the relevant models [
163].
As delivery accuracy enhances, delivery performance improves [
164]. Therefore, RL8 and RL9 were correlated. Working capital can be defined as the ability to pay due debts. Moreover, cash and working capital are used synonymously in some sources [
165]. The cash conversion cycle (CCC) is a measure of working capital and demonstrates how well it is managed [
166]. Accordingly, AS8 and AS9, which were expected to be highly correlated, were correlated in Model 4. Storage cost is also a factor that affects the unit production cost [
167]. Hence, C4 and C5 were correlated in the related model.
Model fit indices for these four structures are given in
Table 15 below. χ2/df, RMSEA, and SRMR indices are at acceptable levels for all models. Alternative Model 2-3-4 also has an acceptable level for CFI. Alternative Model 4 is also acceptable in terms of NFI and AGFI indices. Four alternative models are different from each other in terms of the scope of item, so the chi-square difference test was not carried out. Therefore, statistically, the optimal model is Alternative Model 4, which has better model fit indices values according to the other alternatives.
The t-values of the loading are investigated in order to control whether the factor loadings are statistically significant. The threshold for t-values is 1.96 at the 0.05 significance level [
169,
170]. Factor loading, t-values, and R
2 for the items are given in
Table 16 below. As comprehended in the table, whole t-values of the loadings are statistically significant.
3.5. Stage 5: Convergent and Discriminant Validity
Convergent validity is a type of validity that proves that the observed variables represent the same structure, measure a single conceptual structure, and are related to each other [
171]. Average variance extracted (AVE), composite reliability (CR), and factor loading criteria are considered for this validity [
172]. The AVE value must be higher than or equal to 0.50 [
173]. The threshold value for CR is 0.60 [
174]. The closer it is to 1.00, the better it is considered [
100]. Factor loadings indicate at what level the items are related to the factor, which is called the latent structure [
105,
170]. The loading values are expected to be higher than or equal to 0.708 to ensure convergent validity, and this also achieves indicator reliability [
172]. Lastly, the Cronbach’s Alpha values of relevant factors are investigated and if they are higher the threshold value of 0.70, internal consistency reliability has been ensured [
172,
175,
176]. The scale has convergent validity as it has fulfilled whole all necessary criteria. The factor loading values are given in
Table 16 above. The other values are given in
Table 17 below:
Discriminant validity is a type of validity that is accepted as a prerequisite for analyzing the relationships between latent variables and is calculated separately for each latent variable or for each factor [
112,
177]. The factors should be related to each other to ensure this validity, but the level of this relationship should be low enough for them to be defined as independent structures [
178]. Although structures or concepts are often interrelated, they need to be defined statistically separately. This validity reveals that they are separate structures [
179]. The Fornell-Larcker criterion method is frequently used in the literature to investigate it [
172,
174]. According to this method, the square root of the AVE value of each factor is taken and these values are compared with the correlations of the factors. The square root of AVE must be more than the latent variable correlation for enabling discriminant validity [
172]. The values of the Fornell-Larcker criterion are given in
Table 18 below.
The values given in parentheses show the square roots of the AVE values for the factors, and the remaining values show the correlations between the latent variables in the table above. The square roots of the AVEs of the structures are lower than the correlations of the latent variables with others, so all factors have discriminant validity.
3.6. Stage 6: Nomological Validity
Nomological validity is a type of validity that reveals whether the developed scale reflects the relationships that exist based on theory or previous research. The scale must have nomological validity as well as convergent and discriminant validity to ensure construct validity in scale development studies. The relationships covered and supported by extant research or theoretically accepted principles are identified to test it. Afterwards, it is investigated whether the developed scale is consistent with them [
112].
The most theoretically similar measurement scale to the SCPSS in the extant literature is the SCP scale developed by Sindhuja (2014) [
93]. However, this scale could not be compared with the SCPSS, since the inter-factor correlation of it was not reported by the researcher. For this reason, the correlation coefficient (r) levels between the sub-dimensions of the SCPSS were compared with the existing literature to test the nomological validity. The interpretation of Taylor (1990) [
180] was considered for the r levels. Accordingly, if r is less than or equal to 0.35, it is considered a weak or low correlation. While 0.36 to 0.67 indicates moderate or modest correlation, 0.68 to 0.90 indicates a strong correlation. If r is equal to or above 0.90, it indicates high correlation. Accordingly, all relations between latent variables are modest and statistically significant (
Table 18,
p < 0.05). There is a positive correlation between all variables because its items were generated to point in the same direction. For instance, while the items for responsiveness are generated in the direction of decrease, the items for the cost dimension are generated in the direction of decrease. The findings of this study are in line with the existing literature, as the costs will decrease as the cycle times decrease in the SC [
181]. Cirtita and Glaser-Segura (2012) [
182] conducted a study on SCP measurement using the SCOR model of that period. The latent variables defined are the same except for agility. They have addressed flexibility performance rather than agility. When the findings of the two studies are compared, all correlation levels are moderate except for the correlation between cost and asset management. They found a strong correlation between cost and asset management, while the SCPSS found a modest level correlation. However, most of the study findings are generally coherent with each other, and the conceptual structure of the study is confirmed by the current literature.
3.7. Stage 7: Investigation of Bias Effect
A bias may occur depending on the perceptions of the participants in the data obtained through the questionnaire. It is called the “
bias effect” or the “
common method bias” [
183], which causes a measurement error and distorts the empirical values of the relationship among structures [
184]. It is necessary to check whether the data have a bias effect, and if there is bias, it must be eliminated [
185]. The controlling of bias is tested by adding an artificial observed variable called CMV, which stands for common method bias, to the final model, which exists in theory but does not exist in implementation. This variable affects all observed variables in the model equally [
186].
The CMV artificial variable was added to Alternative Model 4, which is the final model obtained in Stage 4, and it was tested. The new model’s degree of freedom is 451 and the chi-square value is 4355.31 (
p < 0.05). The difference between the degree of freedom of the new model and the final model is 30, and the two models are statistically different since the difference in chi-square values between the two models is more than 43,773. In this case, the optimal model is the final model because it has a lower chi-square value (Δχ
2 = 3393.54, Δdf = 30,
p < 0.05). Therefore, the scale does not have the bias effect. In addition, the inclusion of CMV in the model deteriorates the values of model fit indices (RMSEA = 0.19; SRMR = 0.43; CFI = 0.58; NFI = 0.55; AGFI = 0.39). The final form of the scale is presented in
Appendix B.
4. Discussion
SCM is the systematic and strategic coordination of SC activities among SC members with the aim of enhancing the long-term performance of both individual members and the entire SC [
187]. Therefore, the basis of effective SCM is based on measuring SCP. The most crucial difficulties encountered in measuring SCP are the selection and determination of the accurate and most appropriate measures and measurement systems [
10]. The SCOR model enables performance measures set regularly to be updated in line with SC requirements and ensures a universally accepted and cross-industry standard [
188,
189] and is currently the most widely used instrument for measuring SCP [
190]. On the other hand, it is quite difficult to observe this standard in measurement scales in this field.
There are various scale development studies for measuring SCP in the literature. When the extant scales are investigated, in fact, each scale reflects the structure of that era and the horizon of the researcher. However, some of these scales have some different limitations in terms of their scope and validity. For instance, the SCP scale developed by Lai et al. (2002) [
91] is service industry-oriented, since it was developed based on transportation logistics, and how many of these items implemented in the manufacturing industry should be investigated carefully. The items of the scale developed by Green et al. (2008) [
92] have focused heavily on delivery performance, but SCP needs to be addressed in the way of whole SC processes and performance attributes. The items of the SCP scale developed by Rana and Sharma (2019) [
95] have been designed for the pharmaceutical industry; the utilization of some items in other industries should be examined. The items related to the drug-related costs factor are examples of this. Undoubtedly, the dynamics of the SC for each industry are different from each other, but developing a scale that will cover whole industries will be a guide for future scale adaptation studies. The scale developed by Gawankar et al. (2016) [
94] is a comprehensive scale considering the retail industry in India. Although it is comprehensive in terms of the scope, the convergent and discriminant validity of these eight sub-dimensional scales has not been statistically tested.
There are two scale development studies based on the SCOR model in the extant literature. They have been developed by Lai et al. (2002) [
91] and Sindhuja (2004) [
93] according to this model’s performance measures and the existing literature, and so they were performed in line with the perspective of this study. However, the SC structure has changed considerably over the past decade, and the SCOR model, which has been updated accordingly, has changed greatly. For this reason, the scale developed by Lai et al. (2002) [
91] is different from today’s SC structure. The scale developed by Sindhuja (2014) [
93] is the most similar scale with the SCPSS in terms of its measures and sub-scales. It is based on the SCOR model version of the relevant era, as in this study. However, it has defined a structure called the supplier performance, unlike the SCPSS. The SCOR v13.0 performance attribute does not include this structure as a separate attribute. Moreover, the data collected for the development of the SCPSS have not indicated such a structure. SCOR v13.0 distinctly has a performance attribute named asset management or asset management efficiency. Furthermore, the four common latent variables (reliability, responsiveness, agility, and cost) have been defined differently in terms of their content. The SCPSS has described the structures involved in much more detail. For instance, the performance measures for the reliability dimension are not as comprehensive as those of the SCPSS. It only consists of order fulfillment rate, inventory turns, safety stocks, and inventory obsolescence. On the other hand, responsiveness deals with metrics related to cycle times in a general sense; it is not specific to SC activities such as transportation, storage, and manufacturing. This makes it difficult to identify the source of the problem when identifying SC problems through performance measurement. The agility dimension is more about market, demand, and delivery. However, the SCPSS considers transport and information system flexibility in agility performance. Cost dimension has been taken from a general perspective and consists of inbound and outbound costs, warehousing, inventory holding, and reducing product warranty claims. The SCPSS offers a wide range of cost performance metrics. It is a more up-to-date scale due to the SCOR model version it is based on and offers a more comprehensive scale in terms of its measures.
The disruptive technologies (DTs) offered by the Fourth Industrial Revolution make significant contributions to SCM [
191]. This study also reveals the performance measurements most affected by DTs among the pilot study item pool determined by SCOR v13.0, the extant literature, the effect of DTs, and expert group interviews. The final items of the scale are also the performance measures in which DTs enable the greatest improvement. The DTs include cyber physical systems, internet of things (IoT), artificial intelligence, autonomous robots, big data analytics, blockchain, cloud computing, 3D printers, augmented reality, autonomous (driverless) vehicles, digital twin, horizontal and vertical software integrations, simulation, cyber security, and 5G in this study because they have been widely used in the supply chain context [
168]. According to the findings of this study, these technologies contribute the most to warehouse efficiency and perfect order fulfillment rate in terms of reliability performance. Production cycle time and order fulfillment cycle time provide the highest contribution in terms of responsiveness performance. The item that these technologies contribute the most in terms of agility performance is information system flexibility. They contribute the highest to transportation costs in terms of cost performance, while they contribute the highest to return on working capital and CCC in terms of asset management.
This study is the most comprehensive study that deals with whole aspects of the SCP in the extant literature thanks to the pilot study item pool it generated. This study has revealed the key metrics in SCPM and the performance measures to which DTs contributed the most among the seventy-four meticulously generated items. When the contribution of the study to the understanding of SCP is investigated, cost performance has a significant place in SCP, since the performance metrics related to cost have the highest place among all measures. In other words, the latent variable cost is explained by more observed variables than others. In this respect, it is followed by the reliability latent variable. Reliability performance is based on lessening errors in whole processes at SC, predicting accuracy, and delivering on time, in the right quantity, at the right place. It is also a customer-oriented performance dimension [
192]. “
SCM is the management of upstream and downstream relationships with suppliers and customers to deliver superior customer value at lower cost to the SC as a whole” [
95]. As it is, the main purpose of SCM is to enable superior customer value at the lowest SC cost. Accordingly, the findings of the study are validated by the conceptual structure of SCM.
The performance metrics measuring SC flexibility and SC agility were taken together in this study, while addressing the items regarding the agility performance. The concepts of agility and flexibility are different in the literature. While agility is defined as “
the ability to react to external influences” [
49], flexibility is defined as “
the ability to respond to change with the least loss of time and money” [
193]. However, SC agility performance has been considered as flexibility performance in some research on SCP based on the SCOR model [
194] because flexibility combined with a skilled, knowledgeable, motivated, and empowered workforce discovers agility [
195], so flexibility reveals agility. The findings of this study revealed that DTs, which are among the performance metrics related to both concepts, only contribute to the performance metrics related to flexibility. These items are order flexibility, transportation flexibility, information system flexibility, and market flexibility. The intertwined nature of two concepts and the fact that they are utilized interchangeably in the current literature clearly explain this finding.
The current literature has proved the contribution of DTs to SCP, as have the findings of this study. IoT has contributed to efficient stocktaking and reducing accident risks in warehouses [
196]. Autonomous robots have the effect of increasing efficiency and safety in warehouses. They have also enabled the reduction of human-based operational errors [
197]. Augmented reality is a technology used in warehouses that enhances warehouse efficiency and lessens error rates in warehouses [
198]. The use of 5G technology in warehouse operations has enhanced warehouse efficiency by reducing transportation times, human errors, and accidents in warehouses [
199]. Artificial intelligence, autonomous robots, and 5G are important technologies that boost the perfect order fulfillment rate and shorten the order fulfillment cycle time [
200]. Using robots has shortened production cycle times [
201]. Three-dimensional printers have allowed for lower transportation costs in manufacturing processes [
202,
203]. IoT and big data analytics are among the technologies that contribute to information system flexibility [
204,
205]. Almost all DTs make a significant contribution to the return on working capital and the CCC [
206]. Consequently, this study’s findings are in line with the extant literature.
4.1. Theoretical Contributions
This study enables three main contributions theoretically. Firstly, the most fundamental and profound contribution to a field can be made on measurement [
207]. Although interest in measuring and improving SCP has enhanced considerably over the past decade [
208,
209] and its critical role in SCM has been highlighted, the studies of scale developed in this field are rare [
88]. When the extant scales in the literature are investigated, the necessity of an SCP scale that comprises comprehensive, standardized, and universally accepted performance measures is clear. This study fills this gap and enables the most significant theoretical contribution to the SC literature. It also adapted it to digital SCs, which are the contemporary SCs. Secondly, this study gathers the SCP measures in the literature under a single roof in the item pool for the pilot study and makes them more standard by considering them based on SCOR v13.0. Thirdly, SCP measures were discussed from a wide perspective in the pilot study item pool and contributed to the literature by determining the performance measures that DTs contributed the most among them.
4.2. Managerial Implications
This study enables many contributions practically. SC managers have significant duties to improve SCP [
210] and the SCPSS has the effect of facilitating these duties for them. This study presents a measurement scale based on its measures to overcome the difficulties faced by the SCOR model in practice. Information is significant for SC managers to make decisions [
211]. This measurement scale facilitates the decision-making of SC managers with the information it ensures. It provides them with a more practical and quicker performance measurement opportunity by enabling performance measurement to be performed by measurement scales because it prevents the performance measures from being calculated separately. Therefore, it also speeds up the decision-making processes. This will enable critical benefits to managers in times of crisis. SC managers can compare the present and past of the company in which they operate using the SCPSS. In addition, they could compare the SCPs of the industries in which they operate and accordingly, they could conduct improvement studies at the points where they fall behind according to the industry. They may also investigate how SCP relates to specific operations.
4.3. Limitations and Future Research
The main limitation of this study is that it does not consider sustainability performance in the SC. Nowadays, the sustainability performance is key for SCM [
212,
213]. The reason why sustainability performance measures were not included in this study is that the current SCOR model was version 13.0 at the time of the study and this version did not comprise sustainability performance attributes. Today, the current version for the SCOR model is version 14.0 (v14.0) and one of the major changes in this version is the addition of sustainability to the performance attributes [
188]. Sustainability performance in SCM plays a crucial role not only for the survival of companies, but also for their long-term success and development [
214]. SC managers must simultaneously consider how these decisions affect people, the planet, and the company, with a holistic approach when they make decisions [
215].
The SCOR v14.0 consists of resilience, economic, and sustainability attributes and sub-attributes related to these attributes. These sub-attributes are reliability, responsiveness, agility, cost, profit, assets, environmental, and social. The sustainability attribute in SCOR v14.0 consists of environmental and social sub-attributes [
188]. For further studies, scale development or adaptation studies can be carried out by considering these performance attributes and sub-attributes and their measures. However, there is one issue to take into consideration. The concept of sustainability is discussed in the academic literature in three dimensions: economic, environmental, and social [
216], and it is emphasized that these three dimensions should be handled in a balanced way and with a holistic approach [
217]. Therefore, if the SCP structure is to be considered in three sub-dimensions of resilience, economic, and sustainability in future studies, it is likely that the measures will be distributed differently from the SCOR model due to the conceptual structure of sustainability. Therefore, testing an eight sub-dimensions structure by considering eight sub-attributes will allow performance measures to be more clearly differentiated from each other statistically and conceptually.
SCP needs can be evaluated from different perspectives according to companies, processes, employees, technologies, and strategies [
218], and can be aimed to improve different performance measures [
219]. As an alternative, future researchers can perform a scale adaptation study by selecting measures in line with the company’s needs and goals from the performance measures list in the item pool for the pilot study. Moreover, SCM is a dynamic process [
220] and performance measurement systems need to be compatible with these dynamic environments and processes [
76,
77]. In recent decades, SCs have evolved in light of current developments such as sustainability, the use of information systems, and DTs [
221,
222,
223,
224]. Undoubtedly, the issues discussed or addressed in the SCs of tomorrow will be different from those of today. In line with the current developments, research agenda, and new concepts to be introduced to the literature, the SCP structure can be defined in various ways, within the horizon and knowledge of the researcher, and new scale development studies can be carried out in the future.
5. Conclusions
There are some key indications of their background today, although it is not known for certain what will change in the SCs of the future. Some key trends and indicators must be followed today to perceive the landscape of future SCs [
225]. This scale offers a standard, up-to-date, and comprehensive measurement scale by taking SCP measures in a more standardized and up-to-date position grounded on SCOR v13.0, while also extending it with extant academic literature in terms of the measures. It also has determined the contributions of DTs and the performance metrics that should be considered in digital SCs, since the measurement was performed according to the change in using DTs. Furthermore, this study ensures a measurement system that enables quicker and easier utilization for managerial implications by enabling SCPM to be performed with a measurement scale. Consequently, it will make a crucial contribution to the development of SCM, the foundation of future research, the SC literature, and to the SCs of the future.
This scale was developed with thirty-one items and five factors. As clarified above, this scale has acceptable model fit values for many types of scales (x2/df, RMSEA, SRMR, CFI, NFI, AGFI). The scale has convergent, discriminant, and nomological validity. It does not have the bias effect. The findings of this paper show that the scale can be utilized confidently by researchers and practitioners in their relevant fields. The scale was developed as “Supply Chain Performance Scale as SCOR-Based (SCPSS)”, but it can also be used to measure digital supply chain performance since the measurement is based on the use of DTs. It can be utilized with confidence in studies where the SCP or its sub-scales are variable in their current form.