Product Service System Conﬁguration Based on a PCA-QPSO-SVM Model

: To achieve sustainable development and improve market competitiveness, many manufacturers are transforming from traditional product manufacturing to service manufacturing. In this trend, the product service system (PSS) has become the mainstream of supply to satisfy customers with individualized products and service combinations. The diversiﬁed customer requirements can be realized by the PSS conﬁguration based on modular design. PSS conﬁguration can be deemed as a multi-classiﬁcation problem. Customer requirements are input, and speciﬁc PSS is output. This paper proposes an improved support vector machine (SVM) model optimized by principal component analysis (PCA) and the quantum particle swarm optimization (QPSO) algorithm, which is deﬁned as a PCA-QPSO-SVM model. The model is used to solve the PSS conﬁguration problem. The PCA method is used to reduce the dimension of the customer requirements, and the QPSO is used to optimize the internal parameters of the SVM to improve the prediction accuracy of the SVM classiﬁer. In the case study, a dataset for central air conditioning PSS conﬁguration is used to construct and test the PCA-QPSO-SVM model, and the optimal PSS conﬁguration can be predicted well for speciﬁc customer requirements.


Introduction
As the product market is nearly saturated, it is getting more challenging for manufacturers to satisfy diversified and individualized consumer requirements. To achieve sustainable growth, many manufacturers have transformed from product-oriented manufacturing to service-oriented manufacturing. This trend of product service system (PSS) to enhance competitiveness and sustainability leads to a massive research effort. The PSS method is based on the simultaneous development of products with tangible features and services surrounding the intangible features, to provide an integrated product that can effectively consider all life cycle stages of the product and related services [1,2].
PSS is described as a hybrid solution that comprises products and services for the purpose of increasing value for customers [3]. On the one hand, customers look forward to being provided with services that aim at enhancing the function and economic performance of the product, such as recycling and maintenance [4]. On the other hand, with ever fierce competition and more diversified customer needs, the low-value-added manufacturing paradigm can no longer meet the requirements of the market and the environment [5]. Therefore, PSS integrates the resources of various parties to meet the needs of customers and reduces the material flow in the consumption process by adding services, which is critical to improve social productivity, living standards, and environmental protection. The emergence of PSS has changed the mode of manufacturing and supply. To create maximum value for customers, PSS has to be individually configured. PSS configuration is to provide a concept by selecting appropriate product modules and service modules among the design modules in advance [6].
The foundation of configuration is to design product modules and service modules by modular design. Modular design is a commonly used approach to realize mass production and individualized design. It combines product elements to form a subsystem with specific functions and uses the subsystem as a versatile module to combine other subsystems to form a system and then produce a product with a certain function. Belkadi et al. [7] proposed a modular design process to support the configuration of production systems that meet the needs of regional markets. Rennpferdt et al. [8] verified the importance of modular product architecture for PSS development. To describe the dependencies between activities, Geng et al. [9] proposed a modular design method based on the fuzzy design structure matrix (FDSM) to obtain the result-oriented PSS. The PSS modular design is a very effective way to improve the response speed and reduce the cost of the industry.
The key in configuration is to combine product modules and service modules to satisfy the customer demands. The existing PSS configuration approaches include genetic algorithm (GA) [10,11], ontology modeling [12,13], convolutional neural network (CNN) [14], multi-objective programming [5,15], and so on. These approaches solve the PSS configuration problem as an optimization model with an objective or a fitness function. The goal of the PSS configuration is to provide a PSS concept that meets certain customer demands. Similar customer demands can be provided with the same PSS configuration. Therefore, the PSS configuration problem is regarded as a classification problem in this paper. The support vector machine (SVM) is an effective tool to solve classification problems. It mainly solves small samples and nonlinear problems. SVM has been successfully applied in the field of classification and regression. Demidova et al. [16] used an SVM based on NSGA-II to predict the effect of medical and technical diagnosis. Pławiak et al. [17] predicted Australian credit scoring based on a deep genetic cascade ensemble of SVM classifiers. Sun et al. [18] proposed a DNN decision-tree-based SVM model for speech emotion classification and recognition.
In this paper, PSS configuration is solved by the SVM method. The PSS configuration problem can be decomposed into three sub-problems: requirement feature extraction, SVM parameter optimization, and PSS configuration classification.
The first task is to reduce the dimension of customer requirements. Among the methods to reduce data dimension, principal component analysis (PCA) shows a strong advantage. PCA utilizes high variance to generate new components that store the most valuable information of elements [19]. Asante-Okyere et al. [20] adopted PCA as a dimension reduction method to improve the performance of the optimized least-squares support vector machine (LSSVM) and the adaptive neuro-fuzzy inference system-subtractive clustering method (ANFIS-SCM). Cao et al. [21] proposed a fault diagnosis method based on PCA and the Gaussian mixed model (GMM), and the PCA was used to reduce the data dimension and increase the feature resolution. Henry et al. [22] proposed an integrated framework based on the use of convolutional neural networks (CNN) and PCA to detect stiction and identify the severity of stiction. The PCA acts as a dimension reduction tool to visualize the extracted features. This paper uses the PCA algorithm to reduce the dimension of customer requirements.
The second task is to optimize SVM parameters. The selection of SVM internal parameters affects the classification performance and fitting effect of the SVM model. The commonly used algorithms are GA, particle swarm optimization (PSO), ant colony algorithm, and so on, while these algorithms have to set many parameters and are easy to fall into the problem of local optimization. Lu et al. [23] proposed a quantum particle swarm optimization (QPSO) algorithm from the perspective of quantum mechanics. The QPSO assumes that the particle swarm optimization system is a particle system that satisfies the basic hypothesis of quantum mechanics. In quantum space, particles do not have a specific trajectory, which allows particles to search for the global optimal solution in the entire feasible solution space. The lack of a defined trajectory means that the speed and position of the particles cannot be determined simultaneously in quantum space. The QPSO algorithm uses the Monte Carlo method to calculate the optimal position of the quantum particle to Sustainability 2021, 13, 9450 3 of 22 make sure the global randomness. It has the advantage of a fast convergence rate, and its fitness value is better than that of the traditional PSO [24,25]. This paper takes the advantage of the QPSO algorithm for global random optimization in SVM classification to find the optimal parameters.
The third task is to construct and test a PCA-QPSO-SVM model to predict PSS configuration. The customer requirements and PSS configurations are coded in the model. The PSS configuration is a class label. By comparing the original class label and the predictive class label, the accuracy of the model can be obtained. The PCA-QPSO-SVM model can predict a PSS configuration for specific customer requirements.
The rest of this paper is arranged as follows: In Section 2, the relevant literature is reviewed. The framework of the proposed PSS configuration approach is summarized in Section 3. Section 4 discusses how to use PCA and QPSO to optimize SVM parameters. In Section 5, the PSS configuration model based on PCA-QPSO-SVM is illustrated in detail. Section 6 is a case for an air-conditioner service system to demonstrate the efficacy of the proposed approach. Section 7 is the conclusion of this paper.

Literature Review
This section gives a comprehensive literature review consisting of three parts: PSS design, PSS configuration, and PSS configuration optimization.

PSS Design
PSS is described as a combination of products and services to provide the required utility or function to meet customer needs [26]. To help meet specific customer needs, various PSS development methods have been designed. Durugbo et al. [27] discussed an information flow modeling technique related to the key features of a PSS (such as origin, concepts, and applications). Their proposed model decomposed the iterative process between customers and service providers into different stages to provide total care products. Chiu et al. [28] proposed a process of developing PSS business models to expand current products or services into new market areas. Fargnoli et al. [29] proposed a method based on the collaborative use of PSS quality function deployment, axiomatic design, and service blueprint tools to provide the correlation between customer expectations, PSS components, and PSS modules. Lee et al. [30] utilized a structural services innovation approach to integrating PSS engineering and service engineering in designing intelligent product-service systems. Wang et al. [31] presented a PSS requirements heuristic framework to help engineering designers better improve designs or generate new design concepts.

PSS Configuration
The system framework for PSS configuration was first proposed by Aurich [32]. Based on the determination of the specific products and service features, technology and service configuration were combined to generate a customized PSS in this framework. The existing PSS configuration approaches include GA, multi-objective programming, ontology modeling, and so on. Sheng et al. [11] proposed a PSS configuration optimization method based on a genetic algorithm. Xuanyuan et al. [10] presented a multi-objective optimization for product configuration and used a multi-objective genetic algorithm to find the Pareto optimal solution set from feasible solutions. Song et al. [5] proposed a multi-objective optimization model for product extension service configuration to solve the problems of too many service solutions and low service delivery efficiency in PSS configuration. Dong et al. [12] proposed an ontology-based service product modeling approach for configuration. Shen et al. [13] presented an ontology-based approach to represent service configuration knowledge and developed a product-extended service configuration system. Based on the combination of quality function deployment and screening life cycle modeling, Haber and Fargnoli [33] proposed a product service system method, which is based on the combination of product service system quality function deployment and screening life cycle modeling tools, which is used to customize solutions for different usage patterns and achieve better environmental performance than independent products.
With the support of artificial intelligence, machine learning algorithms are introduced into the configuration study. To understand customer needs by using quantitative approaches, Yu et al. [34] proposed a knowledge-based artificial neural network (ANN) combined with a decision tree to substantiate customer needs to product specification. Zhou et al. [14] trained the CNN to obtain the complex nonlinear mapping relationship between the customer's demand attributes and the basic types of product-service systems, to determine the PSS configuration. Shen et al. [35] proposed a method that combines cluster neural network and rule algorithm to extract configuration rules between service parameters and functional requirements, customer features, or product features to gain higher effectiveness of configuration solution.
The SVM, as a kind of machine learning algorithm, has shown excellent results in solving small samples and multi-classification problems. Nowadays, a large number of scholars have used SVM for prediction and classification. Ahlawat et al. [36] developed a hybrid model of a powerful CNN and an SVM for the recognition of handwritten digits from the ministry dataset. Viloria et al. [37] used an SVM to predict the diagnosis of diabetes mellitus (DM). Zhou et al. [38] proposed a multi-model latent space-inducing ensemble SVM classifier for early dementia diagnosis with neuroimaging data. Shao et al. [39] studied the energy consumption of hotel buildings by establishing a support vector machine energy consumption prediction model.

PSS Configuration Optimization
In the process of predicting the PSS configuration by the SVM method, the optimization of the SVM classifier largely determines the optimization of the PSS configuration scheme.
In the PSS configuration problem, the PSS configuration concepts are classified by the customer requirements. There are a number of requirements that affect classification accuracy. Therefore, the quantity of the characteristics should be reduced, which means the dimension reduction in the SVM algorithm. There are many approaches for feature extraction or dimension reduction, such as neural network, PCA, and so on. Zhang et al. [40] used neural networks to extract more sensitive features with a shallow structure, thereby improving the accuracy of small sample classification. Kontonatsios et al. [41] developed a novel feature extraction method based on a neural network. Unlike the previous unsupervised feature extraction methods, this paper extracts document features in a supervised environment. Xiao et al. [42] presented a feature extractor based on Deep Convolutional Neural Networks to extract patterns shared by a family from entropy graphs automatically. Zhang et al. [43] used a PCA based on the AdaBoost algorithm to detect breast cancer. Ratnasari et al. [44] used a threshold-based region of interest (ROI) and PCA to reduce X-ray images, which obtained the best gray-level threshold of 150. Ma et al. [45] leveraged a deep convolutional neural network to extract image features and leveraged a PCA algorithm to achieve dimension reduction. Negi et al. [46] proposed a method that combines PCA and uncorrelated linear discriminant analysis (ULDA) to obtain the best features that control upper limb motion. Therefore, this paper selects PCA for feature extraction of the dataset. The basic idea of PCA is to use fewer mutually independent features to replace a large amount of information of the original features. It is favored by a large number of scholars.
On the other hand, the accuracy of the SVM classifier prediction depends heavily on the selection of the kernel function and penalty factor. Many scholars have used some optimization algorithms to optimize the parameters of SVM, such as the GA, grid search (GS), PSO, and so on.
(1) A genetic algorithm is a computational model that simulates the biological evolution process of natural selection and the genetic mechanism of Darwin's biological evolution theory. It is a method of searching for the optimal solution by simulating the natural evolution process. GA is often combined with SVM to optimize the parameters of SVM. Huang et al. [47] proposed the GA-SVM model to analyze the quantitative contribution of climate change and human activities to changes in vegetation coverage. The model used genetic algorithms to optimize the loss parameters, kernel function parameters, and loss function epsilon values in the SVM. Based on GA-SVM for rapid and effective screening of human papillomaviruses, Chen et al. [48] proposed a Raman spectroscopy technique that improved the accuracy of the model to optimize the penalty factors and nuclear function parameters in the SVM model. Li et al. [49] used the GA-SVM model to identify and classify flip chips. (2) Grid search is an exhaustive search method. By looping through the possible values of multiple parameters, it generates the parameter with the best performance, which is the optimal parameter. GS is a common method to optimize the parameters of SVM. Lv et al. [50] used PSO-SVM and GS-SVM to predict the corrosion rate of a steel cross-section. Tan et al. [51] proposed a method combining a successive projections algorithm (SPA) with an SVM based on GS-SVM to classify and identify apple samples with different degrees of bruising. Kong et al. [52] used the GS-SVM model to assess marine eutrophication states of coastal waters. (3) Particle swarm optimization is an optimization algorithm that simulates the predation behavior of bird swarms. The iteration process forms the optimal position and optimal direction, hence updating the particle swarm. Many scholars apply the PSO algorithm to optimize SVM parameters. García Nieto et al. [53] proposed a hybrid PSO optimized SVM model to predict the successful growth cycle of spirulina. Liu et al. [54] developed the PSO-SVM model to predict the daily PM2.5 level.
Bonah et al. [55] combined Vis-NIR hyperspectral imaging with pixel analysis and a new CARS-PSO-SVM model to classify foodborne bacterial pathogens.
In the PSO algorithm, the position and moving speed of the particles co-determine the movement trajectory of the particles. In quantum space, particles do not have a specific trajectory, which allows particles to search for the global optimal solution in the entire feasible solution space. The lack of a defined trajectory means that the speed and position of the particles cannot be determined simultaneously. Ch et al. [56] employed the SVM-QPSO model to forecast the streamflow values of Vijayawada station and Polavaram station of Andhra Pradesh in India. Li et al. [57] presented the use of an LSSVM algorithm based on quantum-behaved particle swarm optimization to establish the nonlinear relationship of slope stability. They verified that QPSO-SVM can provide a high degree of accuracy and reliability. Therefore, this paper applies the QPSO to optimize the SVM parameters to solve the PSS configuration problem.

Research Framework
PSS configuration design is based on the modular design concept. Product modules and service modules compose varied PSS configurations. In the design and sales departments, there are many historical data about customers' demands and their purchases schemes. This paper takes full advantage of the historical data to construct and test a PCA-QPSO-SVM model. In this model, PCA is used to reduce the dimension of customer requirements to extract the important features. QPSO is applied to optimize the SVM parameters, and a multi-classification SVM classifies the PSS configurations. The research framework is shown in Figure 1.
(1) Data preparation and preprocessing Customer requirements and the corresponding PSS configurations are obtained from the design and sales database. The representative historical dataset is extracted.
(2) Reduction of the requirement dimension The PCA algorithm is used to reduce the dimension of the customer requirements, then the processed dataset is used in the training set and testing set. (1) Data preparation and preprocessing Customer requirements and the corresponding PSS configurations are obtained from the design and sales database. The representative historical dataset is extracted.
(2) Reduction of the requirement dimension The PCA algorithm is used to reduce the dimension of the customer requirements, then the processed dataset is used in the training set and testing set.

(3) Construction of the QPSO-SVM model
The QPSO algorithm is used to optimize the kernel function б and the penalty factor C for SVM. Functions that satisfy Mercer's theorem [58] can be used as the kernel function. The kernel function is used to map the samples from the original space to the higher-dimension feature space and make them linearly separable in the feature space. The penalty factor is used to control the balance between margin maximization and deviation minimization [59]. By optimizing relevant parameters, the QPSO-SVM model is constructed and tested to verify the accuracy. In the QPSO-SVM model, the customer requirements are the input, and the PSS configuration meeting the customer requirements is the output.

(4) Prediction of the PSS configuration scheme
According to the PCA-QPSO-SVM model, PSS configuration can be predicted according to the inputting of new customer requirements.

Principal Component Analysis
The principal component analysis is a multivariate statistical method for analyzing the correlation between multiple features. The essential idea of PCA is to reveal the internal structure among multiple features through a few principal components. In other words, multiple features in the original data are reduced to a few features, simultaneously the relationships among the original data features are retained as much as possible. PCA has been playing an important role in the fields of artificial intelligence, data mining and image recognition, and so on. The mathematical derivation process of PCA is as follows.
Assuming that the projection of a sample point xi is T i W x on the hyperplane in the new space, then the covariance matrix of the sample points is If the projections of all sample points can be separated as much as possible, then the The QPSO algorithm is used to optimize the kernel function σ and the penalty factor C for SVM. Functions that satisfy Mercer's theorem [58] can be used as the kernel function. The kernel function is used to map the samples from the original space to the higherdimension feature space and make them linearly separable in the feature space. The penalty factor is used to control the balance between margin maximization and deviation minimization [59]. By optimizing relevant parameters, the QPSO-SVM model is constructed and tested to verify the accuracy. In the QPSO-SVM model, the customer requirements are the input, and the PSS configuration meeting the customer requirements is the output.

(4) Prediction of the PSS configuration scheme
According to the PCA-QPSO-SVM model, PSS configuration can be predicted according to the inputting of new customer requirements.

Principal Component Analysis
The principal component analysis is a multivariate statistical method for analyzing the correlation between multiple features. The essential idea of PCA is to reveal the internal structure among multiple features through a few principal components. In other words, multiple features in the original data are reduced to a few features, simultaneously the relationships among the original data features are retained as much as possible. PCA has been playing an important role in the fields of artificial intelligence, data mining and image recognition, and so on. The mathematical derivation process of PCA is as follows.
Assuming that the projection of a sample point x i is W T x i on the hyperplane in the new space, then the covariance matrix of the sample points is ∑ i W T x i x T i W after projection. If the projections of all sample points can be separated as much as possible, then the variance of the sample points would be maximized, so the optimization goal can be written as: where, W= (w 1 , w 2 , . . . , w d ). By using the Lagrange multiplier method for Formula (1), we can obtain: Then the obtained eigenvalues are sorted: Finally, the eigenvector corresponding to the top d' eigenvalues forms W * = (w 1 , w 2 , . . . , w d' ).
The steps of the PCA algorithm are as follows:

•
Step 3: Calculate the sample covariance matrix XX T and decompose the eigenvalues of the covariance matrix XX T .

Quantum Particle Swarm Optimization Algorithm
The PSO algorithm was proposed by Kennedy and Eberhart [60]. It is a random search algorithm based on swarm intelligence developed by simulating the foraging behavior of birds. Its basic idea is to find the optimal solution through collaboration and information sharing among individuals in a group. The mathematical description of the PSO algorithm is as follows: Assuming that the particle swarm size is M, the current optimal position of particle i in the n-dimensional space is expressed as: The current optimal position of the entire particle swarm is: During each iteration, the velocity update formula of each particle is: where, c 1 ,c 2 are learning factors; w is an inertia factor; rand 1 , rand 2 is a random value in [0, 1]; v i (t) is the velocity of the i-th particle, x i (t) is the position of the i-th particle; t is the number of iterations. The position update formula of each particle is: The PSO algorithm needs to set too many parameters, such as inertia factor w, learning factors c 1 , c 2 , etc. It is difficult to find the optimal parameters. Moreover, it is easy to fall into the dilemma of a local optimum due to the lack of randomness in the particle position change. To redress these problems, this paper applies the QPSO algorithm with higher performance. The QPSO algorithm cancels the particle's moving direction attribute and increases the randomness of the particle position change. The calculation process of the QPSO algorithm is as follows: The term mbest is introduced in the QPSO algorithm, which represents the average value of pbest. It can be expressed as: The current optimal position of particle i is: where Φ is uniform value in [0, 1]. The position update formula of each particle is: where µ is a uniform value in [0, 1] and is the new parameter, which can control the degree of contraction and expansion of the particle position generally. The value of alpha is less than 1.

Support Vector Machine
The SVM was originally proposed by Cortes and Vapnik [61]. It is a supervised learning algorithm that classifies datasets. The essential idea of the SVM is to construct a hyperplane in the sample space as a decision line that divides two different samples. The SVM can solve both binary classification problems and multi-class problems. In this paper, PSS configuration is regarded as a multi-class problem. The derivation process of the SVM algorithm is as follows.
Given a set S containing N training samples, S = {(x i , y i ), i = 1, 2, . . . , N}, the expression of classification hyperplane is: where w is the normal vector of the hyperplane and b is the translation distance of the hyperplane.
The objective function of partition hyperplane with the "maximum interval" can be expressed as: To enhance the error tolerance of the SVM classifier, a relaxation variable ξ i and the penalty factor C are introduced into the objective function. The objective function is expressed as: where ξ i is a non-negative slack variable, which is used to improve the generalization ability of the model. C is the penalty factor, which is used to control the balance between margin maximization and deviation minimization. For solving the non-linear classification problem, it is necessary to introduce the kernel function in the SVM. The kernel function is used to map the samples from the original space to the higher-dimension feature space and make them linearly separable in the feature space. In this paper, the Gaussian kernel function is used: The Lagrange multiplier is brought into the objective function. The optimization problem can be expressed as: where α i is the Lagrangian coefficient. The final optimized hyperplane is expressed as:

Optimization of the SVM Parameters
The quantity of the feature attributes affects the accuracy of the SVM model. The dimension of the feature attributes should be reduced first by the PCA algorithm. Then, the QPSO algorithm is used to optimize the C and σ of the SVM classifier based on the processes data. Finally, the optimal parameters are brought into the SVM model to realize the PCA-QPSO-SVM modeling. The PCA-QPSO-SVM optimization process is shown in Figure 2. The steps of constructing the PCA-QPSO-SVM model are as follows:

•
Step 1: Use a PCA algorithm to reduce the dimension of dataset Q to get a new dataset Q'.

•
Step 2: Determine the initial parameters of the QPSO, such as the number of particle swarms, the range of the parameters, the alpha value, and so on.

•
Step 3: Set the fitness function in QPSO. In this paper, the fitness function is the average of SVM cross-validation (CV), and its value represents the classification accuracy of the model. The optimal value pbest and the global optimal value gbest for each particle are updated by iterating the fitness function, where pbest is the penalty factor C, gbest is the kernel function σ.

•
Step 4: Calculate the optimal position mbest of the particle swarm and update the new position of each particle.

•
Step 5: Determine the end condition. When the optimal search reaches the maximum number of iterations, the optimal search ends; otherwise, go to Step 3.

•
Step 6: The optimal parameters (C, σ) are brought into the SVM model to conduct prediction.

PSS Configuration Based on the PCA-QPSO-SVM Model
This paper proposes a PCA-QPSO-SVM model and uses it to solve the PSS configuration problem. The process includes two parts: data collection and construction of the PCA-QPSO-SVM model.

PSS Configuration Based on the PCA-QPSO-SVM Model
This paper proposes a PCA-QPSO-SVM model and uses it to solve the PSS configuration problem. The process includes two parts: data collection and construction of the PCA-QPSO-SVM model.

Data Collecting and Processing
In real life, customers would have different needs for PSS, which generates diversified requirements for PSS configuration. It is vital to collect customer requirements features. Taking central air conditioning as an example, consumers usually use strong stability, high energy saving, and strong reliability to express their requirements. The vague requirements have to be understood and realized by a specific PSS configuration. The proposed PCA-QPSO-SVM model is used to classify customer requirements using PSS configuration as a class label. Individualized PSS configuration can be obtained, which will support the PSS design.
In the PCA-QPSO-SVM model, the input must be a real vector. We assume requirement features of customer C i are represented by E n = {CN i |1 < i < n, CN 1 , . . . , CN i , . . . , CN n }, and each feature CN i has t selectable values. Then numbers can be used to encode the t optional values. For example, E 1 = {0, 0, 2, . . . , 2, 1, −1} represents the requirement features of customer C 1 .
A PSS solution is comprised of product modules and service modules. Assuming there are s product modules in the PSS configuration framework, the module features for the product part can be expressed as P = {P i |1 < i < s}. There is one or more instances for each module characteristic. For example, there exist t available instances for module P i , and the set is {p i j |1 < j < t}. Similarly, the module features for the service part can be expressed as S = {S v |1 < v < u}. Assume there exist q available instances for module S v , and the set is {s u w |1 < w < q}. The product instances and service instances make up diversified PSS configuration instances. Assume there exist l PSS configuration instances, and PSS configuration instances can be expressed as PSS = {PSS k |1 < k < l}. Product and service modules have different selectable instance values, which can be represented by discrete numbers as tags. For example, a PSS configuration instance is expressed as {p 1 2 , p 2 2 , . . . , p i 3 , s 1 1 , s 2 2 , . . . , s u 3 }.

•
Step 1: Determine the product modules and service modules, then combine the corresponding instances to form different PSS configurations. According to the relevant historical data, the 'requirements-configuration' samples are collected to construct the model.

•
Step 2: Reduce the dimension of requirement features by using the PCA algorithm. QPSO is used to perform k-fold cross-validation (CV) to find the best Gaussian kernel function σ and the penalty factor C. For k-fold CV, the entire training set is divided into k subsets with an equal number of samples. One of the subsets is selected as the testing set, and the remaining k-1 subsets are the training set.

•
Step 3: Construct the multi-class SVM model by using the best parameter combination (C, σ) to test the testing set. After constructing a reliable classification model, PSS configuration can be predicted by inputting new customer requirements.

Case Study
Company A is a central air-conditioning manufacturer. To enhance the company's competitiveness and meet the diversified requirements of customers, company A decides to transform towards a service-oriented enterprise and provides customers with individualized PSS configuration. According to the modular design platform, and the PSS configuration design can be realized by combining different product instances and service instances. To quickly and accurately recommend the schemes to customers among diversified PSS configuration instances, the proposed PCA-QPSO-SVM model is applied in company A.
First, 100 data samples are extracted from the design and sales database. Then, the 100 data samples are divided into a training set and a testing set. The number of samples in the training set is 75 and that in the testing set is 25. Finally, the multi-classification PCA-QPSO-SVM model is used to predict the PSS configuration. The customer's need is used as the input of the PCA-QPSO-SVM model, and the PSS configuration scheme is used as the output of the model.
The original design data for constructing the PCA-QPSO-SVM model are shown in Table 4. The distribution of each feature is displayed with a frequency diagram and box plots, as shown in Figures 3 and 4.
It can be seen from Figures 3 and 4 that most of the datasets are relatively evenly distributed and have no abnormal feature attributes. Only features CN 8 , CN 9 , CN 11 have few outliers, which has little effect on the entire datasets, so the features cannot be eliminated easily. To see the correlation between each feature more intuitively, the establishment of a feature relationship heat map is shown in Figure 5.
In Figure 5, the correlation between features (CN 1 , CN 9 ) and (CN 4 , CN 10 ) is higher. Next, by establishing two coordinate systems with CN 1 as the x-axis, CN 9 as the y-axis and CN 4 as the x-axis, and CN 10 as the y-axis, we can observe whether they are linearly or nonlinearly related in the plane distribution. The result of feature correlation is shown in Figure 6.    Table 4. The distribution of each feature is displayed with a frequency diagram and box plots, as shown in Figure 3 and Figure 4.

Samples
Inputs  It can be seen from Figure 6 that the features (CN 1 , CN 9 ) and (CN 4 , CN 10 ) do not show a linear correlation, so the features CN 1 , CN 9 , CN 4 , and CN 10 cannot be eliminated easily.  It can be seen from Figure 3 and Figure 4 that most of the datasets are relatively evenly distributed and have no abnormal feature attributes. Only features CN8, CN9, CN11 have few outliers, which has little effect on the entire datasets, so the features cannot be eliminated easily. To see the correlation between each feature more intuitively, the establishment of a feature relationship heat map is shown in Figure 5. In Figure 5, the correlation between features (CN1, CN9) and (CN4, CN10) is higher. Next, by establishing two coordinate systems with CN1 as the x-axis, CN9 as the y-axis and CN4 as the x-axis, and CN10 as the y-axis, we can observe whether they are linearly or nonlinearly related in the plane distribution. The result of feature correlation is shown in Figure 6.  It can be seen from Figure 3 and Figure 4 that most of the datasets are relatively evenly distributed and have no abnormal feature attributes. Only features CN8, CN9, CN11 have few outliers, which has little effect on the entire datasets, so the features cannot be eliminated easily. To see the correlation between each feature more intuitively, the establishment of a feature relationship heat map is shown in Figure 5. In Figure 5, the correlation between features (CN1, CN9) and (CN4, CN10) is higher. Next, by establishing two coordinate systems with CN1 as the x-axis, CN9 as the y-axis and CN4 as the x-axis, and CN10 as the y-axis, we can observe whether they are linearly or nonlinearly related in the plane distribution. The result of feature correlation is shown in Figure 6. It can be seen from Figure 6 that the features (CN1, CN9) and (CN4, CN10) do not show a linear correlation, so the features CN1, CN9, CN4, and CN10 cannot be eliminated easily.

Dimension Reduction of Requirement Feature
To improve the classification performance of the SVM model, the original feature set is processed to reduce dimension. First, the PCA algorithm is used to reduce the dimen-

Dimension Reduction of Requirement Feature
To improve the classification performance of the SVM model, the original feature set is processed to reduce dimension. First, the PCA algorithm is used to reduce the dimension of 11 requirement features in the original 100 sample data. Because the values of the data samples are all in {−2, -1, 0, 1, 2}, there is no need for StandardScaler processing. Second, the principal components are determined by calculating the covariance of each feature. The result of the calculation is shown in Figure 7. In Figure 7, the first and second principal components account for most of the variance, accounting for 63.62%. Therefore, the parameter set of PCA is n_components = 2. The data after PCA dimension reduction is shown in Table 5.  In Figure 7, the first and second principal components account for most of the variance, accounting for 63.62%. Therefore, the parameter set of PCA is n_components = 2. The data after PCA dimension reduction is shown in Table 5. To demonstrate the distribution of these 100 samples more clearly after PCA dimensionality reduction, this paper visualizes these 100 samples. Take X 1 as the x-axis and X 2 as the y-axis, and the established rectangular coordinate system is shown in Figure 8. To demonstrate the distribution of these 100 samples more clearly after PCA dimensionality reduction, this paper visualizes these 100 samples. Take X1 as the x-axis and X2 as the y-axis, and the established rectangular coordinate system is shown in Figure 8. To visualize the data, the original 11-dimensional features are reduced to 2-dimensional features through the PCA algorithm, thereby reducing nine attribute features. To verify the relationship between the 2-dimensional feature and the original 11-dimensional feature, and we can establish a heat map to reflect the relationship intui- To visualize the data, the original 11-dimensional features are reduced to 2-dimensional features through the PCA algorithm, thereby reducing nine attribute features. To verify the relationship between the 2-dimensional feature and the original 11-dimensional feature, and we can establish a heat map to reflect the relationship intuitively between the original features and PCA principal components, as shown in Figure 9. To visualize the data, the original 11-dimensional features are reduced to 2-dimensional features through the PCA algorithm, thereby reducing nine attribute features. To verify the relationship between the 2-dimensional feature and the original 11-dimensional feature, and we can establish a heat map to reflect the relationship intuitively between the original features and PCA principal components, as shown in Figure  9. In Figure 9, the color from dark to light represents a value from −0.4 to 0.5, if the value of a feature is positive, it means that it is positively related to the principal component. If it is negative, then the opposite. It can be seen from Figure 9 that the correlation coefficient of X1 is extremely low only with CN1, CN2, CN9, CN10, and the correlation coefficients of X2 are extremely low only with CN3, CN4, CN5, CN10. Then it indicates that the substitution effect of X1 and X2 is better.

QPSO-SVM Model Construction and Parameters Setting
First, the 2-dimensional dataset obtained by the PCA algorithm is used as the dataset of the multi-class SVM model. Then, the QPSO algorithm is used for parameter optimization of 2-fold CV to find out the optimal parameter pair (C, б). In the parameter opti- Figure 9. The heat map of the relationship between principal components and eigenvalues.
In Figure 9, the color from dark to light represents a value from −0.4 to 0.5, if the value of a feature is positive, it means that it is positively related to the principal component. If it is negative, then the opposite. It can be seen from Figure 9 that the correlation coefficient of X 1 is extremely low only with CN 1 , CN 2 , CN 9 , CN 10 , and the correlation coefficients of X 2 are extremely low only with CN 3 , CN 4 , CN 5 , CN 10 . Then it indicates that the substitution effect of X 1 and X 2 is better.

QPSO-SVM Model Construction and Parameters Setting
First, the 2-dimensional dataset obtained by the PCA algorithm is used as the dataset of the multi-class SVM model. Then, the QPSO algorithm is used for parameter optimization of 2-fold CV to find out the optimal parameter pair (C, σ). In the parameter optimization process, 75 randomly selected samples were used to train the QPSO-SVM model. In the QPSO algorithm, a parameter pair (C, σ) was obtained by each iteration, when the number of iterations reaches the maximum and output the optimal parameter pair (C, σ). The specific parameter settings of the QPSO algorithm are shown in Table 6. The optimal parameter pair (C, σ) was calculated by setting the parameter value of the QPSO algorithm, and then it was used as the parameter sets the value of the multi-class SVM model for prediction analysis. The entire process was completed by Pycharm 2019.

Prediction and Comparative Analysis of PCA-QPSO-SVM Model
To evaluate the performance of the training model, classification accuracy is used as an index. In each iteration of the PCA-QPSO-SVM model, the training accuracy of training samples is shown in Figure 10.
ti-class SVM model for prediction analysis. The entire process was completed by Pycharm 2019.

Prediction and Comparative Analysis of PCA-QPSO-SVM Model
To evaluate the performance of the training model, classification accuracy is used as an index. In each iteration of the PCA-QPSO-SVM model, the training accuracy of training samples is shown in Figure 10. According to QPSO parameter optimization, when the number of iterations reaches 8, the classification accuracy under 2-fold CV is the highest (95.01%), and the optimal parameter values C = 3.01, б = 2.41 are obtained at this time. Then the optimal parameters are used in the multi-class SVM model, and 25 samples in the testing set are used to test the classification accuracy of the model. The test results are shown in Figure 11. It turns out that all 25 test samples are predicted correctly, and the accuracy rate is 100%, which verifies that the performance of the QPSO-SVM model is excellent. According to QPSO parameter optimization, when the number of iterations reaches 8, the classification accuracy under 2-fold CV is the highest (95.01%), and the optimal parameter values C = 3.01, σ = 2.41 are obtained at this time. Then the optimal parameters are used in the multi-class SVM model, and 25 samples in the testing set are used to test the classification accuracy of the model. The test results are shown in Figure 11. It turns out that all 25 test samples are predicted correctly, and the accuracy rate is 100%, which verifies that the performance of the QPSO-SVM model is excellent. To verify the validity of the PCA-QPSO-SVM model prediction, this paper compares PCA-QPSO-SVM with PCA-PSO-SVM, PSO-SVM, GA-SVM, and GS-SVM. In the experiment, various prediction algorithms are run 20 times, respectively. For each method, the optimal prediction result is selected as the final result. The final results are compared as shown in Table 7. The simulation results show that the prediction accuracy of the PCA-QPSO-SVM model is 100%, which is better than other algorithms. Therefore, the proposed model has a better adaptability and predictive ability. In addition, the mean square error of PCA-QPSO-SVM is 0, indicating that the PCA-QPSO-SVM model has a good predictive modeling effect.   To verify the validity of the PCA-QPSO-SVM model prediction, this paper compares PCA-QPSO-SVM with PCA-PSO-SVM, PSO-SVM, GA-SVM, and GS-SVM. In the experiment, various prediction algorithms are run 20 times, respectively. For each method, the optimal prediction result is selected as the final result. The final results are compared as shown in Table 7. The simulation results show that the prediction accuracy of the PCA-QPSO-SVM model is 100%, which is better than other algorithms. Therefore, the proposed model has a better adaptability and predictive ability. In addition, the mean square error of PCA-QPSO-SVM is 0, indicating that the PCA-QPSO-SVM model has a good predictive modeling effect. To verify the reliability of the model to solve actual problems, this article will test the consistency of the PSS configuration provided by a new customer's demand with the actual situation. The first new customer needs to configure central air-conditioning for the hospital. Its requirements are good stability and low noise (that is, the environmental protection is very high). Because the patient is considered, the comfort is also very high. The input is {2, 2, −1, −1, 0, 2, 1, 2, 1, 1, 0}, and the output result is the configuration plan "4", namely {A 4 , B 2 , C 1 , D 2 , E 1 , F 2 , F 4 , G 1 , H 1 , I 2 , J 3 , K 2 , L 2 , M 3 , N 3 , N 4 }. In practice, A 1 has low noise and good stability. Because E 1 has the characteristics of low wind pressure and small air volume, it is more comfortable. Configuration plan "4" meets the needs of customers and verifies the reliability of the model.

Discussion of Results
In the case study, there exist unimportant requirement features, so it is vital to reduce the dimension of customer requirement features. In this paper, we retain 95% information on requirement features. To facilitate the visualization, customer requirements are reduced from 11 features to 2 principal components by the PCA algorithm. Because the PCA algorithm can filter out some small influencing features by reducing the dimension, it will greatly improve the calculation speed and prediction accuracy of SVM. By comparing it with the PSO-SVM model, the PCA-PSO-SVM reaches the optimal result in the 13th iteration, the PSO-SVM model reaches the optimal result in the 21st iteration. In addition, the PCA-PSO-SVM model has the higher prediction accuracy (92%) than the PSO-SVM model (64%). Moreover, the PCA-PSO-SVM model has the lower mean square error (1.0) than the PSO-SVM model (1.04).
The QPSO algorithm is used to optimize the penalty factor C and the kernel function σ of SVM in this paper. The experimental results show that the optimal parameters (C = 3.01, σ = 2.41) and the highest classification accuracy (95.01%) are obtained by the PCA-QPSO-SVM model. By comparing with the PCA-PSO-SVM model, the proposed approach in this paper reaches the optimal result in the ninth iteration, which is faster than the PCA-PSO-SVM model (the 18th iteration). In terms of prediction accuracy, the proposed approach is 100%, which is higher than the PCA-PSO-SVM model (92%). In addition, the proposed approach has the lower mean square error (0) than the PCA-PSO-SVM model (1.0). These verified that the QPSO algorithm has the advantages of fast calculation ability and strong optimization ability in comparison with other optimization approaches.
By comparing with the GA-SVM model, the PSO algorithm can find the optimal parameters faster and lower mean square error. The PSO algorithm reached the optimal in the 21st iteration, which is faster than the GA-SVM model (the 27th iteration). Although the prediction accuracy the PSO-SVM model is similar to the GA-SVM, the mean square error of the PSO-SVM model is 1.04, which is lower than the GA-SVM model (1.56).
Similarly, by comparing with the GS-SVM model, the PSO algorithm can find the optimal parameters faster and with a better prediction accuracy. The PSO algorithm reached the optimal in the 21st iteration, which is faster than the GA-SVM model (the 38th iteration). The prediction accuracy of the PSO-SVM model is 64%, which is higher than the GS-SVM model. In addition, the mean square error of the PSO-SVM model is 1.04, which is lower than the GS-SVM model (3.32).
In general, the experimental results show that it is effective for predicting PSS configuration by the proposed PCA-QPSO-SVM model. The proposed approach has the highest accuracy and best performance by comparing with other approaches, which can provide a more accurate PSS configuration that meets customer requirements. In the product service area, this is beneficial to customers who will receive a better value offering and the manufacturer who can identify the main aspects of its solution more effectively.
The limitation of the dataset is that the sample are relatively small. The optimal parameters and the highest classification accuracy are reached through 50 iterations.

Conclusions and Future Research
With the increasingly fierce market competition, more and more manufacturers combine products and services into PSS to provide customers with greater value. Manufacturers must accurately determine which configuration meets customer needs better. The PSS configuration is deemed as a multi-classification problem in this paper. A PCA-QPSO-SVM model is proposed to solve this problem. The validation of the proposed model in air conditioner configuration indicates that it can be used as an effective prediction model for PSS configuration. To sum up, the proposed approach reveals the following features.
The PSS configuration design can be realized by machine learning. The historical design and sales data have great value in supporting individualized PSS design. A multiclassification SVM helps PSS providers to easily predict configuration schemes with higher customer satisfaction.
Compared with the conventional SVM model, the proposed model considers the impact of feature dimension and the selection of key parameters. The PCA algorithm has the advantage of convenience in reducing the dimensions of the dataset. In the process of optimizing SVM parameters, QPSO does not fall into the problem of local optimization and has a better fitness value. The usage of QPSO in optimizing SVM parameters can improve classification accuracy.
Although the proposed optimization method can predict the PSS configuration well, it is restricted by the training data. The predicted model performs effectively when there are adequate and reliable data. Completely new customer requirements are hard to process and generate the appropriate PSS configuration. Therefore, in future research, we can calculate the similarity between new customer requirements and previous customer requirements, and the PSS configuration corresponding to the maximum similarity can be recommended to the new customer.
Author Contributions: Conceptualization, Z.C. and X.G.; methodology, Z.C. and X.G.; writingreview and editing, Z.C. and X.G. All authors have read and agreed to the published version of the manuscript.

Conflicts of Interest:
The authors declare no conflict of interest.