Next Article in Journal
Gear Shifting in Biological Energy Transduction
Next Article in Special Issue
Asymmetric Measurement-Device-Independent Quantum Key Distribution through Advantage Distillation
Previous Article in Journal
Benign and Malignant Breast Tumor Classification in Ultrasound and Mammography Images via Fusion of Deep Learning and Handcraft Features
Previous Article in Special Issue
Quantum Image Encryption Based on Quantum DNA Codec and Pixel-Level Scrambling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Applicability of Quantum Machine Learning

by
Sebastian Raubitzek
1,2,* and
Kevin Mallinger
1,2
1
Data Science Research Unit, TU Wien, Favoritenstrasse 9-11/194, 1040 Vienna, Austria
2
SBA Research gGmbH, Floragasse 7/5.OG, 1040 Vienna, Austria
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(7), 992; https://doi.org/10.3390/e25070992
Submission received: 7 June 2023 / Revised: 22 June 2023 / Accepted: 26 June 2023 / Published: 28 June 2023
(This article belongs to the Special Issue Advances in Quantum Computing)

Abstract

:
In this article, we investigate the applicability of quantum machine learning for classification tasks using two quantum classifiers from the Qiskit Python environment: the variational quantum circuit and the quantum kernel estimator (QKE). We provide a first evaluation on the performance of these classifiers when using a hyperparameter search on six widely known and publicly available benchmark datasets and analyze how their performance varies with the number of samples on two artificially generated test classification datasets. As quantum machine learning is based on unitary transformations, this paper explores data structures and application fields that could be particularly suitable for quantum advantages. Hereby, this paper introduces a novel dataset based on concepts from quantum mechanics using the exponential map of a Lie algebra. This dataset will be made publicly available and contributes a novel contribution to the empirical evaluation of quantum supremacy. We further compared the performance of VQC and QKE on six widely applicable datasets to contextualize our results. Our results demonstrate that the VQC and QKE perform better than basic machine learning algorithms, such as advanced linear regression models (Ridge and Lasso). They do not match the accuracy and runtime performance of sophisticated modern boosting classifiers such as XGBoost, LightGBM, or CatBoost. Therefore, we conclude that while quantum machine learning algorithms have the potential to surpass classical machine learning methods in the future, especially when physical quantum infrastructure becomes widely available, they currently lag behind classical approaches. Our investigations also show that classical machine learning approaches have superior performance classifying datasets based on group structures, compared to quantum approaches that particularly use unitary processes. Furthermore, our findings highlight the significant impact of different quantum simulators, feature maps, and quantum circuits on the performance of the employed quantum estimators. This observation emphasizes the need for researchers to provide detailed explanations of their hyperparameter choices for quantum machine learning algorithms, as this aspect is currently overlooked in many studies within the field. To facilitate further research in this area and ensure the transparency of our study, we have made the complete code available in a linked GitHub repository.

1. Introduction

Quantum computing has recently gained significant attention due to its potential to solve complex computational problems exponentially faster than classical computers [1]. Quantum machine learning (QML) is an emerging field that combines the power of quantum computing with traditional machine learning techniques to solve real-world problems more efficiently [2,3]. Various QML algorithms have been proposed, such as quantum kernel estimator [4] and variational quantum circuit [5,6], which have shown promising results in diverse applications, including pattern recognition and classification tasks [7,8,9].
In this study, we aim to compare QKE (quantum kernel estimator) and VQC (variational quantum circuit) with powerful classical machine learning methods such as XGBoost [10], Ridge [11], Lasso [12], LightGBM [13], CatBoost [14], and MLP (multilayer perceptron) [15] on six benchmark datasets partially available in the Scikit-learn library [16] as well as artificially generated datasets. To ensure a fair comparison on the benchmark datasets, we perform a randomized search to optimize hyperparameters for each algorithm, thereby providing a comprehensive statistical comparison of their performance. Furthermore, we provide the full program code in a GitHub repository [17] to make our results reproducible and boost research that can potentially build on our approach.
Since quantum machines are not readily accessible, we can only compare these algorithms’ performance on simulated quantum circuits. Although this approach does not reveal the full potential of quantum machine learning, it does highlight how the discussed quantum machine learning methods handle different levels of complexity inherent in the datasets. For this reason, we also developed a method to generate artificial datasets based on quantum mechanical concepts to provide a prototype for a particularly well-suited dataset for quantum machine learning. This will estimate the possible improvements that quantum machine learning algorithms can offer over classical methods in terms of accuracy and efficiency, considering the computational resources needed to simulate quantum circuits.
In this study, we address and partially answer the following research questions:
  • How do QKE and VQC algorithms compare to classical machine learning methods such as XGBoost, Ridge, Lasso, LightGBM, CatBoost, and MLP regarding accuracy and efficiency on simulated quantum circuits?
  • To what extent can a randomized search to find a suitable set of hyperparameters make the performance of quantum machine learning algorithms comparable to classical approaches?
  • What are the limitations and challenges associated with the current state of quantum machine learning, and how can future research address these challenges to unlock the full potential of quantum computing in machine learning applications?
  • Do quantum machine learning algorithms outperform regular machine learning algorithms on datasets constrained by the rules of quantum mechanics? Thus, do they provide a quantum advantage for datasets that exhibit strong symmetry properties in terms of adhering to Lie algebras?
The research presented in this article is partially inspired by the work of Zeguendry et al. [18], which offers an excellent review and introduction to quantum machine learning. However, their article does not delve into the tuning of hyperparameters for the quantum machine learning models employed, nor does it provide ideas on creating best-suited data for quantum machine learning classification tasks. We aim to expand the toolbox of quantum machine learning, first by discussing the space of Hyperparameters and second by providing a prototype for generating “quantum data”. Furthermore, this analysis will help determine the current state of quantum machine learning performance and whether researchers should employ these algorithms in their studies.
We provide the entire program code of our experiments and all the results in a GitHub repository, ensuring the integrity of our findings, fostering research in this field, and offering a comprehensive code for researchers to test quantum machine learning on their classification problems. Thereby, a key contribution of our research is not only the provision of a single implementation of a quantum machine learning algorithm, but also the execution of a randomized search for potential hyperparameters of both classical and quantum machine learning models and a novel approach for generating artificial classification problems based on concepts inherent to quantum mechanics, i.e., Lie groups and algebras.
This article is structured as follows: Section 2 discusses relevant and related work. In Section 3, we describe, reference, and, to some degree, derive all employed techniques. We will not discuss the mathematical details of all employed algorithms here, but rather refer the interested reader to the referenced sources. Section 4 describes our performed experiments in detail, followed by the obtained results in Section 5, which also features a discussion of our findings. Finally, we conclude our findings in Section 6.

2. Related Work

Considerable research was conducted in recent years to advance quantum machine learning environments and their application field. This starts in the data encoding process, in which Schuld and Killoran [3] investigated quantum machine learning in feature Hilbert spaces theoretically. They proposed a framework for constructing quantum embeddings of classical data to enable quantum algorithms that learn and classify data in quantum feature spaces.
Further research was conducted on introducing novel architectural frameworks. For this, Mitarai et al. [19] presented a method called quantum circuit learning (QCL), which uses parameterized quantum circuits to approximate classical functions. QCL can be applied to supervised and unsupervised learning tasks, as well as reinforcement learning.
Havlíček et al. [4] introduced a quantum-enhanced feature space approach using quantum circuits. This work demonstrated that quantum computers can effectively process classical data with quantum kernel methods, offering the potential for exponential speedup in certain applications.
Furthermore, Farhi and Neven [20] explored the use of quantum neural networks for classification tasks on near-term quantum processors. They showed that quantum neural networks can achieve good classification performance with shallow circuits, making them suitable for noisy intermediate-scale quantum (NISQ) devices.
Other research focused on the advancement of applying quantum fundamentals on classical machine learning applications. Hereby, Rebentrost et al. [21] introduced the concept of a quantum support vector machine for big data classification. They showed that the quantum version of the algorithm can offer exponential speedup compared to its classical counterpart, specifically in the kernel evaluation stage.
To advance the application field of quantum machine learning, Liu and Rebentrost [22] proposed a quantum machine learning approach for quantum anomaly detection. They demonstrated that their method can efficiently solve classification problems, even when the data have a high degree of entanglement.
In this regard, it is worth mentioning the work of Broughton et al. [23] introduced TensorFlow Quantum, an open-source library for the rapid prototyping of hybrid quantum-classical models for classical or quantum data. They demonstrated various applications of TensorFlow Quantum, including supervised learning for quantum classification, quantum control, simulating noisy quantum circuits, and quantum approximate optimization. Moreover, they showcased how TensorFlow Quantum can be applied to advanced quantum learning tasks such as meta-learning, layer-wise learning, Hamiltonian learning, sampling thermal states, variational quantum eigensolvers, classification of quantum phase transitions, generative adversarial networks, and reinforcement learning.
In the review paper by Zeguendry et al. [18], the authors present a comprehensive overview of quantum machine learning from the perspective of conventional machine learning techniques. The paper starts by exploring the background of quantum computing, its architecture, and an introduction to quantum algorithms. It then delves into several fundamental algorithms for QML, which form the basis of more complex QML algorithms and can potentially offer performance improvements over classical machine learning algorithms. In the study, the authors implement three machine learning algorithms: quanvolutional neural networks, quantum support vector machines, and variational quantum circuit. They compare the performance of these quantum algorithms with their classical counterparts on various datasets. Specifically, they implement quanvolutional neural networks on a quantum computer to recognize handwritten digits and compare its performance to convolutional neural networks, stating the performance improvements by quantum machine learning.
Despite these advancements, it is important to note that some of the discussed papers may not have used randomized search CV from Scikit-learn to optimize the classical machine learning algorithms, thereby overstating the significance of quantum supremacy. Nevertheless, the above-mentioned works present a comprehensive overview of the state of the art in quantum machine learning for classification, highlighting the potential benefits of using quantum algorithms in various forms and applications.

3. Methodology

This section presents our methodology for comparing the performance of classical and quantum machine learning techniques for classification tasks. Our approach is designed to provide a blueprint for future experiments in this area of research. We employ the Scikit-learn library, focusing on the inbuilt functions to select a good set of hyperparameters, i.e., RandomizedSearchCV to compare classical and quantum machine learning models. We also utilize the Qiskit library to incorporate quantum machine learning techniques into our experiments, [24]. The selected datasets for our study include both real-world and synthetic data, enabling a comprehensive evaluation of the classifiers’ performance.

3.1. Supervised Machine Learning

Supervised machine learning is a subfield of artificial intelligence that focuses on developing algorithms and models to learn patterns and make decisions or predictions based on data [25,26]. The main goal of supervised learning is to predict labels or outputs of new, unseen data given a set of known input–output pairs (training data). This section briefly introduces several classical machine learning techniques used for classification tasks, specifically in the context of supervised learning. These techniques serve as a baseline to evaluate the applicability of quantum machine learning approaches, which are the focus of this paper. Furthermore, we will then introduce the employed quantum machine learning algorithms.
One of the essential aspects of supervised machine learning is the ability to predict/classify data. The models are trained using a labeled dataset, and then the performance of the models is evaluated based on their accuracy in predicting the labels of previously unseen test samples [27]. This evaluation is crucial to estimate the model’s ability to generalize the learned information when making predictions on new, real-world data.
Various techniques, such as cross-validation and train-test splits, are often used to obtain reliable performance estimates of the models [28]. By comparing the performance of different models, researchers and practitioners can determine which model or algorithm is better suited for a specific problem domain.

3.2. Classical Supervised Machine Learning Techniques

The following list describes the employed algorithms that serve as a baseline for the afterwards described and later tested quantum machine learning algorithms.
  • Lasso and Ridge Regression/Classification: Lasso (least absolute shrinkage and selection operator) and Ridge Regression are linear regression techniques that incorporate regularization to prevent overfitting and improve model generalization [11,12]. Lasso uses L1 regularization, which tends to produce sparse solutions, while Ridge Regression uses L2 regularization, which prevents coefficients from becoming too large.
    Both of these regression algorithms can also be used for classification tasks.
  • Multilayer Perceptron: MLP is a type of feedforward artificial neural network with multiple layers of neurons, including input, hidden, and output layers [15]. MLPs are capable of modeling complex non-linear relationships and can be trained using backpropagation.
  • Support Vector Machines (SVM): SVMs are supervised learning models used for classification and regression tasks [29]. They work by finding the optimal hyperplane that separates the data into different classes, maximizing the margin between the classes.
  • Gradient Boosting Machines: Gradient boosting machines are an ensemble learning method that builds a series of weak learners, typically decision trees, to form a strong learner [30]. The weak learners are combined by iteratively adding them to the model while minimizing a loss function. Notable gradient boosting machines for classification tasks include XGBoost [10], CatBoost [14], and LightGBM [13]. These three algorithms have introduced various improvements and optimizations to the original gradient boosting framework, such as efficient tree learning algorithms, handling categorical features, and reducing memory usage.

3.3. Quantum Machine Learning

Quantum machine learning is an emerging interdisciplinary field that leverages the principles of quantum mechanics and quantum computing to improve or develop novel algorithms for machine learning tasks [2]. This section introduces two key quantum machine learning techniques, Variational Quantum Circuit and Quantum Kernel Estimator, and discusses their connections to classical machine learning techniques. Additionally, we briefly introduce Qiskit Machine Learning, a Python package developed by IBM for implementing quantum machine learning algorithms. Furthermore, we want to mention the work done by [18] for a review of quantum machine learning algorithms and a more detailed discussion of the employed algorithms.

3.3.1. Variational Quantum Circuit (VQC)

VQC is a hybrid quantum-classical algorithm that can be viewed as a quantum analog of classical neural networks, specifically the multilayer perceptron [5,6]. VQC employs a parametrized quantum circuit, which is trained using classical optimization techniques to find the optimal parameters for classification tasks. The learned quantum circuit can then be used to classify new data points.
Figure 1 illustrates the schematic depiction of the variational quantum circuit, which involves preprocessing the data, encoding it onto qubits using a feature map, processing it through a variational quantum circuit (Ansatz), measuring the final qubit states, and optimizing the circuit parameters θ , Thus, the main building blocks of the VQC are as follows:
  • Preprocessing: The data are prepared and preprocessed before being encoded onto qubits.
  • Feature map encoding (yellow in the figure): The preprocessed data are encoded onto qubits using a feature map.
  • Variational quantum circuit (Ansatz) (steel-blue in the figure): The encoded data undergo processing through the variational quantum circuit, also known as the Ansatz, which consists of a series of quantum gates and operations.
  • Measurement (orange in the figure): The final state of the qubits is measured, providing probabilities for the different quantum states.
  • Parameter optimization (Optimizer): The variational quantum circuit is optimized by adjusting the parameters θ , such as the rotations of specific quantum gates, to improve the outcome/classification.

3.3.2. Quantum Kernel Estimator

QKE is a technique that leverages the quantum computation of kernel functions to enhance the performance of classical kernel methods, such as support vector machines [4,31]. By computing the kernel matrix using quantum circuits, QKE can capture complex data relationships that may be challenging for classical kernel methods to exploit.
The main building blocks for the employed QKE, which are depicted in Figure 2 are as follows:
  • Data preprocessing: The input data are preprocessed, which may include tasks such as data cleaning, feature scaling, or feature extraction. This step ensures that the data are in an appropriate format for the following quantum feature maps.
  • Feature map encoding (yellow in the figure): The preprocessed data are encoded onto qubits using a feature map.
  • Kernel computation (steel-blue in the figure): Instead of directly computing the kernel matrix from the original data, a kernel function is precomputed using the quantum computing capabilities, meaning that the inner product of two quantum states is estimated on a quantum simulator/circuit. This kernel function captures the similarity between pairs of data points in a high-dimensional feature space.
  • SVM training: The precomputed kernel function is then used as input to the SVM algorithm for model training. The SVM aims to find an optimal hyperplane that separates the data points into different classes with the maximum margin.
Here, we need to mention that in the documentation of Qiskit machine learning, the developers provided a full QKE implementation without the need to use, e.g., Scikit-learn’s SVM-implementation. However, as of the writing of this article, this estimator is no longer available in Qiskit machine learning. Thus, one needs to use a support vector machine implementation from other sources after precomputing the kernel on a quantum simulator.
Figure 2. Schematic depiction of the quantum kernel estimator. The QKE consists of several steps. We colored the steps that are similar to classical support vector machines in light blue and the other steps in yellow and steel-blue. The employed QKE algorithm consists of a support vector machine algorithm with precomputed kernel, i.e., a classical machine learning method that leverages the power of quantum computing to efficiently compute the kernel matrix.
Figure 2. Schematic depiction of the quantum kernel estimator. The QKE consists of several steps. We colored the steps that are similar to classical support vector machines in light blue and the other steps in yellow and steel-blue. The employed QKE algorithm consists of a support vector machine algorithm with precomputed kernel, i.e., a classical machine learning method that leverages the power of quantum computing to efficiently compute the kernel matrix.
Entropy 25 00992 g002

3.3.3. Qiskit Machine Learning

Qiskit Machine Learning is an open-source Python package developed by IBM for implementing quantum machine learning algorithms [24]. This package enables researchers and practitioners to develop and test quantum machine learning algorithms, including VQC and QKE, using IBM’s quantum computing platform. It provides tools for building and simulating quantum circuits, as well as interfaces to classical optimization and machine learning libraries. Thus, we used this environment and the corresponding quantum simulators described in Appendix A for our experiments.

3.4. Accuracy Score for Classification

The accuracy score is a standard metric used to evaluate the performance of classification algorithms. We employed the accuracy score to evaluate all presented experiments. It is defined as the ratio of correct predictions to the total number of predictions. The formula for the accuracy score is defined as follows:
Accuracy = Number of correct predictions Total number of predictions
In Scikit-learn, the accuracy score can be computed using the accuracy_score function from the ‘sklearn.metrics’ module [16]. For more information on the accuracy score and its interpretation, refer to the Scikit-learn documentation [16].

3.5. Datasets

In this study, we used six classification datasets from various sources. Two datasets are part of the Scikit-learn library, while the remaining four are obtained/fetched from OpenML. The datasets are described below:
  • Iris Dataset: A widely known dataset consisting of 150 samples of iris flowers, each with four features (sepal length, sepal width, petal length, and petal width) and one of three species labels (Iris Setosa, Iris Versicolor, or Iris Virginica). This dataset is included in the Scikit-learn library [16].
  • Wine Dataset: A popular dataset for wine classification, which consists of 178 samples of wine, each with 13 features (such as alcohol content, color intensity, and hue) and one of three class labels (class 1, class 2, or class 3). This dataset is also available in the Scikit-learn library [16].
  • Indian Liver Patient Dataset (LPD): This dataset contains 583 records, with 416 liver patient records and 167 non-liver patient records [32]. The dataset includes ten variables: age, gender, total bilirubin, direct bilirubin, total proteins, albumin, A/G ratio, SGPT, SGOT, and Alkphos. The primary task is to classify patients into liver or non-liver patient groups.
  • Breast Cancer Coimbra Dataset: This dataset consists of 10 quantitative predictors and a binary dependent variable, indicating the presence or absence of breast cancer [33,34]. The predictors are anthropometric data and parameters obtainable from routine blood analysis. Accurate prediction models based on these predictors can potentially serve as a biomarker for breast cancer.
  • Teaching Assistant Evaluation Dataset: This dataset includes 151 instances of teaching-assistant (TA) assignments from the Statistics Department at the University of Wisconsin-Madison, with evaluations of their teaching performance over three regular semesters and two summer semesters [35,36]. The class variable is divided into three roughly equal-sized categories (“low”, “medium”, and “high”). There are six attributes, including whether the TA is a native English speaker, the course instructor, the course, the semester type (summer or regular), and the class size.
  • Impedance Spectrum of Breast Tissue Dataset: This dataset contains impedance measurements of freshly excised breast tissue at the following frequencies: 15.625, 31.25, 62.5, 125, 250, 500, and 1000 KHz [37,38]. The primary task is to predict the classification of either the original six classes or four classes by merging the fibro-adenoma, mastopathy, and glandular classes whose discrimination is not crucial.
These datasets were selected for their diverse domains and varied classification tasks, providing a robust testing ground for the quantum classifiers we employed in our experiments. Furthermore, we used artificially generated datasets to control the number of samples. Here, Scikit-learn provides a valuable function called make_classification to generate synthetic classification datasets. This function creates a random n-class classification problem, initially creating clusters of points normally distributed about vertices of an n-informative-dimensional hypercube, and assigns an equal number of clusters to each class [16]. It introduces interdependence between features and adds further noise to the data. The generated data are highly customizable, with options for specifying the number of samples, features, informative features, redundant features, repeated features, classes, clusters per class, and more. For more details on the make_classification function and its parameters, refer to the Scikit-learn documentation available on scikit-learn.org (accessed on 25 June 2023).

3.5.1. Data Obtained from Lie-Algebras

We construct another artificial dataset final dataset for our final evaluation; however, this time, we do this by using tools from the theory of Lie groups. The reason for employing these concepts is that we want to produce data that resembles the complexity inherent to the Qubit-Vectorspace of quantum machine learning and that, furthermore, is generated by applying transformations on vectors that are similar to the manipulations present in quantum machine learning algorithms, e.g., for the VQC, rotations of/around the Bloch-sphere. Thus, overall, we aim to provide random data for a classification task to show a case where the authors assume quantum machine learning algorithms can, because of their inherent structure, outperform classical machine learning algorithms, and thus, provide a prototype on the type of data specifically tailored to address the inherent structure of quantum machine learning. The theoretical foundations of this section are obtained from [39], and thus, the interested reader is referred to this book for a profound introduction to Lie groups. To further explain the employed ideas, we start by introducing the concept of a Lie group G and the corresponding Lie-algebra g .
A Lie group is a mathematical structure that captures the essence of continuous symmetry. Named after the Norwegian mathematician Sophus Lie, Lie groups are ubiquitous in many areas of mathematics and physics, including the study of differential equations, geometry, and quantum mechanics.
A Lie group is a set G that has the structure of both a smooth manifold and a group in such a way that the group operations (multiplication and inversion) are smooth. That is, a Lie group is a group that is also a differentiable manifold, such that the group operations are compatible with the smooth structure.
Thus, a Lie group is a set G equipped with a group structure (i.e., a binary operation G × G G , ( g , h ) g h that is associative, an identity element e G , and an inversion operation G G , g g 1 ) and a smooth manifold structure such that the following conditions are satisfied:
  • The multiplication map μ : G × G G defined by μ ( g , h ) = g h is smooth.
  • The inversion map ι : G G defined by ι ( g ) = g 1 is smooth.
Lie algebra is associated with each Lie group, a vector space equipped with a binary operation called the Lie bracket. The Lie algebra captures the local structure of the Lie group near the identity element, meaning that the Lie algebra of a Lie group G is the tangent space at the identity, denoted T e G , equipped with the Lie bracket operation. The Lie bracket is defined in terms of the group operation and the differential.
There is a map from the Lie algebra to the Lie group called the exponential map, denoted exp : T e G G . The exponential map provides a way to generate new group elements from elements of the Lie algebra. In particular, given an element X of the Lie algebra, exp ( X ) is a group element close to the identity if X is ‘small’. We will exploit this concept to generate random data associated with a specific group:
We start with a set of generators T a contained within the Lie-algebra g of a Lie group G, where a = 1 , 2 , d g , i.e., the dimension of the Lie-algebra. We can then create elements g G by employing:
g = e i a θ a T a , where θ a 0 , 2 π .
We used the condition for our θ a -values without loss of generality due to the periodicity of the exponential function. To generate our random data, we randomly choose our θ a and create an element of our group. We then apply this element to a corresponding base vector of our vector space.
Specifically, in our example, we use the Lie-group SU(2). The special unitary group of degree 2, denoted as SU(2), is a Lie group of 2 × 2 unitary matrices with determinant 1.
SU ( 2 ) = U C 2 × 2 : U U = I , det ( U ) = 1
The corresponding Lie algebra, su ( 2 ) , consists of 2 × 2 Hermitian traceless matrices, i.e., the Pauli matrices:
σ 1 = 0 1 1 0 , σ 2 = 0 i i 0 , σ 3 = 1 0 0 1
The commutation relations of the Pauli matrices form the structure of the su ( 2 ) Lie algebra:
[ σ i , σ j ] = 2 i ε i j k σ k
where [ · , · ] denotes the commutator and ε i j k is the Levi-Civita symbol.
To generate a classification dataset from this algebra, we use the following procedure:
  • Find a set of random parameters θ 0 ; π , ϕ 0 ; 2 π , λ 0 ; 2 π ;
  • 2.
    We then create an element U of SU(2) using these these randomly set parameters: U = e i θ σ 1 + ϕ σ 2 + λ σ 3 ;
    3.
    Next, we take one of the basevectors from C 2 , denoted as v ^ to create a new complex vector v using the previously obtained matrix U such that: v = U · v ^ ;
    4.
    This vector is then separated into four features F j such that:
    F 1 = Re v 1
    F 2 = Im v 1
    F 3 = Re v 2
    F 4 = Im v 2 ,
    where v 1 and v 2 denotes the individual components of the vector v , and Re [ ] and Im [ ] denote their respective real and imaginary parts;
    5.
    Finally, we assign a class label C to this collection of features such that:
    C = 0 if θ < π 2 1 if θ > π 2 ,
    and collect the features and the class label into one sample F 1 , F 2 , F 3 , F 4 , C . We repeat this process N S times, starting with 1, where N S is the number of samples that we want for our dataset.
    Note that this approach can be extended to arbitrary Lie groups, given that one can construct or obtain a Lie group’s generators.

    4. Experimental Design

    In this section, we describe our experimental design, which aims to provide a fair and comprehensive comparison of the performance of classical machine learning (ML) and quantum machine learning techniques, as discussed in Section 3.2 and Section 3.3. Our experiments involve two main components: Firstly, assessing the algorithms’ performance on artificially generated datasets with varying parametrizations, and secondly, evaluating the algorithms’ performance on benchmark datasets using randomized search to optimize hyperparameters, ensuring a fair comparison. By carefully selecting our experimental setup, we avoid the issue of “cherry-picking” only a favorable subset of results, a common problem in machine learning, leading to heavily biased conclusions.

    4.1. Artificially Generated Scikit Datasets

    To generate the synthetic classification dataset, we utilized Scikit-learn’s make_classification function. We employed two features and two classes while varying the number of samples to obtain a performance curve illustrating how the chosen algorithms’ performance changes depending on the sample size.
    We partitioned each dataset such that 20% of the original data were reserved as a test set to evaluate the trained algorithm, producing the accuracy score used for our assessment. Furthermore, each dataset was normalized such that all features are within the unit interval 0 , 1 .
    As a baseline, we employed the seven classical machine learning algorithms described in Section 3.2, namely Lasso, Ridge, MLP, SVM, XGBoost, LightGBM, and CatBoost. We used two different parameterizations for the classical machine learning algorithms for our comparisons. Firstly, we applied the out-of-the-box implementation without any hyperparameter optimization. Secondly, we used an optimized version of each algorithm found through Scikit-learn’s RandomizedSearchCV by testing 20 different models.
    We then examined 20 distinct parameter configurations, each for the VQC and QKE classifiers, randomly selected from a predefined parameter distribution. Appendix A discusses the parameter grids for all utilized algorithms and all experiments.

    4.2. Artificially Generated SU(2) Datasets

    For our synthetic SU ( 2 ) classification dataset, we used the concepts previously discussed in Section 3.5.1. We employed two complex features, i.e., resulting in four continuous real features, and two classes while varying the number of samples to obtain a performance curve illustrating how the chosen algorithms’ performance changes depending on the sample size.
    We partitioned each dataset such that 20% of the original data were reserved as a test set to evaluate the trained algorithm, producing the accuracy score used for our assessment. Furthermore, each dataset was normalized such that all features are within the unit interval 0 , 1 .
    As a baseline, we employed the seven classical machine learning algorithms described in Section 3.2, namely Lasso, Ridge, MLP, SVM, XGBoost, LightGBM, and CatBoost. We used two different parameterizations for the classical machine learning algorithms for our comparisons. Firstly, we applied the out-of-the-box implementation without any hyperparameter optimization. Secondly, we used an optimized version of each algorithm found through Scikit-learn’s RandomizedSearchCV by testing 20 different models.
    We then examined 20 distinct parameter configurations, each for the VQC and QKE classifiers, randomly selected from a predefined parameter distribution. Appendix A discusses the parameter grids for all utilized algorithms and all experiments.

    4.3. Benchmark Datasets and Hyperparameter Optimization

    Our last experiment was to test the two employed quantum machine learning algorithms against the classical machine learning algorithms on six benchmark datasets (Section 3.5). For this reason, we employed Scikit-learn’s RandomizedSearchCV to test 20 randomly parameterized models for each algorithm to report the best of these tests. Again, we used a train-test-split to keep 20% of the original data to test the trained algorithm. Furthermore, each dataset was normalized such that all features are within the unit interval  0 , 1 .

    5. Results

    In this section, we present the results of our experiments, comparing the performance of classical machine learning and quantum machine learning techniques on both artificially generated datasets and benchmark datasets (Section 3.5). By analyzing the results, we aim to draw meaningful insights into the strengths and weaknesses of each approach and provide a blueprint for future studies in the area. Everything was calculated on a Lenovo ThinkCentre machine using an Intel(R) Core(TM) i7-4770 CPU 3.40GHz and 16GB RAM and Linux 20.04. We used python 3.6 and the included packages are the following:
    • numpy version: 1.18.5
    • sklearn version: 0.23.1
    • catboost version: 0.26.1
    • xgboost version: 1.2.1
    • lightgbm version: 3.2.1
    • qiskit version: {‘qiskit-terra’: ‘0.19.2’, ‘qiskit-aer’: ‘0.10.3’, ‘qiskit-ignis’: ‘0.7.0’, ‘qiskit-ibmq-provider’: ‘0.18.3’, ‘qiskit-aqua’: None, ‘qiskit’: ‘0.34.2’, ‘qiskit-nature’: ‘0.3.1’, ‘qiskit-finance’: None, ‘qiskit-optimization’: None, ‘qiskit-machine-learning’: ‘0.3.1’}
    • qiskit_machine_learning version: 0.3.1

    5.1. Performance on Artificially Generated Scikit Datasets

    In this section, we compare the performance of quantum machine learning algorithms and classical machine learning algorithms on artificially generated classification datasets. The comprehensive experimental setup can be found in Section 4.1.
    Regarding accuracy and runtime, our findings are presented in Table 1 and Table 2 and Figure 3, Figure 4 and Figure 5. The measured runtime includes hyperparameter tuning via randomized search and five-fold cross-validating, training, and testing the model.
    While QML algorithms perform reasonably well, we observe that they are not a match for properly trained and/or sophisticated state-of-the-art classifiers. Even out-of-the-box implementations of state-of-the-art ML algorithms outperform QML algorithms on these artificially generated classification datasets.
    The accuracy of the algorithms varies depending on the dataset size, with larger datasets posing more challenges. CatBoost performed best in our experiments, both out-of-the-box and when optimized in terms of high accuracy over all experiments. The quantum kernel estimator is the fifth-best algorithm overall in terms of accuracy, though it outperforms CatBoost regarding the runtime for CatBoost’s optimized version. XGBoost and support vector classification (SVC) follow closely, with competitive performances in terms of accuracy. However, variational quantum circuit struggles to achieve high accuracy compared to sophisticated boosting classifiers or support vector machines. Furthermore, we observe the best performance in terms of runtime for the two linear models, Lasso and Ridge. We need to point out that Lasso and Ridge both feature increased runtimes for the datasets of size 50; this is most likely due to the optimizer needing an increased number of iterations due to the small number of samples and their relatively scattered distribution of data points.
    Other algorithms, such as multilayer perceptron, Ridge regression, Lasso regression, and LightGBM, exhibit varying performances depending on dataset size and optimization. Despite some reasonable results from QKE, we conclude that classical ML algorithms, particularly sophisticated boosting classifiers, should be chosen to tackle similar problems due to their ease of implementation, better runtime, and overall superior performance.
    In summary, while QML algorithms have shown some promise, they cannot yet compete with state-of-the-art classical ML algorithms on artificially generated classification datasets in terms of accuracy and runtime.

    5.2. Performance on Artificially Generated SU2 Datasets

    In this section, we compare the performance of quantum machine learning algorithms and classical machine learning algorithms on artificially generated classification datasets based on Lie group structures. The detailed experimental setup can be found in Section 4.2.
    Regarding accuracy and runtime, our findings are presented in Table 3 and Table 4 and Figure 6, Figure 7 and Figure 8. While QML algorithms perform reasonably well, we observe that they are not a match for properly trained and/or sophisticated state-of-the-art classifiers. Even out-of-the-box implementations of state-of-the-art ML algorithms outperform QML algorithms on artificially generated classification datasets that are particularly suited for QML.
    The accuracy of the algorithms varies depending on the dataset size, with larger datasets providing increased accuracy for most algorithms. CatBoost performed best in our experiments, both out-of-the-box and when optimized in terms of high accuracy over all experiments. The quantum kernel estimator is the fifth-best algorithm overall in terms of accuracy. However, we observe that, on average, CatBoost with improved hyperparameters performs best over all experiments, but is outperformed by the best QKE implementation for 100 and 500 data points. Thus, we conclude that quantum kernel estimators can capture the complexity of this SU ( 2 ) -generated dataset, but overall, one is better off with an out-of-the-box CatBoost implementation. This means that we do not observe a quantum advantage for this type of data, but rather that the employed quantum kernel estimator behaves similarly to classical machine learning algorithms, i.e., it exhibits reasonable performance but does not perform best for all datasets, even the ones created by exploiting quantum symmetry properties.
    Other algorithms, such as multilayer perceptron, Ridge regression, Lasso regression, and LightGBM, exhibit varying performances depending on dataset size and optimization. Despite some reasonable results from QKE, we conclude that classical ML algorithms, particularly sophisticated boosting classifiers, should be chosen to tackle similar problems due to their ease of implementation, better runtime, and overall superior performance. Furthermore, we again observe the best performance in terms of runtime for the two linear models, Lasso and Ridge. Moreover, again, we observe that Lasso and Ridge both feature increased runtimes for the datasets of size 50.
    In summary, while QML algorithms have shown some promise, they cannot yet compete with state-of-the-art classical ML algorithms even on these SU ( 2 ) -datasets, where the authors intended to provide evidence for the quantum advantage for datasets generated from symmetry properties inherent to quantum mechanics.

    5.3. Results on Benchmark Datasets

    In this section, we discuss the performance of quantum machine learning and classical machine learning algorithms on six benchmark datasets described in Section 3.5. We include results for the quantum classifiers detailed in Section 3.3 and the classical machine learning classifiers discussed in Section 3.2. The scores/accuracies were obtained using randomized search cross-validation from Scikit-learn with 20 models and five-fold cross-validation.
    Our results, shown in Table 5, display the best five-fold cross-validation scores (upper table) and the scores of the best model evaluated on an unseen test subset of the original data (lower table), which makes up 20% of the original data. We observe varying performances of the algorithms on these benchmark datasets.
    Notably, both the variational quantum circuit and the quantum kernel estimator classifier show competitive performance on several datasets but do not consistently outperform classical ML algorithms. In particular, QKE achieves a perfect score on the Iris dataset, but its performance varies across the other datasets.
    Classical ML algorithms, such as multilayer perceptron, support vector machines, XGBoost, LightGBM, and CatBoost, exhibit strong performance across all datasets, with some algorithms achieving perfect scores on multiple datasets. CatBoost consistently performs well, ranking as the top-performing algorithm on three of the six datasets. Ridge and Lasso regression show high accuracy on Iris and Wine datasets but perform poorly on the others.
    When comparing the runtimes of the experiments, as presented in Table 6, it becomes evident that QML algorithms take substantially longer to execute than their classical counterparts. For instance, the VQC and QKE classifiers take hours to days to complete on various datasets, whereas classical ML algorithms such as Ridge, Lasso, MLP, SVM, XGBoost, LightGBM, and CatBoost typically take seconds to minutes.
    This significant difference in runtimes could be attributed to the inherent complexity and resource requirements of QML algorithms, which generally demand specialized quantum hardware and simulators. On the other hand, classical ML algorithms are optimized for execution on conventional hardware, making them more efficient and faster to run.
    In conclusion, while QML algorithms such as VQC and QKE demonstrate potential in achieving competitive performance on certain datasets, their relatively longer runtimes and less consistent performance across the benchmark datasets may limit their practical applicability compared to classical ML algorithms. Classical ML algorithms, such as CatBoost, XGBoost, and LightGBM, continue to offer superior and more consistent performance with faster execution times, solidifying their place as reliable and powerful tools for classification tasks.

    5.4. Comparison and Discussion

    In this study, we have compared the performance of quantum machine learning and classical machine learning algorithms on six benchmark datasets and two types of artificially generated classification datasets. We included results for quantum classifiers, such as variational quantum circuit and quantum kernel estimator, and classical machine learning classifiers, such as CatBoost, XGBoost, and LightGBM. Our experiments showed that while QML algorithms demonstrate potential in achieving competitive performance on certain datasets, they do not consistently outperform classical ML algorithms. Additionally, their longer runtimes for the whole process, i.e., hyperparameter tuning via randomized search and five-fold cross-validation, the corresponding training and testing, and less consistent performance across the benchmark datasets, may limit their practical applicability compared to classical ML algorithms, which continue to offer superior and more consistent performance with faster execution times. Furthermore, we constructed artificial datasets with the structure and rulings of quantum Mechanics in mind, i.e., we used symmetry properties and unitary transformations to generate a classification dataset from SU ( 2 ) -matrices in order to demonstrate an advantage of quantum machine learning algorithms to tackle problems with an inherent structure relatable to that of quantum circuits and quantum mechanics overall. However, also for these datasets, the employed quantum machine learning algorithms performed reasonably but did not outperform sophisticated boost classifiers. Thus, we cannot conclude a quantum advantage for these datasets.
    It is essential to highlight that the QML algorithms’ performance in our experiments was based on simulated quantum infrastructures. This is a significant limitation to consider, as the specific constraints and characteristics of the simulated hardware may influence the performance of these algorithms. Furthermore, given the rapid advancement of quantum technologies and hardware, this constraint might be obsolete in the near future.
    The impact of quantum simulators, feature maps, and quantum circuits on the performance of quantum estimators stems from the fact that these components play crucial roles in shaping the behavior and capabilities of quantum machine learning algorithms. Quantum simulators, which emulate quantum systems on classical computers, introduce various levels of approximation and noise, leading to deviations from ideal quantum behavior. Different simulators may employ distinct algorithms and techniques, resulting in variations in performance.
    Feature maps, responsible for encoding classical data into quantum states, determine how effectively the quantum system can capture and process information. The choice of feature map can greatly influence the ability of quantum algorithms to extract meaningful features and represent the data in a quantum-mechanical space.
    Similarly, quantum circuits, composed of quantum gates and operations, define the computational steps performed on the encoded data. Different circuit designs and configurations can affect the expressiveness and depth of the quantum computation, potentially impacting the accuracy and efficiency of the quantum estimators.
    Considering the diverse options for quantum simulators, feature maps, and quantum circuits, it becomes essential for researchers to provide detailed explanations of their hyperparameter choices. This entails clarifying the rationale behind selecting a specific simulator, feature map, or circuit design, as well as the associated parameters and their values. By providing such explanations, researchers can enhance the reproducibility and comparability of results, enabling the scientific community to better understand the strengths and limitations of different quantum machine learning algorithms.
    Unfortunately, the current state of the field often overlooks the thorough discussion of hyperparameter choices in many studies. This omission restricts the transparency and interpretability of research outcomes and hinders the advancement of quantum machine learning. To address this issue, researchers should embrace a culture of providing comprehensive documentation regarding hyperparameter selection, sharing insights into the decision-making process, and discussing the potential implications of different choices.
    By encouraging researchers to provide detailed explanations of hyperparameter choices and corresponding code, we can foster a more robust and transparent research environment in quantum machine learning. This approach enables the replication and comparison of results, promotes knowledge sharing, and ultimately contributes to the development of reliable and effective quantum machine learning algorithms. Additionally, our program code serves as introductory material, providing easy-to-use implementations and a foundation for comparing quantum machine learning and classical machine learning (CML) algorithms.
    One possible direction for future research is exploring quantum ensemble classifiers and, consequently, quantum boosting classifiers, as suggested by Schuld et al. [40]. This approach might help in improving the capabilities of QML algorithms and make them more competitive with state-of-the-art classical ML algorithms in terms of high accuracies.
    Finally, the relatively lower performance of the employed quantum machine learning algorithms compared to, for example, the employed boosting classifiers might be attributed to quantum machine learning, being constrained by specific rules of quantum mechanics.
    In the authors’ opinion, quantum machine learning might be constrained by the unitary transformations inherent in, for example, the variational quantum circuits. These transformations are part of the unitary group U ( n ) . Thus, all transformations are constrained by symmetry properties. Classical machine learning models are not constrained by these limitations, meaning that, for instance, different activation functions in neural networks do not preserve certain distance metrics or probabilities when processing data. However, expanding the set of transformations of quantum machine learning and getting rid of possible constraints might improve the capabilities of quantum machine learning models such that these algorithms might be better capable of capturing the information of more complex data. However, this needs to be discussed in the context of quantum computers such that one determines what all possible transformations on a quantum computer are. This means that future research needs to consider the applicability of advanced mathematical frameworks for quantum machine learning regarding the formal requirements of quantum computers.
    Furthermore, another constraint of quantum machine learning is that it, and quantum mechanics in general, relies on Hermitian matrices, e.g., to provide real-valued eigenvalues of observables. However, breaking this constraint might be another way to broaden the capabilities of quantum machine learning to better capture complexity, e.g., by using non-Hermitian kernels in a quantum kernel estimator. Here, we want to mention the book by Moiseyev [41], which introduces non-Hermitian quantum mechanics. Furthermore, quantum computers, in general, might provide a testing ground for non-Hermitian quantum mechanics in comparison to Hermitian quantum mechanics. However, at this point, this is rather speculative, but given that natural data are nearly always corrupted by noise and symmetries are never truly perfect in nature, breaking constraints and symmetries might be ideas to expand the capabilities of QML.

    6. Conclusions

    In this research, we have explored the applicability of quantum machine learning for classification tasks by examining the performance of variational quantum circuit and quantum kernel estimator algorithms. Our comparison of these quantum classifiers with classical machine learning algorithms, such as XGBoost, Ridge, Lasso, LightGBM, CatBoost, and MLP, on six benchmark datasets and artificially generated classification datasets demonstrated that QML algorithms can achieve competitive performance on certain datasets. However, they do not consistently outperform their classical ML counterparts, particularly with regard to runtime performance and accuracy. Quite the contrary, classical machine learning algorithms still demonstrate superior performance, especially in terms of increased accuracy, in most of our experiments. Furthermore, we cannot conclude a quantum advantage even for artificial data built by data manipulations inherent to quantum mechanics.
    As our study’s performance comparison relied on simulated quantum circuits, it is important to consider the limitations and characteristics of simulated hardware, which may affect the true potential of quantum machine learning. Given the rapid advancement of quantum technologies and hardware, these constraints may become less relevant in the future.
    Quantum simulators, feature maps, and quantum circuits significantly influence quantum estimator performance; hence, a detailed discussion of the chosen hyperparameters is essential. The absence of such a discussion in current research limits the interpretation and replication of experiments. Thus, we aim to encourage transparency in decision-making processes to promote a robust research environment, aiding in knowledge sharing and the creation of reliable quantum machine learning algorithms.
    Despite the current limitations, this study has shed light on the potential and challenges of quantum machine learning compared to classical approaches. Thus, by providing our complete code in a GitHub repository, we hope to foster transparency, encourage further research in this field, and offer a foundation for other researchers to build upon as they explore the world of quantum machine learning. Furthermore, the developed SU ( 2 ) -data creation might serve as a quantum data prototype for future experiments, and both quantum and regular machine learning algorithms can be tested for their accuracy on datasets like these.
    Future research should also consider exploring quantum ensemble classifiers and quantum boosting classifiers, as well as addressing the limitations imposed by the specific rules of quantum mechanics. By breaking constraints and symmetries and expanding the set of transformations in quantum machine learning, researchers may be able to unlock its full potential.

    Author Contributions

    Conceptualization, S.R. and K.M.; Methodology, S.R. and K.M.; Software, S.R. and K.M.; Validation, S.R. and K.M.; Formal analysis, S.R. and K.M.; Investigation, S.R. and K.M.; Resources, K.M.; Data curation, S.R. and K.M.; Writing—original draft, S.R. and Kevin Mallinger; Writing—review & editing, S.R. and K.M.; Visualization, S.R. and K.M. All authors have read and agreed to the published version of the manuscript.

    Funding

    Open Access Funding by TU Wien.

    Acknowledgments

    The authors acknowledge the funding by the TU Wien Bibliothek for financial support through its Open Access Funding Program.

    Conflicts of Interest

    The authors declare no conflict of interest.

    Appendix A. Parametrization

    This Appendix lists the parameter grids for all employed algorithms per the implementations from Scikit-learn and Qiskit [16,24]. Thus, for further explanations on the parameters and how they influence the discussed algorithm, the reader is referred to the respective sources, which we linked in Section 3.2 and Section 3.3.

    Appendix A.1. Ridge

    param_grid = {
    ’alpha’: [0.001, 0.01, 0.1, 1, 10, 100],
    ’fit_intercept’: [True, False],
    ’normalize’: [True, False],
    ’copy_X’: [True, False],
    ’max_iter’: [100, 500, 1000],
    ’tol’: [1e-4, 1e-3, 1e-2],
    ’solver’: [’auto’, ’svd’, ’cholesky’, ’lsqr’, ’sparse_cg’, ’sag’, ’saga’],
    ’random_state’: [42]
    }

    Appendix A.2. Lasso

    param_grid = {
    ’alpha’: [0.001, 0.01, 0.1, 1, 10, 100],
    ’fit_intercept’: [True, False],
    ’normalize’: [True, False],
    ’precompute’: [True, False],
    ’copy_X’: [True, False],
    ’max_iter’: [100, 500, 1000],
    ’tol’: [1e-4, 1e-3, 1e-2],
    ’warm_start’: [True, False],
    ’positive’: [True, False],
    ’random_state’: [42],
    ’selection’: [’cyclic’, ’random’]
    }

    Appendix A.3. SVM

    param_grid = {
    ’C’: [0.1, 1, 10, 100],
    ’kernel’: [’linear’, ’poly’, ’rbf’, ’sigmoid’],
    ’degree’: [2, 3, 4],
    ’gamma’: [’scale’, ’auto’],
    ’coef0’: [0.0, 1.0, 2.0],
    ’shrinking’: [True, False],
    ’probability’: [False],
    ’tol’: [1e-4, 1e-3, 1e-2],
    ’cache_size’: [200],
    ’class_weight’: [None, ’balanced’],
    ’verbose’: [False],
    ’max_iter’: [200, 300, 400],
    ’decision_function_shape’: [’ovr’, ’ovo’],
    ’break_ties’: [False],
    ’random_state’: [42]
    }

    Appendix A.4. MLP

    param_grid = {
    ’hidden_layer_sizes’: [(50,), (100,), (150,)],
    ’activation’: [’relu’, ’tanh’],
    ’solver’: [’adam’, ’sgd’],
    ’alpha’: [0.0001, 0.001, 0.01],
    ’learning_rate’: [’constant’, ’invscaling’, ’adaptive’],
    ’max_iter’: [200, 300, 400]
    }

    Appendix A.5. XGBoost

    param_grid = {
    ’max_depth’: [3, 5, 7, 10],
    ’learning_rate’: [0.01, 0.05, 0.1, 0.2],
    ’n_estimators’: [50, 100, 150, 200],
    ’subsample’: [0.5, 0.8, 1],
    ’colsample_bytree’: [0.5, 0.8, 1]
    }

    Appendix A.6. LightGBM

    param_grid = {
    ’max_depth’: [3, 5, 7, 10],
    ’learning_rate’: [0.01, 0.05, 0.1, 0.2],
    ’n_estimators’: [50, 100, 150, 200],
    ’subsample’: [0.5, 0.8, 1],
    ’colsample_bytree’: [0.5, 0.8, 1]
    }

    Appendix A.7. CatBoost

    param_grid = {
    ’iterations’: [50, 100, 150, 200],
    ’learning_rate’: [0.01, 0.05, 0.1, 0.2],
    ’depth’: [3, 5, 7, 10],
    ’l2_leaf_reg’: [1, 3, 5, 7, 9],
    }

    Appendix A.8. QKE

    For this Algorithm, we precomputed the kernel matrix using Qiskit and then performed the support vector classification via the vanilla SVM algorithm from Scikit-learn.
    param_grid = {
    ’feature_map’: [PauliFeatureMap, ZFeatureMap, ZZFeatureMap],
    ’quantum_instance’: [
    QuantumInstance(Aer.get_backend(’aer_simulator’), shots=1024),
    QuantumInstance(Aer.get_backend(’qasm_simulator’), shots=1024),
    QuantumInstance(Aer.get_backend(’statevector_simulator’), shots=1024)
    ],
    ’C’ : np.logspace(-3, 3, 9),
    }

    Appendix A.9. VQC

    param_grid = {
    ’feature_map’: [PauliFeatureMap, ZFeatureMap, ZZFeatureMap],
    ’ansatz’: [EfficientSU2, TwoLocal, RealAmplitudes],
    ’optimizer’: [
    COBYLA(maxiter=max_iter),
    SPSA(maxiter=max_iter),
    NFT(maxiter=max_iter),
    ],
    ’quantum_instance’: [
    QuantumInstance(Aer.get_backend(’aer_simulator’), shots=1024),
    QuantumInstance(Aer.get_backend(’qasm_simulator’), shots=1024),
    QuantumInstance(Aer.get_backend(’statevector_simulator’), shots=1024)
    ],
    }

    References

    1. Nielsen, M.A.; Chuang, I.L. Quantum Computation and Quantum Information: 10th Anniversary Edition, 10th ed.; Cambridge University Press: Cambridge, MA, USA, 2011. [Google Scholar]
    2. Biamonte, J.; Wittek, P.; Pancotti, N.; Rebentrost, P.; Wiebe, N.; Lloyd, S. Quantum machine learning. Nature 2017, 549, 195–202. [Google Scholar] [CrossRef] [PubMed] [Green Version]
    3. Schuld, M.; Sinayskiy, I.; Petruccione, F. An introduction to quantum machine learning. Contemp. Phys. 2015, 56, 172–185. [Google Scholar] [CrossRef] [Green Version]
    4. Havlíček, V.; Córcoles, A.D.; Temme, K.; Harrow, A.W.; Kandala, A.; Chow, J.M.; Gambetta, J.M. Supervised learning with quantum-enhanced feature spaces. Nature 2019, 567, 209–212. [Google Scholar] [CrossRef] [PubMed] [Green Version]
    5. Cerezo, M.; Arrasmith, A.; Babbush, R.; Benjamin, S.C.; Endo, S.; Fujii, K.; McClean, J.R.; Mitarai, K.; Yuan, X.; Cincio, L.; et al. Variational quantum algorithms. Nat. Rev. Phys. 2021, 3, 625–644. [Google Scholar] [CrossRef]
    6. Griol-Barres, I.; Milla, S.; Cebrián, A.; Mansoori, Y.; Millet, J. Variational Quantum Circuits for Machine Learning. An Application for the Detection of Weak Signals. Appl. Sci. 2021, 11, 6427. [Google Scholar] [CrossRef]
    7. Kuppusamy, P.; Yaswanth Kumar, N.; Dontireddy, J.; Iwendi, C. Quantum Computing and Quantum Machine Learning Classification—A Survey. In Proceedings of the 2022 IEEE 4th International Conference on Cybernetics, Cognition and Machine Learning Applications (ICCCMLA), Goa, India, 8–9 October 2022; pp. 200–204. [Google Scholar] [CrossRef]
    8. Blance, A.; Spannowsky, M. Quantum machine learning for particle physics using a variational quantum classifier. J. High Energy Phys. 2021, 2021, 212. [Google Scholar] [CrossRef]
    9. Abohashima, Z.; Elhoseny, M.; Houssein, E.H.; Mohamed, W.M. Classification with Quantum Machine Learning: A Survey. arXiv 2020, arXiv:2006.12270. [Google Scholar]
    10. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
    11. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
    12. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
    13. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 3149–3157. [Google Scholar]
    14. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased Boosting with Categorical Features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, Montréal, ON, Canada, 2–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 6639–6649. [Google Scholar]
    15. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation; Technical report; California Univ San Diego La Jolla Inst for Cognitive Science: San Diego, CA, USA, 1985. [Google Scholar]
    16. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 282–290. [Google Scholar]
    17. Raubitzek, S. Quantum_Machine_Learning. Preprints 2023, 2023050833. [Google Scholar] [CrossRef]
    18. Zeguendry, A.; Jarir, Z.; Quafafou, M. Quantum Machine Learning: A Review and Case Studies. Entropy 2023, 25, 287. [Google Scholar] [CrossRef]
    19. Mitarai, K.; Negoro, M.; Kitagawa, M.; Fujii, K. Quantum circuit learning. Phys. Rev. A 2018, 98, 032309. [Google Scholar] [CrossRef] [Green Version]
    20. Farhi, E.; Neven, H. Classification with quantum neural networks on near term processors. arXiv 2018, arXiv:1802.06002. [Google Scholar]
    21. Rebentrost, P.; Mohseni, M.; Lloyd, S. Quantum support vector machine for big data classification. Phys. Rev. Lett. 2014, 113, 130503. [Google Scholar] [CrossRef] [PubMed] [Green Version]
    22. Liu, D.; Rebentrost, P. Quantum machine learning for quantum anomaly detection. Phys. Rev. A 2019, 100, 042328. [Google Scholar] [CrossRef] [Green Version]
    23. Broughton, M.; Verdon, G.; McCourt, T.; Martinez, A.J.; Yoo, J.H.; Isakov, S.V.; King, A.D.; Smelyanskiy, V.N.; Neven, H. TensorFlow Quantum: A Software Framework for Quantum Machine Learning. arXiv 2020, arXiv:2003.02989. [Google Scholar]
    24. Qiskit Contributors. Qiskit: An Open-source Framework for Quantum Computing. Zenodo 2023. [Google Scholar] [CrossRef]
    25. Bishop, C.M. Pattern Recognition and Machine Learning (Information Science and Statistics); Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
    26. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2013. [Google Scholar]
    27. Kotsiantis, S.B. Supervised Machine Learning: A Review of Classification Techniques. In Proceedings of the 2007 Conference on Emerging Artificial Intelligence Applications in Computer Engineering: Real Word AI Systems with Applications in EHealth, HCI, Information Retrieval and Pervasive Technologies; IOS Press: Amsterdam, The Netherlands, 2007; pp. 3–24. [Google Scholar]
    28. Refaeilzadeh, P.; Tang, L.; Liu, H. Cross-Validation. In Encyclopedia of Database Systems; Liu, L., ÖZSU, M.T., Eds.; Springer: Boston, MA, USA, 2009; pp. 532–538. [Google Scholar] [CrossRef]
    29. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
    30. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
    31. Schuld, M.; Killoran, N. Quantum Machine Learning in Feature Hilbert Spaces. Phys. Rev. Lett. 2019, 122, 040504. [Google Scholar] [CrossRef] [Green Version]
    32. Ramana, B.V.; Babu, M.S.P.; Venkateswarlu, N.B. LPD (Indian Liver Patient Dataset) Data Set. 2012. Available online: https://archive.ics.uci.edu/ml/datasets/ILPD+(Indian+Liver+Patient+Dataset) (accessed on 25 June 2023).
    33. Patrício, M.; Pereira, J.; Crisóstomo, J.; Matafome, P.; Gomes, M.; Seiça, R.; Caramelo, F. Using Resistin, glucose, age and BMI to predict the presence of breast cancer. BMC Cancer 2018, 18, 29. [Google Scholar] [CrossRef] [PubMed] [Green Version]
    34. Crisóstomo, J.; Matafome, P.; Santos-Silva, D.; Gomes, A.L.; Gomes, M.; Patrício, M.; Letra, L.; Sarmento-Ribeiro, A.B.; Santos, L.; Seiça, R. Hyperresistinemia and metabolic dysregulation: A risky crosstalk in obese breast cancer. Endocrine 2016, 53, 433–442. [Google Scholar] [CrossRef] [PubMed]
    35. Loh, W.Y.; Shih, Y.S. Split selection methods for classification trees. Stat. Sin. 1997, 7, 815–840. [Google Scholar]
    36. Lim, T.S.; Loh, W.Y.; Shih, Y.S. A comparison of prediction accuracy, complexity, and training time of thirty-three old and new classification algorithms. Mach. Learn. 2000, 40, 203–228. [Google Scholar] [CrossRef]
    37. Marques de Sá, J.; Jossinet, J. Breast Tissue Impedance Data Set. 2002. Available online: https://archive.ics.uci.edu/ml/datasets/Breast+Tissue (accessed on 25 June 2023).
    38. Estrela da Silva, J.; Marques de Sá, J.P.; Jossinet, J. Classification of breast tissue by electrical impedance spectroscopy. Med. Biol. Eng. Comput. 2000, 38, 26–30. [Google Scholar] [CrossRef] [PubMed]
    39. Georgi, H. Lie Algebras in Particle Physics: From Isospin to Unified Theories; CRC Press: Boca Raton, FL, USA, 2019. [Google Scholar] [CrossRef]
    40. Schuld, M.; Petruccione, F. Quantum ensembles of quantum classifiers. Sci. Rep. 2018, 8, 2772. [Google Scholar] [CrossRef] [Green Version]
    41. Moiseyev, N. Non-Hermitian Quantum Mechanics; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar] [CrossRef]
    Figure 1. Schematic depiction of the variational quantum circuit. The VQC consists of several steps. We colored the steps that are similar to classical neural networks in light blue and the other steps in yellow, steel-blue, and orange.
    Figure 1. Schematic depiction of the variational quantum circuit. The VQC consists of several steps. We colored the steps that are similar to classical neural networks in light blue and the other steps in yellow, steel-blue, and orange.
    Entropy 25 00992 g001
    Figure 3. These figures depict the results from our experiments, comparing the five best QML and classical ML algorithms on artificially generated datasets in terms of accuracy. The upper part illustrates the accuracy of the algorithms on different sample sizes, while the lower part demonstrates how the runtimes change with increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm.
    Figure 3. These figures depict the results from our experiments, comparing the five best QML and classical ML algorithms on artificially generated datasets in terms of accuracy. The upper part illustrates the accuracy of the algorithms on different sample sizes, while the lower part demonstrates how the runtimes change with increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm.
    Entropy 25 00992 g003
    Figure 4. These figures depict the results from our experiments, comparing differently parameterized classical machine learning algorithms on artificially generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the run times change with the increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms.
    Figure 4. These figures depict the results from our experiments, comparing differently parameterized classical machine learning algorithms on artificially generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the run times change with the increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms.
    Entropy 25 00992 g004
    Figure 5. These figures depict the results from our experiments for the artificially generated datasets, comparing differently parameterized QML algorithms on artificially generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the runtimes change with the increasing size of the test datasets. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator.
    Figure 5. These figures depict the results from our experiments for the artificially generated datasets, comparing differently parameterized QML algorithms on artificially generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the runtimes change with the increasing size of the test datasets. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator.
    Entropy 25 00992 g005
    Figure 6. These figures depict the results from our experiments, comparing the five best QML and classical ML algorithms in terms of accuracy on datasets using the exponential map to create SU ( 2 ) -transformations on complex vectors. The upper part illustrates the accuracy of the algorithms on different sample sizes, while the lower part demonstrates how the runtimes change with the increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and, more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm.
    Figure 6. These figures depict the results from our experiments, comparing the five best QML and classical ML algorithms in terms of accuracy on datasets using the exponential map to create SU ( 2 ) -transformations on complex vectors. The upper part illustrates the accuracy of the algorithms on different sample sizes, while the lower part demonstrates how the runtimes change with the increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and, more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm.
    Entropy 25 00992 g006
    Figure 7. These figures depict the results from our experiments, comparing differently parameterized classical machine learning algorithms on the SU ( 2 ) -generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the run times change with the increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms.
    Figure 7. These figures depict the results from our experiments, comparing differently parameterized classical machine learning algorithms on the SU ( 2 ) -generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the run times change with the increasing size of the test dataset. The right part contains the legend, indicating which algorithms were used, and more specifically, the different parametrizations of the employed machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms.
    Entropy 25 00992 g007
    Figure 8. These figures depict the results from our experiments for the artificially generated datasets, comparing differently parameterized QML algorithms on the SU ( 2 ) -generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the runtimes change with the increasing size of the test datasets. The right part contains the legend, indicating which algorithms were used and, more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator.
    Figure 8. These figures depict the results from our experiments for the artificially generated datasets, comparing differently parameterized QML algorithms on the SU ( 2 ) -generated datasets. The upper part illustrates the behavior of the accuracies, while the lower part demonstrates how the runtimes change with the increasing size of the test datasets. The right part contains the legend, indicating which algorithms were used and, more specifically, the different parametrizations of the employed quantum machine learning algorithms. Furthermore, the legend is sorted in decreasing order of the average accuracy of the employed algorithms. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator.
    Entropy 25 00992 g008
    Table 1. This table presents the scores/accuracies of our experiments conducted on artificially generated classification datasets of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in decreasing order of the average accuracy over all different sample sizes of each algorithm. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Table 1. This table presents the scores/accuracies of our experiments conducted on artificially generated classification datasets of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in decreasing order of the average accuracy over all different sample sizes of each algorithm. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Algorithm/ParametrizationSize 50Size 100Size 250Size 500Size 1000Size 1500Size 2000Average
    OutOfTheBox, CatBoost, results1.01.00.980.970.9250.930.94250.963929
    RandomSearch, CatBoost, results1.01.00.960.960.9350.9366670.94250.962024
    RandomSearch, SVM, results1.01.00.940.960.9450.930.93750.958929
    RandomSearch, XGBoost, results1.00.950.980.960.930.9333330.94250.956548
    QKE, PauliFeatureMap, statevector-simulator, 1000.01.01.00.960.930.930.930.9250.953571
    OutOfTheBox, XGBoost, results1.00.950.940.960.910.9366670.950.949524
    OutOfTheBox, SVM, results1.01.00.920.920.940.9333330.930.949048
    QKE, ZZFeatureMap, statevector-simulator, 177.827941003892281.01.00.940.930.9150.9266670.92250.947738
    QKE, ZFeatureMap, statevector-simulator, 5.6234132519034911.01.00.920.910.9250.930.93750.946071
    RandomSearch, MLP, results1.01.00.940.880.9050.9333330.940.942619
    OutOfTheBox, MLP, results1.01.00.940.890.9050.9166670.92750.939881
    OutOfTheBox, Ridge, results1.01.00.940.880.90.8966670.90250.93131
    QKE, ZFeatureMap, qasm-simulator, 5.6234132519034911.01.00.940.820.910.920.90250.9275
    QKE, ZZFeatureMap, statevector-simulator, 31.6227766016837931.00.950.920.880.880.9266670.91750.924881
    QKE, PauliFeatureMap, statevector-simulator, 5.6234132519034911.00.950.920.850.8950.930.920.923571
    QKE, ZFeatureMap, statevector-simulator, 0.17782794100389231.00.950.90.880.90.920.91250.923214
    QKE, ZFeatureMap, aer-simulator, 0.17782794100389231.00.950.90.870.9050.920.91250.9225
    RandomSearch, Ridge, results1.01.00.90.880.880.8933330.90250.922262
    QKE, ZZFeatureMap, qasm-simulator, 5.6234132519034911.00.950.920.860.890.910.91750.921071
    QKE, PauliFeatureMap, qasm-simulator, 5.6234132519034911.00.950.920.860.890.910.91750.921071
    VQC, ZFeatureMap, EfficientSU2, COBYLA, statevector-simulator1.00.950.90.90.920.8933330.880.920476
    RandomSearch, Lasso, results1.01.00.940.820.8950.8933330.890.919762
    VQC, ZFeatureMap, EfficientSU2, COBYLA, qasm-simulator1.00.950.90.880.920.910.8450.915
    QKE, PauliFeatureMap, aer-simulator, 1.00.90.950.920.890.890.930.910.912857
    VQC, ZFeatureMap, EfficientSU2, SPSA, qasm-simulator1.00.950.90.860.9250.910.8450.912857
    VQC, ZFeatureMap, EfficientSU2, COBYLA, aer-simulator1.00.950.920.880.90.9066670.82750.912024
    VQC, ZFeatureMap, EfficientSU2, SPSA, statevector-simulator1.00.950.920.870.890.890.8350.907857
    VQC, ZFeatureMap, RealAmplitudes, COBYLA, aer-simulator1.00.950.90.860.9050.850.8650.904286
    RandomSearch, LightGBM, results0.41.00.980.950.930.9333330.94750.877262
    OutOfTheBox, LightGBM, results0.41.00.960.940.9250.9366670.93750.87131
    VQC, PauliFeatureMap, EfficientSU2, SPSA, qasm-simulator0.90.750.90.840.890.860.86750.858214
    VQC, ZFeatureMap, EfficientSU2, NFT, statevector-simulator1.00.950.860.720.90.7766670.770.85381
    QKE, PauliFeatureMap, aer-simulator, 31.6227766016837931.00.850.960.70.8750.8266670.7350.849524
    QKE, ZFeatureMap, aer-simulator, 31.6227766016837931.01.00.880.620.8350.7366670.74750.83131
    QKE, PauliFeatureMap, aer-simulator, 1000.01.00.850.960.580.870.8266670.6650.821667
    VQC, PauliFeatureMap, EfficientSU2, SPSA, aer-simulator0.80.750.90.730.8450.860.85250.819643
    VQC, PauliFeatureMap, EfficientSU2, NFT, statevector-simulator0.80.650.90.80.840.7833330.84750.802976
    QKE, ZFeatureMap, qasm-simulator, 177.827941003892280.91.00.880.570.8750.730.63750.798929
    VQC, ZZFeatureMap, EfficientSU2, COBYLA, aer-simulator0.70.70.90.710.820.8266670.8350.784524
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, qasm-simulator0.80.70.90.620.7750.8166670.7850.770952
    VQC, ZZFeatureMap, RealAmplitudes, NFT, qasm-simulator0.70.70.90.860.7750.7866670.5350.750952
    VQC, PauliFeatureMap, RealAmplitudes, NFT, qasm-simulator0.60.70.90.490.80.7633330.780.719048
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, aer-simulator0.50.650.840.730.830.830.5750.707857
    QKE, PauliFeatureMap, aer-simulator, 0.031622776601683790.40.350.90.650.860.9233330.82750.701548
    QKE, PauliFeatureMap, aer-simulator, 0.0056234132519034910.40.350.90.490.750.7666670.82750.640595
    QKE, PauliFeatureMap, qasm-simulator, 0.0056234132519034910.40.350.90.490.750.7666670.82750.640595
    QKE, ZFeatureMap, statevector-simulator, 0.0056234132519034910.40.350.840.490.630.850.830.627143
    VQC, ZFeatureMap, TwoLocal, SPSA, statevector-simulator0.70.650.520.510.520.4933330.580.567619
    QKE, PauliFeatureMap, qasm-simulator, 0.0010.40.350.90.490.480.7533330.49750.552976
    OutOfTheBox, Lasso, results0.40.350.50.490.480.5066670.49750.460595
    VQC, ZZFeatureMap, TwoLocal, COBYLA, qasm-simulator0.20.350.280.350.2250.2166670.39750.288452
    VQC, PauliFeatureMap, TwoLocal, SPSA, qasm-simulator0.20.350.260.380.1850.2233330.40.285476
    VQC, PauliFeatureMap, TwoLocal, COBYLA, statevector-simulator0.20.350.280.360.190.2233330.390.284762
    VQC, PauliFeatureMap, TwoLocal, SPSA, statevector-simulator0.20.350.280.360.190.2233330.390.284762
    Table 2. This table presents the runtimes of our experiments conducted on artificially generated classification datasets of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in increasing order of the average runtime over all different sample sizes of each algorithm. The measured runtime includes hyperparameter tuning via randomized search and five-fold cross-validating, training, and testing the model. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Table 2. This table presents the runtimes of our experiments conducted on artificially generated classification datasets of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in increasing order of the average runtime over all different sample sizes of each algorithm. The measured runtime includes hyperparameter tuning via randomized search and five-fold cross-validating, training, and testing the model. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Algorithm/ParametrizationSize 50Size 100Size 250Size 500Size 1000Size 1500Size 2000Average
    OutOfTheBox, Lasso, results0.0014730.0011620.0011580.0011230.0011410.0011530.0011590.001196
    OutOfTheBox, Ridge, results0.0029330.0015530.0014330.0018940.0026280.0025750.0024360.002207
    OutOfTheBox, SVM, results0.0010210.0006480.0010390.0024570.0055010.0172430.02950.008201
    OutOfTheBox, XGBoost, results0.0168810.0171870.0229220.0387510.051110.1518070.1209730.059947
    OutOfTheBox, LightGBM, results0.0096550.0248870.1041070.1248620.18980.4890430.2183430.165814
    RandomSearch, Lasso, results1.0453280.1134130.1022580.1057360.1040310.1205070.1160060.243897
    RandomSearch, Ridge, results1.1207080.1221880.1147060.1759960.2269490.2558450.250670.323866
    RandomSearch, SVM, results1.063760.1355930.1638750.1596990.2031630.3541720.4427410.360429
    OutOfTheBox, MLP, results0.0829530.0911690.1213170.2327710.4516740.9473731.3769650.472032
    OutOfTheBox, CatBoost, results0.3898260.4119650.6543250.7838250.8675951.0852981.19310.769419
    RandomSearch, LightGBM, results1.7118720.3764940.583870.7047150.7283050.8974281.0000390.857532
    RandomSearch, XGBoost, results1.5725410.3991740.4410590.5779690.997761.4676671.3524740.972663
    VQC, ZFeatureMap, TwoLocal, SPSA, statevector-simulator0.5024470.823911.3196022.90786.7595311.8160118.0647256.027718
    VQC, PauliFeatureMap, TwoLocal, COBYLA, statevector-simulator0.5364540.8869451.7578773.4869758.13782114.68888122.794767.469959
    VQC, PauliFeatureMap, TwoLocal, SPSA, statevector-simulator1.9817850.7158291.6210593.4883728.51762415.17018522.3009727.685118
    VQC, PauliFeatureMap, TwoLocal, SPSA, qasm-simulator0.7507191.1544062.534495.00026211.26513719.49394529.0314639.89006
    VQC, ZZFeatureMap, TwoLocal, COBYLA, qasm-simulator0.7348651.0972022.5147034.99083211.89597119.28340629.3182699.976464
    RandomSearch, MLP, results3.8996343.2983375.0036189.65127414.72992420.65224931.20206912.633872
    QKE, ZFeatureMap, statevector-simulator, 0.17782794100389231.3439830.8022862.1708295.96589918.50454636.65992259.88994117.905344
    QKE, ZFeatureMap, statevector-simulator, 0.0056234132519034910.4112960.6974612.1541646.12256419.67081937.29733462.190118.363391
    QKE, PauliFeatureMap, statevector-simulator, 1000.00.4709330.9562692.7212577.281721.35629840.13071667.42290820.048583
    QKE, PauliFeatureMap, statevector-simulator, 5.6234132519034910.5014460.9222372.7756647.45464221.78063740.42603666.75892720.088513
    QKE, ZFeatureMap, statevector-simulator, 5.6234132519034910.3780180.7573632.1416774.96246419.90156541.91300371.45383120.215417
    QKE, ZZFeatureMap, statevector-simulator, 31.6227766016837930.2143860.5672821.6503045.30243720.7762942.61451772.87107820.570899
    QKE, ZZFeatureMap, statevector-simulator, 177.827941003892280.4610930.9435742.7808047.58085722.90681141.95552168.04555320.667745
    RandomSearch, CatBoost, results8.62787810.87314226.72839535.2085736.90227256.25326537.99492930.369779
    VQC, ZFeatureMap, RealAmplitudes, COBYLA, aer-simulator47.43818363.446748192.148143404.2339541060.2916571619.3972052290.222381811.025467
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, qasm-simulator43.11363683.175558166.040938421.2783741064.2385641702.8930062719.340939885.725859
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, aer-simulator45.90950483.201411152.20265509.19561158.9025321654.0659072603.942577886.774312
    VQC, ZFeatureMap, EfficientSU2, COBYLA, statevector-simulator48.54624381.030425190.958188402.1217221044.8558251807.6763572751.241623903.775769
    VQC, ZFeatureMap, EfficientSU2, COBYLA, aer-simulator57.728111100.590997240.174666507.587091253.0805782139.8552183196.072471070.727019
    VQC, ZFeatureMap, EfficientSU2, COBYLA, qasm-simulator59.058898100.862056242.285405507.1717311262.6501432151.5034993191.7455681073.611043
    VQC, ZZFeatureMap, EfficientSU2, COBYLA, aer-simulator59.651649105.629842254.918442601.2451251335.0179042260.3542943366.655011140.496038
    QKE, ZFeatureMap, qasm-simulator, 177.827941003892284.58947813.18480582.633779332.713271337.1029073020.6895795368.2015091451.30219
    QKE, ZZFeatureMap, qasm-simulator, 5.6234132519034914.35278515.92124997.165028390.4720921573.1031973549.6297986282.6702511701.902057
    QKE, PauliFeatureMap, aer-simulator, 0.031622776601683793.54912515.09414498.970568393.4969211581.6622413554.9629276317.3556691709.298799
    QKE, PauliFeatureMap, aer-simulator, 0.0056234132519034913.37325715.31153899.2351390.521311574.1083713555.30486339.0264431710.982974
    QKE, PauliFeatureMap, qasm-simulator, 0.0056234132519034913.81211519.479307101.289711404.4323841636.246863642.9373936307.6050391730.828973
    QKE, PauliFeatureMap, aer-simulator, 31.6227766016837933.84857817.062982101.387533408.699031635.8631363674.9762576555.8115071771.092718
    VQC, ZFeatureMap, EfficientSU2, NFT, statevector-simulator98.831974167.48274394.378037836.9134512197.6521353719.0471165621.1347081862.205737
    VQC, PauliFeatureMap, EfficientSU2, NFT, statevector-simulator103.914165177.047181423.4236031014.9635112338.0783563953.8617235905.4330941988.10309
    VQC, ZZFeatureMap, RealAmplitudes, NFT, qasm-simulator105.987181183.918751427.0167021036.6054732366.4631524052.5210356042.5380152030.721473
    VQC, PauliFeatureMap, RealAmplitudes, NFT, qasm-simulator103.625823180.306618425.4880491041.1609992371.3667154044.8564756048.5739292030.768373
    VQC, ZFeatureMap, EfficientSU2, SPSA, statevector-simulator119.513477200.101417474.1132881008.9328742601.7319174505.3062686781.0897452241.541284
    VQC, ZFeatureMap, EfficientSU2, SPSA, qasm-simulator145.295744256.711762609.7912291272.6750593150.5275375366.1166028009.6490752687.25243
    VQC, PauliFeatureMap, EfficientSU2, SPSA, aer-simulator144.280811259.102175625.1930961502.4769233356.3407995689.8276158454.1442952861.623673
    VQC, PauliFeatureMap, EfficientSU2, SPSA, qasm-simulator152.666649269.680847642.4007471505.7625213388.6629985709.5058268438.9577092872.519614
    QKE, ZFeatureMap, aer-simulator, 31.6227766016837935.99324125.852654166.703792669.2013092934.1695986729.3141112,037.4306873224.095056
    QKE, PauliFeatureMap, qasm-simulator, 5.6234132519034918.38471532.795287206.595473890.4149043753.4888688537.76858915,232.7455424094.599054
    QKE, PauliFeatureMap, qasm-simulator, 0.0017.79209332.566225207.832614896.0422493778.3243518610.33514715,348.8101424125.957546
    QKE, ZFeatureMap, aer-simulator, 0.177827941003892310.51129643.335078276.8107341111.5456144799.03299610,979.13560119,768.0735745284.063556
    QKE, ZFeatureMap, qasm-simulator, 5.62341325190349111.57392943.186982277.2913141113.6643134842.58709410,978.90847619,798.8211565295.147609
    QKE, PauliFeatureMap, aer-simulator, 1000.012.59693851.788837332.2811041434.2086015986.63100613,592.86606524,280.5440756527.273804
    QKE, PauliFeatureMap, aer-simulator, 1.012.26160451.508959332.5618221423.1111355984.90258713,603.95688724,362.832026538.733573
    Table 3. This table presents the scores/accuracies of our experiments conducted on classification datasets generated via SU ( 2 ) generators of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in decreasing order of the average accuracy over all different sample sizes of each algorithm. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Table 3. This table presents the scores/accuracies of our experiments conducted on classification datasets generated via SU ( 2 ) generators of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in decreasing order of the average accuracy over all different sample sizes of each algorithm. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Algorithm/ParametrizationSize 50Size 100Size 250Size 500Size 1000Size 1500Size 2000Average
    RandomSearch, CatBoost, results0.90.550.780.780.880.9066670.9150.815952
    OutOfTheBox, CatBoost, results0.60.70.760.850.8950.9066670.93750.807024
    RandomSearch, XGBoost, results0.60.650.740.840.870.860.94250.786071
    OutOfTheBox, XGBoost, results0.40.750.760.860.890.9266670.90.78381
    QKE, ZFeatureMap, statevector-simulator, 1000.00.70.80.740.790.80.8066670.84750.783452
    RandomSearch, LightGBM, results0.30.60.740.870.890.9033330.91750.745833
    OutOfTheBox, LightGBM, results0.30.60.740.90.850.9233330.8850.742619
    QKE, ZZFeatureMap, statevector-simulator, 177.827941003892280.80.550.540.710.7950.8266670.850.724524
    QKE, PauliFeatureMap, aer-simulator, 5.6234132519034910.70.650.440.730.7250.6833330.720.664048
    QKE, ZZFeatureMap, qasm-simulator, 31.6227766016837930.70.70.70.590.6150.620.6650.655714
    QKE, ZZFeatureMap, statevector-simulator, 0.17782794100389230.80.50.620.550.6450.6933330.73750.649405
    QKE, ZZFeatureMap, aer-simulator, 0.17782794100389230.40.60.70.670.6250.7133330.70750.630833
    QKE, PauliFeatureMap, aer-simulator, 0.17782794100389230.50.650.640.630.6150.680.70.630714
    QKE, PauliFeatureMap, statevector-simulator, 0.17782794100389230.40.650.60.550.750.6666670.730.620952
    OutOfTheBox, MLP, results0.80.550.520.630.5550.560.620.605
    VQC, ZZFeatureMap, EfficientSU2, SPSA, aer-simulator0.70.50.60.60.6050.5766670.6450.60381
    VQC, ZZFeatureMap, EfficientSU2, COBYLA, qasm-simulator0.60.60.620.540.60.6066670.660.60381
    OutOfTheBox, SVM, results0.40.650.440.590.680.6766670.6950.590238
    RandomSearch, SVM, results0.20.70.80.650.780.5033330.47250.586548
    VQC, ZZFeatureMap, EfficientSU2, NFT, statevector-simulator0.50.60.540.560.70.580.610.584286
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, statevector-simulator0.60.650.60.530.5750.5433330.57750.582262
    VQC, PauliFeatureMap, EfficientSU2, COBYLA, aer-simulator0.40.750.520.550.560.6066670.62250.572738
    VQC, PauliFeatureMap, RealAmplitudes, NFT, statevector-simulator0.30.650.520.690.630.5866670.620.570952
    OutOfTheBox, Ridge, results0.70.550.620.480.5750.5333330.5250.569048
    VQC, ZFeatureMap, RealAmplitudes, COBYLA, qasm-simulator0.70.50.520.570.590.5733330.52750.56869
    VQC, ZZFeatureMap, EfficientSU2, NFT, aer-simulator0.40.70.560.580.5750.5433330.61750.567976
    QKE, ZZFeatureMap, aer-simulator, 31.6227766016837930.40.550.60.620.6250.5966670.5550.56381
    QKE, PauliFeatureMap, aer-simulator, 31.6227766016837930.60.450.60.560.570.620.54250.563214
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, aer-simulator0.60.650.480.550.5350.5733330.5450.561905
    QKE, ZFeatureMap, qasm-simulator, 177.827941003892280.60.70.50.520.50.5566670.530.558095
    VQC, ZFeatureMap, EfficientSU2, COBYLA, aer-simulator0.70.60.50.480.5250.480.6150.557143
    VQC, ZZFeatureMap, RealAmplitudes, NFT, qasm-simulator0.60.40.580.470.660.5733330.610.55619
    VQC, ZFeatureMap, EfficientSU2, COBYLA, qasm-simulator0.70.50.560.550.510.5666670.5050.555952
    QKE, PauliFeatureMap, qasm-simulator, 0.17782794100389230.50.350.360.630.6550.6933330.70250.555833
    QKE, ZZFeatureMap, qasm-simulator, 0.0010.70.450.660.570.540.4666670.47250.55131
    VQC, ZFeatureMap, RealAmplitudes, SPSA, aer-simulator0.70.50.50.590.4950.5166670.5550.550952
    VQC, ZFeatureMap, EfficientSU2, SPSA, statevector-simulator0.30.80.440.590.550.5633330.5950.548333
    VQC, PauliFeatureMap, RealAmplitudes, NFT, qasm-simulator0.40.550.60.540.590.5533330.580.544762
    VQC, ZFeatureMap, RealAmplitudes, COBYLA, aer-simulator0.80.450.480.480.520.5133330.5450.54119
    VQC, ZFeatureMap, EfficientSU2, COBYLA, statevector-simulator0.50.550.560.560.540.5166670.560.540952
    QKE, ZFeatureMap, aer-simulator, 1000.00.60.550.60.520.5050.4866670.520.540238
    VQC, PauliFeatureMap, EfficientSU2, SPSA, qasm-simulator0.40.450.560.640.5750.5333330.60750.537976
    VQC, ZFeatureMap, RealAmplitudes, NFT, aer-simulator0.50.40.620.520.580.5333330.58750.534405
    VQC, PauliFeatureMap, EfficientSU2, SPSA, aer-simulator0.40.450.540.540.640.560.60.532857
    QKE, ZFeatureMap, statevector-simulator, 0.17782794100389230.50.450.540.540.5850.5566670.5550.532381
    QKE, ZFeatureMap, aer-simulator, 1.00.40.50.40.630.560.6033330.5750.524048
    QKE, ZFeatureMap, qasm-simulator, 1000.00.60.50.440.590.5250.4433330.50.514048
    QKE, PauliFeatureMap, statevector-simulator, 0.031622776601683790.50.450.580.560.460.4766670.56750.513452
    RandomSearch, MLP, results0.30.250.50.60.6350.640.66750.513214
    QKE, ZFeatureMap, aer-simulator, 177.827941003892280.40.650.540.460.4950.480.5650.512857
    QKE, ZFeatureMap, qasm-simulator, 0.0056234132519034910.50.650.480.470.5150.480.4850.511429
    RandomSearch, Ridge, results0.40.550.520.460.560.520.56750.511071
    OutOfTheBox, Lasso, results0.50.70.40.510.4950.460.4950.508571
    RandomSearch, Lasso, results0.60.450.440.450.5150.5333330.540.504048
    Table 4. This table presents the scores/accuracies of our experiments conducted on classification datasets generated via SU ( 2 ) generators of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in increasing order of the average runtime over all different sample sizes of each algorithm. The measured runtime includes hyperparameter tuning via randomized search and five-fold cross-validating, training, and testing the model. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Table 4. This table presents the scores/accuracies of our experiments conducted on classification datasets generated via SU ( 2 ) generators of varying sizes, e.g., 50 and 100. Given these different dataset sizes, this table is sorted in increasing order of the average runtime over all different sample sizes of each algorithm. The measured runtime includes hyperparameter tuning via randomized search and five-fold cross-validating, training, and testing the model. The parametrization for the QKE is as follows: QKE, feature map, quantum simulator, C-Value for the SVM algorithm. The parametrization for the VQC is as follows: VQC, feature map, Ansatz, optimizer, quantum simulator. For the classical machine learning algorithms, OutOfTheBox means that we did not tune the hyperparameters of the employed algorithm and RandomSearch refers to hyperparameter optimization via a randomized search.
    Algorithm/ParametrizationSize 50Size 100Size 250Size 500Size 1000Size 1500Size 2000Average
    OutOfTheBox, Lasso, results0.0041030.0006460.0006610.0012490.0008040.0006940.0006910.001264
    OutOfTheBox, Ridge, results0.0037330.0028980.0027860.0296880.0018990.0019290.001780.006388
    OutOfTheBox, SVM, results0.001110.0009190.0022980.0060780.0121390.0259350.046670.013593
    RandomSearch, Lasso, results1.0556540.1230470.1034570.1042520.1175910.1243010.1158390.249163
    RandomSearch, Ridge, results1.0843480.1227410.1382480.1455620.1567830.1299150.142620.274317
    OutOfTheBox, XGBoost, results0.0266160.0409690.2072651.3484210.1902740.1238780.2054960.306131
    RandomSearch, SVM, results1.1339570.1679290.1311670.1478770.2333140.4448340.4382090.385327
    OutOfTheBox, CatBoost, results1.0724320.4398860.684830.8136990.9366631.131451.2522420.904457
    RandomSearch, XGBoost, results1.5553240.3987610.5178640.8171181.793722.0564362.2895631.346969
    OutOfTheBox, LightGBM, results0.1344990.9351280.3193971.1250765.6464884.9082064.3121012.482985
    OutOfTheBox, MLP, results0.4473430.3852580.3030561.8746622.1654865.5828016.9861282.534962
    RandomSearch, LightGBM, results3.767520.7264690.8924961.3494115.9424325.5695743.7418423.141392
    QKE, ZFeatureMap, statevector-simulator, 1000.00.5888031.1051123.1152289.69281824.48824546.86338175.60530523.065556
    QKE, ZZFeatureMap, statevector-simulator, 177.827941003892281.0233772.0147386.13374414.0885536.20986963.51069897.57521731.508028
    QKE, PauliFeatureMap, statevector-simulator, 0.17782794100389231.1100542.1547836.59103614.48780236.08956262.91284897.6427331.569831
    RandomSearch, MLP, results16.9045215.1291565.67144642.57679638.56464170.80703564.60204836.322235
    QKE, ZFeatureMap, statevector-simulator, 0.17782794100389231.2350191.9846075.54759415.33380841.97768679.052726127.71062838.977438
    QKE, PauliFeatureMap, statevector-simulator, 0.031622776601683791.4535792.8040569.08531319.44072147.63334781.740004128.18663641.477665
    QKE, ZZFeatureMap, statevector-simulator, 0.17782794100389232.1942364.85654910.04320720.39173957.65883497.515617151.07067649.104408
    RandomSearch, CatBoost, results18.35017435.65474240.72586870.6595668.78844668.95894943.68561949.546194
    VQC, ZFeatureMap, RealAmplitudes, COBYLA, qasm-simulator55.500573100.423827241.59017577.4002581286.9628672164.2822463315.1858711105.906545
    VQC, ZFeatureMap, RealAmplitudes, COBYLA, aer-simulator61.174412116.296597274.944359672.2159721509.9872982624.7105164058.8192831331.164062
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, statevector-simulator68.600875123.207802380.446356770.7161851635.899192621.2428223805.0212061343.590634
    VQC, PauliFeatureMap, EfficientSU2, COBYLA, aer-simulator89.832315163.70887480.279572975.2912552084.1745643407.8449345050.4334051750.223559
    VQC, ZZFeatureMap, EfficientSU2, COBYLA, qasm-simulator88.21128163.988243480.35886973.6145662068.9503563425.5647595057.8497611751.219689
    VQC, ZFeatureMap, EfficientSU2, COBYLA, qasm-simulator85.847133156.496491381.392026888.4611742300.5276293878.069956136.1752471975.281379
    VQC, ZFeatureMap, EfficientSU2, COBYLA, statevector-simulator103.51928191.066923456.0172771079.391812305.4440053940.2122295958.2774532004.846997
    VQC, ZFeatureMap, RealAmplitudes, NFT, aer-simulator111.03018203.235006491.9789031181.6688282620.5447814428.847016770.5382132258.263274
    VQC, ZFeatureMap, EfficientSU2, COBYLA, aer-simulator113.765615205.589837516.2898811202.2886622663.1219474533.3780976730.846632280.754381
    VQC, ZZFeatureMap, RealAmplitudes, COBYLA, aer-simulator111.686165209.034186638.2941791296.485542869.7089464781.9810677074.9254792426.016509
    VQC, PauliFeatureMap, RealAmplitudes, NFT, qasm-simulator163.24924303.991817936.8980121915.6989944145.4669836900.4320810,273.3740653519.873027
    VQC, ZFeatureMap, EfficientSU2, SPSA, statevector-simulator190.048137341.454534823.8758591930.6378184188.2379676999.25880710,586.64483580.02256
    VQC, PauliFeatureMap, RealAmplitudes, NFT, statevector-simulator190.485358349.9068651108.1806722248.3452324814.1438067883.61088511,821.0595174059.390334
    VQC, ZFeatureMap, RealAmplitudes, SPSA, aer-simulator195.45132357.101921856.6798862110.8579924766.4498268178.59332312,549.7993714144.99052
    VQC, ZZFeatureMap, RealAmplitudes, NFT, qasm-simulator224.602174405.4979281270.7415522674.1411095725.2172529676.19154714,306.048814897.491482
    VQC, ZZFeatureMap, EfficientSU2, NFT, statevector-simulator243.54281457.1722661372.5286442784.4221745896.2971669666.12619114,179.3313144942.774366
    QKE, ZFeatureMap, aer-simulator, 1.09.84710440.399745258.3792681171.6123934847.63838110,994.73662919,569.1320425270.249366
    VQC, ZZFeatureMap, EfficientSU2, NFT, aer-simulator259.240503473.1264741408.6010572903.5637376301.78696310,332.95920115,345.3144365289.227482
    VQC, PauliFeatureMap, EfficientSU2, SPSA, aer-simulator316.140029574.3014721727.7001923530.7229186706.0760429861.49159514,726.3884765348.974389
    QKE, ZFeatureMap, aer-simulator, 1000.010.89817146.773357297.3901071339.7844665587.10790411,607.62334419,440.970925475.79261
    VQC, ZZFeatureMap, EfficientSU2, SPSA, aer-simulator243.897139463.5862311368.883242789.1339916570.67682111,456.70576517,726.7965885802.811396
    VQC, PauliFeatureMap, EfficientSU2, SPSA, qasm-simulator348.216144639.1317721898.4698023945.8072498437.84889313,815.44667320,511.0341767085.136387
    QKE, ZFeatureMap, qasm-simulator, 1000.011.57945147.675482344.5437751568.9039396191.83091214,908.46770127,002.7660787153.681048
    QKE, ZFeatureMap, aer-simulator, 177.8279410038922814.16361956.856619359.7933431620.4826266647.99084915,084.13776426,961.3267137249.250219
    QKE, ZFeatureMap, qasm-simulator, 177.8279410038922816.3571777.129608482.4788772237.681529219.95434418,899.04655226,623.3124878222.28008
    QKE, ZFeatureMap, qasm-simulator, 0.00562341325190349116.18445968.030962439.8891232003.145868339.15707218,939.86641833,875.1898229097.351959
    QKE, PauliFeatureMap, aer-simulator, 31.62277660168379316.82244670.391611549.2859962267.7943919148.30649920,490.13138936,687.6388089890.05302
    QKE, ZZFeatureMap, aer-simulator, 31.62277660168379317.38223470.921393552.7202362290.3051189223.0182420,681.45066836,991.5540659975.335993
    QKE, ZZFeatureMap, aer-simulator, 0.177827941003892319.61800680.653612632.0122982628.4070389714.43148920,666.72584436,766.37877610,072.603866
    QKE, PauliFeatureMap, aer-simulator, 5.62341325190349120.0346181.805468657.4373842646.60001810,751.04372224,303.41059442,050.18660111,501.502628
    QKE, ZZFeatureMap, qasm-simulator, 0.00122.47487194.5939748.536393061.49290811,095.03755724,179.10849442,833.54406111,719.255454
    QKE, PauliFeatureMap, aer-simulator, 0.177827941003892315.7044964.293166539.3724322052.76724510,360.38173528,219.13410353,610.13857913,551.684536
    QKE, PauliFeatureMap, qasm-simulator, 0.177827941003892328.777769121.675706961.2485343951.05299216,159.34356135,692.43133448,691.26245115,086.541764
    QKE, ZZFeatureMap, qasm-simulator, 31.62277660168379328.201021110.795119877.1412223647.41301716,257.20780538,819.79606569,300.66174918,434.459428
    Table 5. These tables present the scores/accuracies of our experiments conducted on publicly available classification datasets. The upper table displays the best five-fold cross-validation scores, obtained using randomized search cross-validation from Scikit-learn, which were employed to identify the optimal model. The lower table shows the scores of the best model evaluated on an unseen test subset of the original data. We include results for the six datasets described in Section 3.5, the quantum classifiers detailed in Section 3.3, and the classical machine learning classifiers discussed in Section 3.2.
    Table 5. These tables present the scores/accuracies of our experiments conducted on publicly available classification datasets. The upper table displays the best five-fold cross-validation scores, obtained using randomized search cross-validation from Scikit-learn, which were employed to identify the optimal model. The lower table shows the scores of the best model evaluated on an unseen test subset of the original data. We include results for the six datasets described in Section 3.5, the quantum classifiers detailed in Section 3.3, and the classical machine learning classifiers discussed in Section 3.2.
    Classifier\DatasetIrisWineILPDBC-CoimbraTAEBreast-Tissue
    VQC0.8170.8170.7060.5990.4170.339
    QKE0.9080.8530.7060.6200.4830.382
    Ridge0.9140.8750.0800.0530.053<0.001
    Lasso0.9140.8700.0850.0040.004<0.001
    MLP0.9750.9370.7120.6870.4250.406
    SVM0.9580.7590.7060.6300.4500.382
    XGBoost0.9580.9860.6950.6560.5330.441
    LightGBM0.9670.9860.6990.6660.4750.393
    CatBoost0.9500.9790.7020.6880.5250.440
           
    Classifier\DatasetIrisWineILPDBC-CoimbraTAEBreast-Tissue
    VQC0.7670.6390.7440.5410.3880.334
    QKE1.00.8330.7440.7920.6130.409
    Ridge0.9470.8780.1150.234<0.001<0.001
    Lasso0.9450.8820.1150.296<0.001<0.001
    MLP1.01.00.7690.8750.3870.455
    SVM1.00.9720.7430.8750.3550.455
    XGBoost1.01.00.7350.9170.5330.441
    LightGBM1.01.00.7520.9170.4190.455
    CatBoost1.01.00.7440.9170.6450.545
    Table 6. This table presents the combined runtimes of our experiments conducted on well-known and publicly available classification datasets. The runtimes include both the five-fold randomized search cross-validation process from Scikit-learn, which was employed to identify the optimal model, and the evaluation of the best model on an unseen test subset of the original data. We include results for the six datasets described in Section 3.5, the quantum classifiers detailed in Section 3.3, and the classical machine learning classifiers discussed in Section 3.2.
    Table 6. This table presents the combined runtimes of our experiments conducted on well-known and publicly available classification datasets. The runtimes include both the five-fold randomized search cross-validation process from Scikit-learn, which was employed to identify the optimal model, and the evaluation of the best model on an unseen test subset of the original data. We include results for the six datasets described in Section 3.5, the quantum classifiers detailed in Section 3.3, and the classical machine learning classifiers discussed in Section 3.2.
    Classifier\
    Dataset
    IrisWineILPDBC-CoimbraTAEBreast-Tissue
    VQC3:32:16.5476051 day, 13:56:59.4551852 days, 23:03:26.3988569:55:17.9074432:46:25.9215539:01:58.623806
    QKE2:03:57.92115421:41:38.7382557 days, 6:30:41.1796765:02:26.4300011:28:54.0697253:37:05.655104
    Ridge0:00:00.1750090:00:00.4967710:00:00.3992290:00:00.2408570:00:00.2096000:00:00.296966
    Lasso0:00:00.1730510:00:00.1814440:00:00.2374550:00:00.1922570:00:00.2295080:00:00.225531
    MLP0:00:16.8762880:00:10.4774200:00:26.7489070:00:10.9512290:00:08.4752630:00:13.729790
    SVM0:00:00.1433530:00:00.1654310:00:00.4844850:00:00.1806940:00:00.2285080:00:00.226784
    XGBoost0:00:03.8090850:00:04.0304250:00:04.7526270:00:02.7441220:00:05.8203710:00:06.864497
    LightGBM0:00:02.9711640:00:03.1807700:00:03.0625530:00:01.4621740:00:03.0566150:00:04.540870
    CatBoost0:00:06.4659750:00:18.5116120:00:11.3529440:00:07.4604600:00:06.9648210:00:26.639070
    Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

    Share and Cite

    MDPI and ACS Style

    Raubitzek, S.; Mallinger, K. On the Applicability of Quantum Machine Learning. Entropy 2023, 25, 992. https://doi.org/10.3390/e25070992

    AMA Style

    Raubitzek S, Mallinger K. On the Applicability of Quantum Machine Learning. Entropy. 2023; 25(7):992. https://doi.org/10.3390/e25070992

    Chicago/Turabian Style

    Raubitzek, Sebastian, and Kevin Mallinger. 2023. "On the Applicability of Quantum Machine Learning" Entropy 25, no. 7: 992. https://doi.org/10.3390/e25070992

    Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

    Article Metrics

    Back to TopTop