Next Article in Journal
A Systematic Review of Machine Learning in Credit Card Fraud Detection Under Original Class Imbalance
Previous Article in Journal
A Web-Based Digital Twin Framework for Interactive E-Learning in Engineering Education
Previous Article in Special Issue
Mapping the Chemical Space of Antiviral Peptides with Half-Space Proximal and Metadata Networks Through Interactive Data Mining
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sparse Keyword Data Analysis Using Bayesian Pattern Mining

Department of Data Science, Cheongju University, Cheongju-si 28503, Chungbuk, Republic of Korea
Computers 2025, 14(10), 436; https://doi.org/10.3390/computers14100436
Submission received: 11 September 2025 / Revised: 11 October 2025 / Accepted: 13 October 2025 / Published: 14 October 2025

Abstract

Keyword data analysis aims to extract and interpret meaningful relationships from large collections of text documents. A major challenge in this process arises from the extreme sparsity of document–keyword matrices, where the majority of elements are zeros due to zero inflation. To address this issue, this study proposes a probabilistic framework called Bayesian Pattern Mining (BPM), which integrates Bayesian inference into association rule mining (ARM). The proposed method estimates both the expected values and credible intervals of interestingness measures such as confidence and lift, providing a probabilistic evaluation of keyword associations. Experiments conducted on 9436 quantum computing patent documents, from which 175 representative keywords were extracted, demonstrate that BPM yields more stable and interpretable associations than conventional ARM. By incorporating credible intervals, BPM reduces the risk of biased decisions under sparsity and enhances the reliability of keyword-based technology analysis, offering a rigorous approach for knowledge discovery in zero-inflated text data.

1. Introduction

The exponential growth of digital information has made text data one of the most widely used and important resources in modern data analysis. With the rapid accumulation of documents such as research papers, patents, reports, and online content, extracting useful insights from textual information has become a critical research topic [1,2,3,4,5,6,7]. Among the various approaches to text mining, keyword analysis has played a central role because keywords provide a compact and meaningful representation of document content. By analyzing keywords, researchers can identify emerging themes, discover semantic structures, and capture technological and social trends [1,2,8,9,10,11,12,13,14]. A standard process in keyword analysis begins with constructing a document–keyword matrix, where rows represent documents and columns represent keywords. Each entry in this matrix denotes the frequency with which a keyword occurs in a document. However, in practice, such matrices are characterized by extreme sparsity. That is, extreme sparsity arises because even if a particular keyword appears in only a single document, it is still represented as a separate column in the document–keyword matrix. As a result, that column has a non-zero entry in only one row, while all other rows contain zero, which leads to a highly sparse and zero-inflated structure. This leads to the well-known zero-inflated problem, which has been reported as a major obstacle to the reliable performance of analytical models [3,4,15,16,17,18,19,20,21,22,23]. When sparsity is severe, traditional frequency-based methods may provide unstable or biased results, making it difficult to draw meaningful conclusions.
Over the years, a variety of methods have been proposed to address sparsity in keyword analysis. Classical approaches include weighting schemes such as term frequency–inverse document frequency (TF–IDF) [24,25,26,27], which adjust raw frequencies to better reflect the discriminative power of keywords across documents. More recently, probabilistic models, graph-based approaches, and embedding methods (e.g., Word2Vec, BERT) have been applied to keyword analysis, aiming to capture latent semantic relationships beyond raw co-occurrence counts [3,5,28,29,30,31,32,33,34]. While these methods represent significant progress, they are not free from limitations. Frequency-based and TF–IDF approaches often remain sensitive to sparsity. Graph-based and embedding approaches provide richer representations but may require large corpora, heavy computation, or domain-specific fine-tuning, which can limit their applicability in practice. Moreover, these methods typically rely on point estimates, which may not adequately account for uncertainty in sparse and noisy data. Against this backdrop, Bayesian statistics has recently attracted increasing attention as a powerful tool for handling sparsity and uncertainty in text analysis [16,18,19,20,32]. By incorporating prior information and using posterior distributions, Bayesian methods can provide both point estimates and uncertainty intervals, thereby supporting more robust inference. In particular, Bayesian models have shown promise in zero-inflated count data analysis, such as in biomedical, ecological, and patent domains [16,19,21]. Building on this line of research, Bayesian approaches offer an opportunity to enhance keyword analysis, where sparsity and zero inflation are pervasive. The objective of this study is therefore to develop a robust methodology for sparse keyword data analysis that explicitly addresses the zero-inflated problem. We propose Bayesian Pattern Mining (BPM), which integrates Bayesian inference with association rule mining (ARM). Traditional ARM has been widely used in fields such as market basket analysis and text mining [35,36,37], where it discovers frequent co-occurrence patterns among items or keywords. However, ARM relies only on single-value measures such as support, confidence, and lift, without quantifying the uncertainty of these measures. This limitation becomes critical when analyzing sparse document–keyword matrices, where biased or unstable estimates may lead to incorrect interpretations.
Our proposed BPM framework addresses this challenge by introducing Bayesian inference into pattern mining. Specifically, BPM applies Bayesian priors and likelihoods to association rules, yielding posterior distributions for interestingness measures. From these posterior distributions, BPM derives not only the expected (mean) values but also credible intervals for support, confidence, and lift. This dual estimation enables analysts to make decisions that account for uncertainty, thereby reducing bias and improving robustness in sparse settings. For example, in cases where ARM would produce a misleadingly high confidence due to data sparsity, BPM can reveal wide credible intervals that caution against over-interpretation. Despite the progress achieved by recent text mining methods, there remains a clear research gap between semantic representation and probabilistic uncertainty modeling. Embedding-based and large language model (LLM)-based models (e.g., Word2Vec, BERT, GPT) have advanced the representation of keyword meaning, yet they typically rely on large corpora and deterministic embeddings, without explicitly quantifying the uncertainty caused by data sparsity. Moreover, these approaches seldom address the zero-inflated nature of document–keyword matrix. In contrast, our proposed BPM framework directly incorporates Bayesian inference into rule-based keyword analysis to estimate both the mean and credible intervals of association measures. This enables a rigorous probabilistic treatment of uncertainty, thereby filling a methodological gap in the current literature on sparse text analysis.
The contributions of this paper are fourfold as follows: First, we introduce BPM as a novel approach that combines Bayesian inference with association rule mining to mitigate the zero-inflated sparsity problem in keyword data. Second, we demonstrate how Bayesian credible intervals can enhance the interpretability of association measures, offering a probabilistic framework that extends traditional ARM. Third, using a large-scale dataset of 9436 patent documents in the field of quantum computing, we validate the effectiveness of BPM compared to conventional ARM, showing that BPM provides more stable and reliable keyword associations. Lastly, we illustrate how BPM can support technology forecasting and patent analysis by offering more trustworthy insights into the associative structures among technological keywords. The remainder of this paper is organized as follows. Section 2 reviews the research background, including keyword analysis and association rule mining. Section 3 presents the proposed BPM methodology in detail. Section 4 reports experimental results using the quantum computing patent dataset. Section 5 provides a discussion of theoretical and practical implications, followed by conclusions and future research directions.

2. Research Background

2.1. Keyword Analysis

Keyword analysis is to analyze the document–keyword matrix constructed from the text document data. Figure 1 shows text preprocessing and mining of collected text documents.
We search the text documents related to target issues from diverse resources. By text processing, the collected text documents are changed to a corpus. A corpus is a dataset that has been structured and transformed into a form that can be processed by a computer [24]. Using text mining techniques, we construct the document–keyword matrix from the corpus. Using this matrix, we can perform various keyword analyses. Keyword analysis has been widely used as a fundamental approach in text document studies to capture the social and technological trends of documents. In general, keyword analysis is also to extract the terms that frequently occur in the corpus, followed by quantifying their distribution and statistical significance [24,25,26]. Keywords provide a meaningful representation of document content and build associative structures such as graph or topic models [8]. Most studies have primarily relied on frequency counts to identify keywords occurring in documents. However, keyword analysis methods based on frequency are often limited by the sparsity and skewness of keyword distributions caused by the zero-inflated problem in the document–term matrix [3,20]. To address this problem in the early stages of keyword analysis, Salton and Buckley (1988) introduced weighting schemes such as term frequency–inverse document frequency (TF–IDF) and probabilistic models [25]. So, they enhanced the discriminative power of keywords across documents. Some recent research has used keyword network models based on graph structure, social network analysis, and probabilistic approaches [3,5,28,29,30]. Bayesian keyword analysis has also been actively researched in various fields recently [3,16,30,31]. Jun (2024) proposed a zero-inflated model based on Bayesian statistics to analyze keyword data [3], because Bayesian data analysis provides a way to solve the sparsity problem of zero-overcount data [18,19,32].
Recently, new methods have been actively studied in patent keyword analysis, which extracts and analyzes keywords from patent documents [7,20,31]. These studies have used advanced statistical analysis methods as well as machine learning algorithms for patent keyword analysis. Keyword analysis has been making significant contributions in solving problems in various domains. For example, patent keyword analysis has been used to predict technological trends and understand technological structures within a target domain [33]. In keyword analysis, a fundamental challenge concerns the construction of an appropriate semantic representation of keywords, as traditional frequency-based or co-occurrence methods are often sparse and limited in capturing semantic relations. Recent studies have therefore compared alternative approaches such as co-word networks, word embeddings, and network embeddings, demonstrating that their effectiveness in bibliometric research largely depends on corpus characteristics and analytical objectives [10,34]. There are still many issues to be resolved in the field of keyword analysis, and data sparsity of zero inflation is also a critical issue that needs to be addressed. In our paper, we study an analysis method to solve this problem.
In addition to keyword-based methods, a significant body of research has focused on multiclass text classification and clustering to find latent structures in text corpora. Many of these approaches are grounded in probabilistic modeling, including the use of Dirichlet distributions, multinomial models, and Bayesian inference. For example, Campello et al. (2013) studied density-based clustering and hierarchical density estimation to group documents based on probabilistic similarity measures [38]. Also, Aljalbout et al. (2018) integrated deep learning with clustering to improve representation learning for large-scale corpora [39]. In scientific text analysis, Beltagy et al. (2019) developed pretrained language models such as SciBERT, which showed strong performance not only in classification tasks but also in entity and relation extraction, closely connected to rule mining and keyword association [40]. These previous studies demonstrated that their probabilistic approaches to clustering and classification are powerful tools for text mining, complementing keyword analysis by capturing semantic and structural relationships at multiple levels.
While embedding-based methods such as Word2Vec, BERT, and SciBERT [10,34] provide powerful representations of semantic similarity, they typically assume dense vector spaces and do not directly address the statistical sparsity and uncertainty inherent in document–keyword matrices. LLMs further extend this capability to contextual semantics but are computationally intensive and lack mechanisms to quantify probabilistic confidence in inferred associations. In contrast, BPM focuses on probabilistic rule inference, offering a complementary approach that estimates the reliability of associations rather than semantic proximity.

2.2. Association Rule Mining

ARM is a data mining method designed to uncover hidden relationships among items in large datasets [35]. This was introduced as an algorithm for application in market basket analysis. ARM consists of items and transactions. An item is a distinct attribute, object, or keyword that can appear in a dataset. A transaction is a collection of items that occur together in one record, such as a shopping basket or a document. We build a matrix consisting of transactions and items as rows and columns, respectively. Figure 2 shows the transaction–item matrix for ARM.
In Figure 2, each element b i j of this matrix has a binary value of 1 or 0. That is, if the jth item occurs in the ith transaction, b i j = 1 . This matrix is input data for general ARM. To evaluate the performance of rules extracted through ARM, we use three measures: support, confidence, and lift. When an association rule is expressed as X Y , X is the antecedent item and Y is the consequent item [35,36]. The support of X Y is defined as follows [35,36].
S u p p o r t X Y = P X Y
In Equation (1), the S u p p o r t X Y is the probability that a transaction contains X Y . The confidence of X Y is computed by following equation [35,36].
C o n f i d e n c e X Y = P X Y P ( X )
In Equation (2), the C o n f i d e n c e X Y is the conditional probability that a transaction having Y given X . The remaining measure for evaluating the performance of ARM is lift, which is defined as follows [35,36].
L i f t X Y = C o n f i d e n c e X Y P ( Y )
The lift value is not a probability in Equation (3). This measure has a value greater than or equal to 0. If the lift value is greater than 1, X and Y are complementary, and if it is less than 1, they are substitutes. Therefore, we find the meaningful rules from given data using support, confidence, and lift measures.

3. Bayesian Pattern Mining for Sparse Keyword Data Analysis

In ARM, three primary measures (support, confidence, and lift) are commonly used to evaluate the strength of associations between items. In this study, these measures are further extended under a Bayesian framework. The term confidence refers to its conventional definition in ARM, representing the conditional probability of observing the consequent given the antecedent. To avoid confusion, the term credible interval is used instead of confidence interval when describing Bayesian posterior intervals. In this paper, we propose a method of Bayesian Pattern Mining, an association rule mining method based on Bayesian inference, to solve the sparsity problem that occurs during keyword data analysis. In general, the keyword data have a zero-inflated problem. That is, since a significant number of elements in the keyword data have zero values, the data sparsity problem occurs due to zero excess. Our proposed method consists of two analysis procedures. The first procedure is to build a document–keyword matrix from the collected text documents to make the data to be used for Bayesian Pattern Mining. The procedure of building document–keyword matrix is shown as follows. Using the result of building document–keyword matrix in Algorithm 1, we perform the BPM to analyze the sparse keyword data as follows.
Algorithm 1 Building Document–Keyword Matrix
Input:
    - Document data
    - Unique string separator
Output:
    - Extracted keywords
    - Document–keyword matrix
1. Collecting text data
- documents, papers, patents, reports, news articles, legal documents
- social media and chat comments, online comments
2. Checking data schema
- removing duplicated documents
- missing value imputation
3. Normalizing text data
- removing spaces, numbers, symbols, etc.
- lowercasing, using stopwords by user dictionary
4. Lemmatization and stemming
5. Parsing
- corpus, text database
6. Constructing document–keyword matrix
- sparse matrix, keyword extraction
- each element representing frequency of keywords occurring in each document
7. Changing element value from count to binary
- binary document–keyword matrix
- using binary matrix for transaction data for association rule mining
The input and output of Algorithm 1 are the collected text documents and document–keyword matrix, respectively. We collect text data from various data sources. The collected data goes through a preprocessing with text mining after data checking such as removing duplicate documents and missing value imputation. Using the data normalizing, lemmatization, stemming, and parsing techniques of text mining, we construct the document–keyword matrix. At this stage, the preprocessing of various text mining such as removing spaces, lowercasing, stemming, and corpus parsing is performed. The rows and columns of the final document–keyword matrix represent documents and keywords, respectively, and each element indicates the frequency of a keyword contained in a document. Typically, the number of keywords is much larger than the number of documents, which causes the document–keyword matrix to become extremely sparse, with most entries equal to zero, thereby giving rise to a zero-inflated problem. In this paper, we study a new method to solve this problem. In Algorithm 1, all non-zero count values in this matrix are changed to one. This binary matrix becomes the transaction data used in the association rule mining.
To solve the sparsity problem of zero inflation that occurs in text data analysis, the method we study in this paper is sparse text data analysis based on Bayesian Pattern Mining (BPM). The BPM, a combination of Bayesian statistics and ARM, is a methodology that infers the uncertainty of association rules probabilistically by reflecting prior knowledge in the model. For example, ARM provides a single value representing the lift measure, but BPM provides not only a single lift value but also a credible interval for the lift, giving you more information to understand and utilize the association rules. This characteristic of BPM enables it to effectively address the sparsity issue inherent in keyword data analysis. Pattern mining is the process of discovering meaningful patterns from data consisting of items and transactions. By applying Bayesian statistics, the significance of patterns is evaluated using posterior distributions derived from the prior distributions and likelihood functions. Algorithm 2 shows the procedure of BPM, from the sparse document–keyword matrix to summary statistics of extracted association rules.
Algorithm 2 Bayesian Pattern Mining
Input:
    - sparse document–keyword matrix with binary data
Output:
    - means and credible intervals of association rules
1. Constructing 2 × 2 contingency table
- X: antecedent item (keyword)
- Y: consequent item (keyword)
2. Building posterior distribution
- prior: P ~ D i r i c h l e t α 11 , α 10 , α 01 , α 00 ,   α i j > 0
- likelihood: n | P ~ M u l t i n o m i a l ( N , P ) ,   N = n 11 + n 10 + n 01 + n 00
- posterior: P | n ~ D i r i c h l e t ( α 11 + n 11 , α 10 + n 10 , α 01 + n 01 , α 00 + n 00 )
3. Obtaining interesting measures
- support: S X Y
- confidence: C X Y
- lift: L X Y
4. Estimating mean and credible interval
- drawing samples from posterior distribution
- estimating probability values, expectation, and credible interval
In BPM, each document and keyword in the document–keyword matrix becomes a transaction and item, respectively. In BPM, we use the document–keyword matrix as the transaction–item matrix. The element f i j of this matrix has a value of 0 or 1, as follows.
f i j 0,1 ,   i = 1 ,   2 , n ,   j = 1 ,   2 , p
In Equation (4), n and p are the numbers of documents (transactions) and keywords (items), respectively. If the f i j value is 1, it means that the jth keyword is included in the ith document. If the antecedent and consequent of the association rule are represented as X and Y, respectively, the following partition table is created [35,37].
In Table 1, each item is a keyword of document–keyword matrix. Also, n i j represents the number of X = i and Y = j ( i , j 0,1 ). Let p i j be the probability of X = i and Y = j , P = p 11 , p 10 , p 01 , p 00 and i , j = 1 0 p i j = 1 . In our BPM, we use the Dirichlet distribution as a prior, P ~ D i r i c h l e t α 11 , α 10 , α 01 , α 00 , α i j > 0 in Equation (5).
f P = Γ α 11 + α 10 + α 01 + α 00 Γ α 11 Γ α 10 Γ α 01 Γ α 00 p 11 α 11 1 p 10 α 10 1 p 01 α 01 1 p 00 α 00 1 ,   p 11 , p 10 , p 01 , p 00 0
The likelihood function uses a multinomial distribution n | P ~ M u l t i n o m i a l ( N , P ) , N = n 11 + n 10 + n 01 + n 00 . In this study, we set the Dirichlet prior parameter α = 1 for all elements, corresponding to a uniform noninformative prior. This choice assumes equal prior probability across all possible rule outcomes, reflecting the absence of domain-specific prior knowledge. Such a setting prevents overfitting on rare keyword pairs and provides stable posterior estimates. Alternative α configurations can be employed when prior information about item co-occurrence is available, which will be explored in future work. Using the prior distribution and likelihood function, we derive the posterior distribution as Equation (6).
P | n ~ D i r i c h l e t ( α 11 + n 11 , α 10 + n 10 , α 01 + n 01 , α 00 + n 00 )
For each interestingness measure (support, confidence, and lift), posterior distributions were estimated using Markov Chain Monte Carlo (MCMC) sampling [41]. The posterior mean represents the expected value under the Bayesian model, while the credible interval corresponds to the 95% highest posterior density (HPD) region, calculated from the 2.5th and 97.5th percentiles of the posterior samples [41]. These credible intervals quantify the uncertainty of each measure and provide a probabilistic interpretation that extends beyond the point estimates obtained from conventional ARM. We derive the interesting measures of association rules from the contingency table and probability distribution as follows.
S X Y = p 11 ,   C X Y = p 11 p 11 + p 10 ,   L X Y = p 11 ( p 11 + p 10 ) ( p 11 + p 01 )
In Equation (7), S X Y , C X Y , and L X Y are support, confidence, and lift of association rules, respectively. We use samples drawn from the posterior distribution to obtain probability, expectation, and credible interval for each interesting measure. We show how to apply Bayesian inference to association rules with support, confidence, and lift. Lastly, we apply Bayesian inference to association rules, and estimate means and credible intervals of confidence and lift from the association rules in Algorithm 3.
Algorithm 3 Bayesian Inference for Association Rules
Input:
    - n 11 , n 10 , n 01 , n 00 of contingency table
    - prior, likelihood (M samples), posterior
- threshold t of confidence
Output:
    - summary statistics of confidence and lift
    - credible intervals of confidence and lift
    - probability values of P ( L X Y > 1 ) and P ( C X Y > t )
1. Obtaining parameters for posterior distribution
- α i j * = α i j + n i j for all elements
2. Drawing M samples from posterior distribution
- posterior: P ( m ) ~ D i r i c h l e t α i j * ,   α i j * > 0
3. Computing interesting measures per sample
- S X Y = p 11 ( m )
- C X Y = p 11 ( m ) p 11 ( m ) + p 10 ( m )
- L X Y = C X Y p 11 ( m ) + p 01 ( m )
4. Estimating summary statistics
- calculating posterior mean and 95% credible intervals
- estimating probability values of P ( L X Y > 1 )   and   P ( C X Y > t )
The input of Bayesian inference for association rules are n 11 , n 10 , n 01 , n 00 of contingency table and M samples drawn from posterior distribution. We determine threshold t of confidence measure in the input. Our output contains the summary statistics of confidence and lift such as mean and credible interval. The output also provides the probability values of P ( L X Y > 1 ) and P ( C X Y > t ) . Step 3 of Algorithm 3 represents how to compute the interesting measures per sample in total M samples. Using the M samples, we estimate the means and 95% credible intervals of confidence and lift. Using the results of Algorithms 1–3, we made a keyword diagram to understand the target domain.

4. Experimental Results

To evaluate the effectiveness of the proposed BPM, we collected patent documents related to quantum computing technology from two major international patent databases: the Korea Intellectual Property Rights Information Service (KIPRIS) and the United States Patent and Trademark Office (USPTO) [42,43]. The collection period spans from March 2015 to March 2025, during which we obtained a total of 9436 patent documents that were either filed or granted in the domain of quantum computing. Before conducting the analysis, we performed several preprocessing steps using text mining techniques. First, duplicate and incomplete documents were removed and missing values were imputed. Next, standard natural language processing procedures were applied, including tokenization, lowercasing, stemming, and lemmatization. Domain-specific stopwords (e.g., general legal terms in patents such as “system,” “method,” or “apparatus”) were also eliminated using a user-defined stopword dictionary. This process resulted in a clean corpus suitable for keyword extraction.
From the processed corpus, we extracted a total of 175 representative keywords. The selection of the 175 representative keywords was based on both frequency and domain relevance. First, only keywords with occurrence frequencies greater than approximately 1000 across the entire corpus were considered. Second, within this subset, we identified those that were relevant to quantum computing by consulting standard references in the field [44,45,46,47,48,49]. In particular, we relied on the introduction sections of these works, which provide an overview of the fundamental concepts and terminology in quantum computing. This ensured that selected keywords not only appeared frequently but also reflected core technical themes, including qubit, entanglement, gate, error correction, algorithms, and simulation. The final set was then categorized into three main technology domains of quantum computing: core, software, and hardware: core technologies (e.g., qubit, entanglement, gate), software and algorithms (e.g., encryption, program, network), and hardware components (e.g., chip, photon, superconduct). Using these keywords, we constructed a document–keyword matrix where rows represent documents and columns represent keywords. An important characteristic of the dataset is its sparsity. Because the number of keywords is much larger than the number of documents, most entries in the document–keyword matrix are zero. Specifically, over 92% of all matrix entries have zero values, indicating the presence of a severe zero-inflated problem. This sparsity highlights the limitations of traditional ARM methods and underscores the necessity of our proposed BPM approach, which incorporates credible intervals to reduce decision bias under zero inflation.
Quantum computing is a computational paradigm based on the principles of quantum mechanics using qubit to achieve speedups beyond classical bit-based computers [44,45,46,47,48,49]. Whereas classical computers represent and process information using bits that take values of either 0 or 1, quantum computers employ qubits that can exist in a superposition of both 0 and 1 simultaneously [46,48]. This enables the solution of certain computational problems with significantly higher efficiency and speed compared to classical computers. So, we analyzed the patents to build technology structure of quantum computing. Figure 3 shows the constructed patent document–keyword matrix.
In Figure 3, the keywords and frequency values of all keywords extracted from patent documents are showed. Since the word endings were removed during the preprocessing of text mining, the keywords shown in Figure 3 represent only the root of each word. For ARM analysis, we converted the count values in this matrix to binary data. That is, all values greater than or equal to 1 were converted to 1. If a specific keyword was included in each patent document, it was represented as 1, otherwise 0. Figure 4 shows resulting binary patent document–keyword matrix.
Among all keywords, we selected the final keywords according to the three technology categories of quantum computing: core, software, and hardware, as shown in Table 2.
According to the keywords, there are also representative keywords that are included in two categories at the same time. For example, the keyword algorithm falls into both the core and software categories. Among them, we selected 25 keywords as follows; atom, bit, calculation, chip, cloud, code, computing, connect, data, electron, encryption, hybrid, inform, logic, memory, network, photon, processor, program, quantum, qubit, random, security, superconduct, and vector. We first show the performance results of ARM in Table 3.
Table 3 shows the top 30 in descending order of confidence. We made it possible to compare not only the confidence value of each rule, but also the values of lift and support measures. The left-hand side (LHS) and right-hand side (RHS) are the antecedent and consequent items. From the results in Table 3, we see that rules b i t q u a n t u m and p h o t o n q u a n t u m are the largest, followed by c h i p q u a n t u m and a t o m q u a n t u m . For example, the probability of quantum appearing under the condition that the keyword photon appears is 0.9806. This means that knowledge of photons is essential for quantum technology development. Additionally, the lift value of rule c h i p q u a n t u m was 1.0773, which was greater than 1. This means that chip and quantum technologies are complementary to each other. We also found that the support value of rule q u b i t q u a n t u m is 0.2134, which means that the keywords qubit and quantum appeared together in 21.34% of all patent documents. Next, we show the top 30 ARM rules by lift measure in Table 4.
In Table 4, we found that the lift value of ARM rules e n c r y p t i o n s e c u r i t y and s e c u r i t y e n c r y p t i o n are the largest. The lift value is 5.9356. This indicates a very strong interdependence between the two technologies based on keywords encryption and security. This means that as the technology based on encryption is developed, technology based on security is also developed accordingly. The interpretation and use of confidence and support in Table 4 are the same as in Table 3.
As confirmed in Table 3 and Table 4, ARM provides only one value for each of the three measures—confidence, lift, and support. Therefore, we must rely on a single value for decision-making. In analyzing sparse data with zero inflation, the decision process can lead to bias. So, to overcome this problem, we proposed a method of extracting association rules, called BPM. Table 5 shows the top 30 association rules of BPM by confidence measure.
In Table 5, the confidence measure provides three values such as mean, lower, and upper. The mean of confidence in Table 5 is the same concept as the confidence of ARM in Table 3. The BPM enables us to make unbiased decisions about association rules because it provides credible intervals for confidence that ARM does not. In this paper, we obtained a 95% credible interval. We also illustrate the top 30 rules of BPM by lift measure in Table 6.
Similar to the results of confidence in Table 5, Table 6 presents the mean of the lift measure and the lower and upper limits of the 95% credible interval. We found that the lift values for rules, e n c r y p t i o n s e c u r i t y and s e c u r i t y e n c r y p t i o n were identical, with a mean of 5.9397. This result was similar to that in Table 4. In addition, Table 6 provided the lower and upper values of the credible interval, 5.5412 and 6.3394, as well as the mean lift. In other words, the keywords encryption and security have a highly complementary relationship, with lift values ranging from a minimum of 5.5412 to a maximum of 6.3394 at the 95% confidence level, thereby enabling efficient decision-making. Using the results in Table 5, we constructed the keyword diagram shown in Figure 5.
From the results of Figure 5, we verified that all keywords influence the keywords quantum and computing. Also, keyword computing affects keyword quantum. All arrows have a mean and credible interval of the confidence measure. For example, the confidence value of association rule c o m p u t i n g q u a n t u m is 0.9401, and its credible interval is the bounds between 0.9347 and 0.9455. Next, we show another keyword diagram by lift measure in Figure 6.
From the results in Figure 6, we can see that there are four keyword components for quantum computing. The smallest component is the rule consisting of two keywords (logic and qubit), with a lift value of 2.0029, and the lower and upper bounds of the credible interval from 1.8417 to 2.1603, respectively. In this paper, we used the R data language and its packages to perform our experiments [50,51].

5. Discussion and Implications

5.1. Theoretical Implications

The experimental results of this study provide several important insights into both the theoretical development of text mining methodologies and their practical applications in real-world contexts. In this section, we discuss the implications of our proposed BPM method, highlighting how it differs from existing approaches and how it contributes to the field of keyword data analysis. This study makes a methodological contribution by extending the classical framework of ARM with Bayesian inference. While ARM has been widely used to extract frequent co-occurrence patterns in document–keyword matrices [35,36,37], it is limited in that it provides only single-value measures of support, confidence, and lift. Such point estimates are particularly vulnerable to bias and instability in sparse and zero-inflated datasets, where the majority of matrix entries are zero [3,20]. By contrast, BPM introduces a probabilistic framework in which these measures are derived from posterior distributions, enabling the estimation of both mean values and credible intervals. This extension allows analysts to incorporate uncertainty directly into decision-making, thus reducing the likelihood of biased interpretations. To our knowledge, this is one of the first attempts to explicitly integrate Bayesian credible intervals into rule-based keyword analysis. In this respect, BPM not only strengthens the theoretical underpinnings of association rule mining but also bridges the gap between statistical zero-inflated modeling [16,18,21] and text mining approaches.

5.2. Practical Implications

From a practical standpoint, BPM provides an effective tool for analyzing sparse keyword datasets across a variety of application domains. In our case study of 9436 quantum computing patent documents collected from KIPRIS and USPTO, BPM revealed reliable associations among 175 extracted keywords. These results can inform technology forecasting, competitive intelligence, and strategic planning by identifying stable relationships among emerging technologies. Unlike embedding-based models (e.g., Word2Vec, BERT) that require large-scale corpora and heavy computation [10,34], BPM can be applied effectively in situations where datasets are small, domain-specific, or highly sparse. Moreover, BPM does not require training large neural architectures, which makes it more accessible to practitioners in fields such as intellectual property management, R&D strategy, and policymaking. By providing credible intervals for association measures, BPM also enhances interpretability, allowing decision-makers to gauge the robustness of results before acting on them.

5.3. Differentiation from Existing Work

This research also differs from existing studies in important ways. Traditional frequency-based methods such as TF–IDF [25,27] or co-word analysis capture keyword salience but do not quantify the uncertainty associated with sparsity. Recent advances using semantic embeddings or LLMs [10,34] provide richer representations of meaning but often fail to address the structural sparsity of document–keyword matrices. Furthermore, prior zero-inflated modeling research has been primarily applied to regression analysis or biomedical count data [16,18,21], rather than to association rules in text mining. By directly embedding Bayesian inference into ARM, our BPM approach introduces a unique mechanism for mitigating decision bias under sparsity. This methodological innovation positions BPM as a complementary tool that can coexist with semantic and embedding-based methods. In future work, BPM could be integrated with semantic embeddings to leverage both probabilistic inference and semantic representation, thereby enhancing the reliability and applicability of keyword-based text mining across diverse domains.
In this study, we focused primarily on single-word keywords in constructing the document–keyword matrix. The motivation for this choice is that our goal is not to perform topic modeling or semantic phrase extraction, but to evaluate associative relationships between items (keywords) using confidence and lift measures. Nevertheless, the proposed BPM framework is not restricted to single terms. Since n-grams such as bigrams and trigrams (e.g., quantum computing, quantum entanglement, quantum cryptography) can also be represented as items in the transaction–item matrix, our method can be extended to incorporate them. Future research will consider this extension to enhance the discriminatory power of the analysis.
Although the extracted rules were ranked quantitatively according to confidence and lift measures (Table 5 and Table 6), this study did not incorporate external expert validation or semantic comparison for evaluating the relevance of the extracted rules. The purpose of our work was to introduce a probabilistic framework that quantifies uncertainty in rule strength through Bayesian credible intervals, providing an internal measure of reliability rather than an external relevance assessment. Nevertheless, integrating expert judgment or external validation, for example, comparing rule patterns with domain knowledge or large language model outputs, represents an important direction for future research. This combination of statistical and semantic validation could further enhance the interpretability and applicability of BPM.

6. Conclusions

In this paper, we proposed BPM as a novel approach for analyzing sparse and zero-inflated keyword data. Unlike traditional ARM, which provides only point estimates of support, confidence, and lift, BPM incorporates Bayesian inference to estimate both the mean values and credible intervals of these measures. This dual estimation enables more robust and unbiased decision-making in the presence of sparsity, which is a critical limitation of conventional keyword analysis methods. To verify the effectiveness of our approach, we applied BPM to a dataset of 9436 patent documents related to quantum computing collected from KIPRIS and USPTO over a ten-year period (2015–2025). From these documents, 175 representative keywords were extracted after preprocessing, and a highly sparse document–keyword matrix was constructed. Our experimental results showed that BPM can provide interpretable and stable associations between keywords, with credible intervals that reduce the risk of biased interpretation under zero-inflated conditions.
This study contributes to the literature in three ways: (i) it introduces a Bayesian extension of ARM that addresses the zero-inflated sparsity problem in keyword data, (ii) it demonstrates the empirical utility of BPM through a large-scale patent analysis in the field of quantum computing, and (iii) it highlights the theoretical and practical implications of incorporating uncertainty into rule-based keyword analysis. Nevertheless, some limitations remain.
Furthermore, we acknowledge that the empirical validation used ARM as the only baseline, since BPM was developed as a direct Bayesian extension of ARM. However, to strengthen future evaluation, BPM will be compared with embedding-based and LLM-based approaches that provide semantic representations of keyword relationships. Combining semantic embedding with Bayesian inference may enable a more comprehensive framework that captures both semantic proximity and probabilistic uncertainty, thereby enhancing interpretability and predictive capability.
This study focused on improving ARM and therefore used ARM as the primary baseline. While comparisons with state-of-the-art semantic or embedding-based approaches such as TF–IDF, word embeddings, or LLMs were not conducted, future work will explore integrating BPM with such methods to combine probabilistic inference and semantic representation. We also plan to extend BPM to other domains, including social media and biomedical text, to further demonstrate its versatility.
In conclusion, BPM provides a probabilistic framework that enhances the interpretability and reliability of keyword data analysis. By explicitly addressing sparsity and zero-inflated problems, our approach lays a foundation for future studies that seek to combine Bayesian inference with modern text mining techniques to achieve more comprehensive and trustworthy insights.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author the data having been purchased and being currently used for research projects.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Dai, Z.; Zhao, X.; Cui, B. TFIDF Text Keyword Mining Method Based on Hadoop Distributed Platform Under Massive Data. In Proceedings of the 2024 IEEE 2nd International Conference on Image Processing and Computer Applications (ICIPCA), Shenyang, China, 28–30 June 2024; pp. 1844–1848. [Google Scholar]
  2. Hu, H.; Chen, J.; Hu, H. Digital Trade Related Policy Text Classification and Quantification Based on TF-IDF Keyword Algorithm. In Proceedings of the 2024 International Symposium on Intelligent Robotics and Systems (ISoIRS), Changsha, China, 14–16 June 2024; pp. 284–288. [Google Scholar]
  3. Jun, S. Patent Keyword Analysis Using Bayesian Zero-Inflated Model and Text Mining. Stats 2024, 7, 827–841. [Google Scholar] [CrossRef]
  4. Jun, S. Zero-Inflated Text Data Analysis using Generative Adversarial Networks and Statistical Modeling. Computers 2023, 12, 258. [Google Scholar] [CrossRef]
  5. Kim, J.-M.; Jun, S. Graphical causal inference and copula regression model for apple keywords by text mining. Adv. Eng. Inform. 2015, 29, 918–929. [Google Scholar] [CrossRef]
  6. Singh, S.; Gupta, S.; Singh, V.; Narmadha, T.; Karthikeyan, K.; Chavan, G.T. Text Mining for Knowledge Discovery and Information Analysis. In Proceedings of the 2024 15th International Conference on Computing Communication and Networking Technologies (ICCCNT), Kamand, India, 24–28 June 2024; p. 61001. [Google Scholar]
  7. Xue, D.; Shao, Z. Patent text mining based hydrogen energy technology evolution path identification. Int. J. Hydrogen Energy 2024, 49, 699–710. [Google Scholar] [CrossRef]
  8. Bzhalava, L.; Kaivo-oja, J.; Hassan, S.S. Digital business foresight: Keyword-based analysis and CorEx topic modeling. Futures 2024, 155, 103303. [Google Scholar] [CrossRef]
  9. Feng, L.; Niu, Y.; Liu, Z.; Wang, J.; Zhang, K. Discovering Technology Opportunity by Keyword-Based Patent Analysis: A Hybrid Approach of Morphology Analysis and USIT. Sustainability 2020, 12, 136. [Google Scholar] [CrossRef]
  10. Chen, G.; Hong, S.; Du, C.; Wang, P.; Yang, Z.; Xiao, L. Comparing semantic representation methods for keyword analysis in bibliometric research. J. Informetr. 2024, 18, 101529. [Google Scholar] [CrossRef]
  11. Jain, K.; Srivastava, M. Which Technologies Will Drive the Battery Electric Vehicle Industry?: A Keyword Network Based Roadmapping. In Proceedings of the 2024 Portland International Conference on Management of Engineering and Technology (PICMET), Portland, OR, USA, 4–8 August 2024; pp. 1–9. [Google Scholar]
  12. Jun, S. Keyword Data Analysis Using Generative Models Based on Statistics and Machine Learning Algorithms. Electronics 2024, 13, 798. [Google Scholar] [CrossRef]
  13. Lee, J.; Jeong, H. Keyword analysis of artificial intelligence education policy in South Korea. IEEE Access 2023, 11, 102408–102417. [Google Scholar] [CrossRef]
  14. Shin, H.; Lee, H.J.; Cho, S. General-use unsupervised keyword extraction model for keyword analysis. Expert Syst. Appl. 2023, 233, 120889. [Google Scholar] [CrossRef]
  15. de Souza, H.C.C.; Louzada, F.; Ramos, P.L.; de Oliveira Júnior, M.R.; Perdoná, G.D.S.C.A. Bayesian approach for the zero-inflated cure model: An application in a Brazilian invasive cervical cancer database. J. Appl. Stat. 2022, 49, 3178–3194. [Google Scholar] [CrossRef] [PubMed]
  16. Hwang, B.S. A Bayesian joint model for continuous and zero-inflated count data in developmental toxicity studies. Commun. Stat. Appl. Methods 2022, 29, 239–250. [Google Scholar] [CrossRef]
  17. Lee, K.H.; Coull, B.A.; Moscicki, A.B.; Paster, B.J.; Starr, J.R. Bayesian variable selection for multivariate zero-inflated models: Application to microbiome count data. Biostatistics 2020, 21, 499–517. [Google Scholar] [CrossRef] [PubMed]
  18. Lu, L.; Fu, Y.; Chu, P.; Zhang, X. A Bayesian Analysis of Zero-Inflated Count Data: An Application to Youth Fitness Survey. In Proceedings of the Tenth International Conference on Computational Intelligence and Security, Kunming, China, 15–16 November 2014; pp. 699–703. [Google Scholar]
  19. Neelon, B.; Chung, D. The LZIP: A Bayesian Latent Factor Model for Correlated Zero-Inflated Counts. Biometrics 2017, 73, 185–196. [Google Scholar] [CrossRef]
  20. Park, S.; Jun, S. Zero-Inflated Patent Data Analysis Using Compound Poisson Models. Appl. Sci. 2023, 13, 4505. [Google Scholar] [CrossRef]
  21. Sidumo, B.; Sonono, E.; Takaidza, I. Count Regression and Machine Learning Techniques for Zero-Inflated Overdispersed Count Data: Application to Ecological Data. Ann. Data Sci. 2024, 11, 803–817. [Google Scholar] [CrossRef]
  22. Wanitjirattikal, P.; Shi, C. A Bayesian zero-inflated binomial regression and its application in dose-finding study. J. Biopharm. Stat. 2020, 30, 322–333. [Google Scholar] [CrossRef]
  23. Young, D.S.; Roemmele, E.S.; Yeh, P. Zero-inflated modeling part I: Traditional zero-inflated count regression models, their applications, and computational tools. WIREs Comput. Stat. 2022, 14, e1541. [Google Scholar] [CrossRef]
  24. Feinerer, I.; Hornik, K. Package ‘tm’ Version 0.7-16, Text Mining Package; CRAN of R Project; R Foundation for Statistical Computing: Vienna, Austria, 2025. [Google Scholar]
  25. Salton, G.; Buckley, C. Term-weighting approaches in automatic text retrieval. Inf. Process. Manag. 1988, 24, 513–523. [Google Scholar] [CrossRef]
  26. Hogg, R.V.; McKean, J.M.; Craig, A.T. Introduction to Mathematical Statistics, 8th ed.; Pearson: Upper Saddle River, NJ, USA, 2018. [Google Scholar]
  27. Ramos, J. Using tf-idf to determine word relevance in document queries. In Proceedings of the First Instructional Conference on Machine Learning, Bangalore, India, 21–23 October 2021; Volume 242, pp. 29–48. [Google Scholar]
  28. Sucar, L.E. Probabilistic Graphical Models Principles and Applications; Springer: New York, NY, USA, 2015. [Google Scholar]
  29. Yang, X.; Sun, B.; Liu, S. Study of technology communities and dominant technology lock-in in the Internet of Things domain—Based on social network analysis of patent network. Inf. Process. Manag. 2025, 62, 103959. [Google Scholar] [CrossRef]
  30. Lu, Y.; Zheng, Q.; Quinn, D. Introducing Causal Inference Using Bayesian Networks and do-Calculus. J. Stat. Data Sci. Educ. 2023, 31, 3–17. [Google Scholar] [CrossRef]
  31. Park, S.; Jun, S. Patent Analysis Using Bayesian Data Analysis and Network Modeling. Appl. Sci. 2022, 12, 1423. [Google Scholar] [CrossRef]
  32. Workie, M.S.; Azene, A.G. Bayesian zero-inflated regression model with application to under-five child mortality. J. Big Data 2021, 8, 4. [Google Scholar] [CrossRef]
  33. Roper, A.T.; Cunningham, S.W.; Porter, A.L.; Mason, T.W.; Rossini, F.A.; Banks, J. Forecasting and Management of Technology; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  34. Hu, M.; Mu, Y.; Jin, H. A bibliometric analysis of advances in CO2 reduction technology based on patents. Appl. Energy 2025, 382, 125193. [Google Scholar] [CrossRef]
  35. Agrawal, R.; Imieliński, T.; Swami, A. Mining association rules between sets of items in large databases. In Proceedings of the ACM SIGMOD International Conference on Management of Data, Washington, DC, USA, 1 June 1993; pp. 207–216. [Google Scholar]
  36. Han, J.; Kamber, M.; Pei, J. Data Mining: Concepts and Techniques, 3rd ed.; Morgan Kaufmann: Waltham, MA, USA, 2012. [Google Scholar]
  37. Tan, P.; Kumar, V.; Srivastava, J. Selecting the right interestingness measure for association patterns. In Proceedings of the eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Edmonton, Alberta, Canada, 23–26 July 2002; pp. 32–41. [Google Scholar]
  38. Campello, R.J.G.B.; Moulavi, D.; Sander, J. Density-Based Clustering Based on Hierarchical Density Estimates. In Advances in Knowledge Discovery and Data Mining, Proceedings of the PAKDD 2013, Gold Coast, Australia, 14–17 April 2013; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2013; Volume 7819, pp. 160–172. [Google Scholar]
  39. Aljalbout, E.; Golkov, V.; Siddiqui, Y.; Strobel, M.; Cremers, D. Clustering with Deep Learning: Taxonomy and New Methods. arXiv 2018, arXiv:1801.07648. [Google Scholar] [CrossRef]
  40. Beltagy, I.; Lo, K.; Cohan, A. SCIBERT: A Pretrained Language Model for Scientific Text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 3615–3620. [Google Scholar]
  41. Efron, B.; Hastie, T. Computer Age Statistical Inference; Cambridge University Press: New York, NY, USA, 2016. [Google Scholar]
  42. KIPRIS. Korea Intellectual Property Rights Information Service. Available online: www.kipris.or.kr (accessed on 1 March 2025).
  43. USPTO. The United States Patent and Trademark Office. Available online: http://www.uspto.gov (accessed on 15 March 2025).
  44. Bernhardt, C. Quantum Computing for Everyone; MIT Press: Cambridge, MA, USA, 2020. [Google Scholar]
  45. Blais, A.; Grimsmo, A.L.; Girvin, S.M.; Wallraff, A. Circuit quantum electrodynamics. Rev. Mod. Phys. 2021, 93, 025005. [Google Scholar] [CrossRef]
  46. Lee, P.Y.; Ji, H.; Cheng, R. Quantum Computing and Information: A Scaffolding Approach, 2nd ed.; Polaris QCI Publishing: Middletown, NY, USA, 2025. [Google Scholar]
  47. Montanaro, A. Quantum Algorithms: An Overview. NPJ Quantum Inf. 2016, 2, 15023. [Google Scholar] [CrossRef]
  48. Nielsen, M.A.; Chuang, I.L. Quantum Computation and Quantum Information, 10th ed.; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
  49. Preskill, J. Quantum computing in the NISQ era and beyond. Quantum 2018, 2, 79. [Google Scholar] [CrossRef]
  50. R Development Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2024; Available online: http://www.R-project.org (accessed on 1 March 2025).
  51. Hahsler, M.; Buchta, C.; Gruen, B.; Hornik, K.; Borgelt, C.; Johnson, I.; Ledmi, M. Package ‘arules’ 1.7-11, Mining Association Rules and Frequent Itemsets; CRAN of R Project; R Foundation for Statistical Computing: Vienna, Austria, 2025. [Google Scholar]
Figure 1. Text preprocessing and mining of document data.
Figure 1. Text preprocessing and mining of document data.
Computers 14 00436 g001
Figure 2. Transaction–item matrix.
Figure 2. Transaction–item matrix.
Computers 14 00436 g002
Figure 3. Patent document–keyword matrix (part of the entire data).
Figure 3. Patent document–keyword matrix (part of the entire data).
Computers 14 00436 g003
Figure 4. Binary patent document–keyword matrix (part of the entire data).
Figure 4. Binary patent document–keyword matrix (part of the entire data).
Computers 14 00436 g004
Figure 5. Keyword diagram by confidence measure of BPM.
Figure 5. Keyword diagram by confidence measure of BPM.
Computers 14 00436 g005
Figure 6. Keyword diagram by lift measure of BPM.
Figure 6. Keyword diagram by lift measure of BPM.
Computers 14 00436 g006
Table 1. Contingency table of antecedent and consequent items.
Table 1. Contingency table of antecedent and consequent items.
Item Y = 1 Y = 0 Row Sum
X = 1 n 11 n 10 n 1 ·
X = 0 n 01 n 00 n 0 ·
Column sum n · 1 n · 0 Total sum = N
Table 2. Representative keywords related to quantum computing.
Table 2. Representative keywords related to quantum computing.
Technology CategoryRepresentative Keywords
Corealgorithm, bit, calcul, circuit, comput, entangl, error, gate,
Hamiltonian, ion, measur, photon, quantum, qubit, sequenc,
signal, simul, state, superconduct, timetrap,
Softwarealgorithm, authent, circuit, cloud, code, connect, data, electron,
encrypt, execut, function, gate, graph, implement, includ, inform, key, logic, model, modul, network, optim, perform, processor,
program, random, sampl, secret, secur, simul, structur, technolog,
Hardwarechip, electron, energi, ion, memori, optic, optical, photon, reson, storage, substrat, superconduct, trap
Table 3. Top 30 rules of ARM by confidence measure.
Table 3. Top 30 rules of ARM by confidence measure.
LHSRHSConfidenceLiftSupport
bitquantum0.98091.07770.1271
photonquantum0.98061.07730.0379
chipquantum0.97661.07300.0761
atomquantum0.97141.06730.0255
superconductquantum0.96591.06120.0669
processorquantum0.95961.05430.0995
qubitquantum0.94941.04310.2134
logicquantum0.94931.04300.0643
programquantum0.94891.04250.0717
hybridquantum0.94241.03540.0491
computingquantum0.94021.03290.7622
electronquantum0.93301.02510.0792
codequantum0.92661.01810.0583
hybridcomputing0.92391.13960.0482
connectquantum0.92371.01480.0870
chipcomputing0.91181.12480.0710
memoryquantum0.90000.98880.0454
informquantum0.89970.98850.1493
networkquantum0.89860.98720.1150
programcomputing0.89491.10380.0676
calculationquantum0.89190.98000.1603
cloudcomputing0.88571.09250.0299
vectorquantum0.88340.97060.0325
cloudquantum0.87620.96270.0296
atomcomputing0.87351.07740.0230
superconductcomputing0.86981.07280.0602
dataquantum0.86830.95400.2066
processorcomputing0.86541.06750.0897
electroncomputing0.85461.05420.0725
qubitcomputing0.85351.05270.1919
Table 4. Top 30 rules of ARM by lift measure.
Table 4. Top 30 rules of ARM by lift measure.
LHSRHSLiftConfidenceSupport
encryptionsecurity5.93560.55670.0411
securityencryption5.93560.43820.0411
randomencryption4.99030.36840.0203
encryptionrandom4.99030.27470.0203
processormemory4.90560.24740.0256
memoryprocessor4.90560.50850.0256
atomphoton4.33200.16730.0044
photonatom4.33200.11390.0044
superconductchip4.17920.32560.0225
chipsuperconduct4.17920.28930.0225
encryptioncloud3.69800.12500.0092
cloudencryption3.69800.27300.0092
randomsecurity3.63730.34110.0188
securityrandom3.63730.20020.0188
cloudrandom3.05650.16830.0057
randomcloud3.05650.10330.0057
memorylogic2.89090.19570.0099
logicmemory2.89090.14580.0099
chipconnect2.88010.27130.0211
connectchip2.88010.22440.0211
securitycloud2.87720.09730.0091
cloudsecurity2.87720.26980.0091
chipbit2.77570.35950.0280
bitchip2.77570.21620.0280
processorhybrid2.64000.13770.0143
hybridprocessor2.64000.27370.0143
superconductconnect2.60000.24500.0170
connectsuperconduct2.60000.18000.0170
memoryprogram2.39400.18090.0091
programmemory2.39400.12070.0091
Table 5. Top 30 rules of BPM by confidence measure.
Table 5. Top 30 rules of BPM by confidence measure.
LHSRHSCredible Interval of Confidence MeasureLiftSupport
MeanLowerUpperMeanMean
bitquantum0.98040.97190.98821.07730.1271
photonquantum0.97930.96310.99101.07600.0379
chipquantum0.97600.96360.98551.07240.0761
atomquantum0.96960.94430.98791.06540.0255
superconductquantum0.96520.94960.97781.06050.0669
processorquantum0.95900.94590.97011.05380.0995
qubitquantum0.94930.93920.95841.04300.2134
logicquantum0.94870.93120.96471.04250.0643
programquantum0.94800.93030.96331.04170.0717
hybridquantum0.94160.91920.96021.03450.0491
computingquantum0.94010.93470.94551.03300.7622
electronquantum0.93270.91590.94921.02500.0792
codequantum0.92610.90530.94431.01750.0583
hybridcomputing0.92310.89770.94541.13880.0482
connectquantum0.92300.90470.93991.01420.0870
chipcomputing0.91170.89040.93121.12460.0710
informquantum0.89930.88360.91370.98830.1493
memoryquantum 0.89920.87100.92430.98790.0454
networkquantum0.89830.88140.91520.98700.1150
programcomputing0.89470.87160.91551.10360.0676
calculationquantum0.89170.87640.90650.97990.1603
cloudcomputing0.88390.84660.91711.09030.0299
vectorquantum0.88250.84500.91450.96960.0325
cloudquantum0.87450.83620.90750.96080.0296
atomcomputing0.87210.82820.91041.07570.0230
superconductcomputing0.86930.84130.89441.07210.0602
dataquantum0.86800.85400.88190.95390.2066
processorcomputing0.86520.84160.88631.06740.0897
electroncomputing0.85380.82820.87801.05310.0725
qubitcomputing0.85320.83730.86791.05260.1919
Table 6. Top 30 rules of BPM by lift measure.
Table 6. Top 30 rules of BPM by lift measure.
LHSRHSCredible Interval of Lift MeasureConfidenceSupport
MeanLowerUpperMeanMean
securityencryption5.93975.54126.33940.43800.0411
encryptionsecurity5.93125.53996.35630.55680.0411
encryptionrandom4.98074.47135.51690.27440.0203
randomencryption4.97684.45995.51090.36800.0203
processormemory4.90914.49425.34580.24780.0256
memoryprocessor4.90094.48585.31850.50860.0256
chipsuperconduct4.18673.76004.61490.29020.0225
superconductchip4.18303.75224.60940.32630.0225
securityrandom3.64743.25064.05970.20120.0188
randomsecurity3.63093.23224.04810.34060.0188
connectchip2.87652.58293.19060.22440.0211
chipconnect2.87432.57063.21140.27080.0211
bitchip2.77642.52423.03260.21640.0280
chipbit2.77542.53633.02490.35980.0280
processorhybrid2.64582.29263.01250.13820.0143
hybridprocessor2.64242.29012.99870.27410.0143
superconductconnect2.60762.30192.94250.24600.0170
connectsuperconduct2.60542.29292.94120.18060.0170
datacloud2.32392.08222.55320.07880.0187
clouddata2.32112.09412.54950.55200.0187
networksecurity2.29422.07812.51220.21540.0275
securitynetwork2.29032.08672.50470.29300.0275
encryptiondata2.11501.96422.27030.50330.0371
dataencryption2.11291.95852.26960.15610.0371
logicqubit2.00291.84172.16030.45040.0305
qubitlogic2.00051.83662.17060.13570.0305
securityinform1.91951.75002.09270.31840.0298
informsecurity1.91751.74362.09180.18000.0298
encryptionnetwork1.91011.67622.15740.24450.0180
networkencryption1.90851.68222.14420.14120.0180
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jun, S. Sparse Keyword Data Analysis Using Bayesian Pattern Mining. Computers 2025, 14, 436. https://doi.org/10.3390/computers14100436

AMA Style

Jun S. Sparse Keyword Data Analysis Using Bayesian Pattern Mining. Computers. 2025; 14(10):436. https://doi.org/10.3390/computers14100436

Chicago/Turabian Style

Jun, Sunghae. 2025. "Sparse Keyword Data Analysis Using Bayesian Pattern Mining" Computers 14, no. 10: 436. https://doi.org/10.3390/computers14100436

APA Style

Jun, S. (2025). Sparse Keyword Data Analysis Using Bayesian Pattern Mining. Computers, 14(10), 436. https://doi.org/10.3390/computers14100436

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop