Next Article in Journal
Large-Scale Screening and Machine Learning to Predict the Computation-Ready, Experimental Metal-Organic Frameworks for CO2 Capture from Air
Previous Article in Journal
Early Intervention Using Fecal Microbiota Transplantation Combined with Probiotics Influence the Growth Performance, Diarrhea, and Intestinal Barrier Function of Piglets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Patent Data Analysis of Artificial Intelligence Using Bayesian Interval Estimation

1
Department of Mathematics, University of Arkansas, Fort Smith, AR 72913, USA
2
Department of Big Data and Statistics, Cheongju University, Chungbuk 28503, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(2), 570; https://doi.org/10.3390/app10020570
Submission received: 15 December 2019 / Revised: 2 January 2020 / Accepted: 10 January 2020 / Published: 13 January 2020

Abstract

:
Technology analysis is one of the important tasks in technology and industrial management. Much information about technology is contained in the patent documents. So, patent data analysis is required for technology analysis. The existing patent analyses relied on the quantitative analysis of the collected patent documents. However, in the technology analysis, expert prior knowledge should also be considered. In this paper, we study the patent analysis method using Bayesian inference which considers prior experience of experts and likelihood function of patent data at the same time. For keyword data analysis, we use Bayesian predictive interval estimation with count data distributions such as Poisson. Using the proposed models, we forecast the future trends of technological keywords of artificial intelligence (AI) in order to know the future technology of AI. We perform a case study to provide how the proposed method can be applied to real areas. In this paper, we retrieve the patent documents related to AI technology, and analyze them to find the technological trend of AI. From the results of AI technology case study, we can find which technological keywords are more important or critical in the entire structure of AI industry. The existing methods for patent keyword analysis were depended on the collected patent documents at present. But, in technology analysis, the prior knowledge by domain experts is as important as the collected patent documents. So, we propose a method based on Bayesian inference for technology analysis using the patent documents. Our method considers the patent data analysis with the prior knowledge from domain experts.

1. Introduction

Many researches related to patent data analysis have been performed in various fields of management of technology (MOT) [1,2,3,4,5]. They analyzed the patent documents for technology transfer, new product development, technology forecasting, etc. A patent contains detailed information on the technology developed, including the title and abstract of the invention, claims, dates of filing and registration, citation information, technological classification codes, names of inventor and applicant, and so forth [6]. Because of this characteristic of patent, many researchers have carried out technology analysis on their fields using patent document data [7]. In addition, they proposed diverse methods for technology analysis using patent data. Kim et al. (2019) studied on a statistical method based on functional count data modeling for patent data analysis [8]. They extracted technological keywords from the patent documents applied by Apple company and analyzed them by functional count data models based on Poisson, negative binomial, and hurdle Poisson distributions [8]. Using the result of their patent analysis, they found technological structure of Apple for understanding the innovation and evolution of Apple’s technologies [8]. Uhm el al. (2017) also proposed a statistical inference for the patent keyword analysis [9]. They estimated the confidence intervals of extracted keywords from the patent documents [9]. In addition, they provided the sustainable technology areas of given technology domain using the results of the keyword interval estimations [9]. Most of the previous studies for patent analysis were the results of objective and quantitative analysis using only the collected patent data. If the traditional technology analysis process can reflect the technological knowledge provided by the expert group of given technological field, we can expect a more reliable and applicable analysis results. In this paper, we use the Bayesian inference to propose a methodology for patent analysis that reflects the expert group’s technological knowledge of given technology domain. Statistics performs data analysis on two theories based on frequentists and Bayesian [10]. For data analysis, the frequentist carries out statistical analysis by only the likelihood of given data. In contrast, the Bayesian performs the statistical analysis by combining the likelihood of given data with the prior information of the data domain [11]. In our research, we analyze the patent data using Bayesian inference approach. This is because the Bayesian analysis method can add the experts’ knowledge to the model by using the prior distribution.
Also, we collect the patent documents related to artificial intelligence (AI) for our technology analysis. This is because AI has a lot of influence on our society as well as technology. PwC insists that AI is an important factor for companies to stay competitive in a rapidly changing economic environment [12]. In recent years, AI has been rapidly expanding to the areas that require delicate human skills combined with cognitive science and technology [13]. Most countries consider machine learning, deep learning, natural language processing, and computer vision as their AI technology portfolio [14]. Among them, China is the most active in the development of artificial intelligence technology [14]. Understanding AI through patent data analysis related to AI technology is an important task in MOT. Therefore, we propose a patent analysis method based on Bayesian interval estimation for understanding AI technology.
We first retrieve the AI patents from the patent databases in the world. Next, using the preprocessing by text mining methods, we extract the technological keywords and construct a structured data for our Bayesian data analysis. Finally, we analyze the structured AI patent data using Bayesian interval estimation for technology forecasting of AI. We organized our paper as follows. Section 2 introduces the patent analysis and technology forecasting. We propose our Bayesian interval estimation for patent keyword analysis in Section 3. In Section 4, we illustrate a case study with the technological keywords of AI technology. Lastly, we show our conclusions in Section 5.

2. Patent Analysis and Technology Forecasting

The patent system in the world encourages the research and development of technology by protecting the inventors’ exclusive rights to registered patents for a certain period of time [7]. So, a patent contains various and detailed information of developed technology [6]. We can get valuable knowledge for technology management from the results of patent analysis. Patent analysis is to analyze the data contained in patent documents such as title of invention, abstract, claims, citations, international patent classification (IPC) codes applied, and registered dates, etc. Kim et al. (2019) studied on a method for patent keyword analysis using time series and copular models [15]. They extracted the technological keywords from the searched patent documents and analyzed them for technology forecasting [15]. Kim et al. (2018) analyzed the IPC codes extracted from the patent documents related to AI [16]. They also used Bayesian regression and social networks analysis for the patent IPC codes analysis [16].
Technology forecasting is to forecast the future states of target technology [9]. Uhm et al. (2017) proposed an interval estimation for patent keyword analysis [9]. According to the change of intervals of patent keywords, they found technological structure of target technology [9]. This paper only focused on the collected patent data. That is, the authors did not consider the experience and knowledge of domain experts. But, in most technology analysis based on patent documents, the prior knowledge of domain experts is important to understand and analyze the technology. To solve this problem of previous researches, we propose a new method for patent keyword analysis using Bayesian prediction interval estimation.

3. Bayesian Prediction Interval Estimation for Patent Data Analysis

The frequentist estimates the model parameters using only the observed data. But, the Bayesian adds the prior belief in data to the observed data for the model parameter estimating [17]. The main difference between frequentist and Bayesian is the use of prior belief in data for statistical inference such as estimation, testing, modeling, etc. [10]. Because patent data analysis is an analysis of technology, more explanatory results can be expected when the expert’s knowledge is reflected in data analysis. So, we consider the Bayesian approach for patent data analysis. Basically, the Bayesian approach is based on Bayesian rules as follows [18].
P ( θ | x ) = P ( θ ) P ( x | θ ) P ( x )
where x and y are variables representing the patent keywords. That is, Bayes’ rule is dependent on the conditional probability. In addition, P ( x ) = θ P ( θ ) P ( x | θ ) . In Bayesian inference, P ( θ | x ) and P ( θ ) are posterior and prior respectively. Also, P ( x | θ ) is likelihood function based on observed data. The P ( x ) does not depend on θ so we can omit it as follow [10].
P ( θ | x )   P ( θ ) P ( x | θ )
This is core formula of Bayesian inference. In this paper, the θ represents the keyword extracted from patent documents, and we are finally interested in P ( θ | x ) called posterior density. So, Bayesian inference reflects not only the collected data but also the domain knowledge to which the data belongs, through posterior distribution as follow.
P o s t e r i o r P r i o r × L i k e l i h o o d
This feature of Bayesian statistics is useful for analyzing patent data. This is because patent data represents technology and technology requires the experience of domain experts. We apply Bayesian inference to patent data analysis for technology analysis as shown in Table 1.
In this paper, we combine the domain expert knowledge with collected patent data for technology patent analysis. Using this expression based on prior, likelihood and posterior, we compute Bayesian confidence interval for each keyword ( θ ). The 100 ( 1 α ) % Bayesian confidence interval C α of keyword θ is defined as follows [10,17].
P ( θ C α | x ) = ( 1 α )
The θ represents the frequency of the keyword having posterior distribution and is defined as follows [10,17].
θ C α P ( θ | x ) d θ = 1 α
In this case, we find the largest C α that satisfies
P ( θ C α | x ) 1 α
In general, Bayesian confidence intervals are not unique, because we can take many intervals satisfied with the proper probability coverage for a given posterior density. In this paper, x i represents ith keyword in the patent-keyword matrix. Let n keywords data follows a normal distribution as follows.
x 1 ,   x 2 ,   ,   x n   ~   N ( θ , σ 2 )
where σ 2 is known, and the parameter θ is distributed as N ( μ , v 2 ) . Then the posterior distribution of θ is shown as follow [10,17].
N ( ( 1 σ 2 σ 2 + n v 2 ) x ¯ + ( σ 2 σ 2 + n v 2 ) μ , σ 2 n ( 1 σ 2 σ 2 + n v 2 ) )
So, we can take 100 ( 1 α ) % Bayesian credible interval as follow [10,13].
( ( 1 σ 2 σ 2 + n v 2 ) x ¯ + ( σ 2 σ 2 + n v 2 ) μ ) ± z α / 2 σ n 1 σ 2 σ 2 + n v 2
Among the various credible intervals, we select the interval with the largest posterior density function of θ . If the confidence interval contains many θ s with large value of posterior probability density function, the confidence interval is shortened. Using this Bayesian interval based on posterior density, we carry out patent data analysis as follow.
In Figure 1, we first select the target technology to be analyzed. In this study, we choose AI technology as the target technology. Since the Fourth Industrial Revolution, AI has dominated most technological fields. We search the patent documents related to AI from the patent databases in the world. Using the collected AI patent documents, we build patent-keyword matrix as a structured data for patent data analysis. We apply various text mining techniques to this preprocessing. We use R data language and its packages for the preprocessing based on text mining [19,20]. The row and column of the structured data (patent-keyword matrix) are patent and keyword respectively. Each value of the matrix represents the occurred frequency of each keyword in a patent document. Next, we estimate the Bayesian credible intervals of AI keywords for AI technology analysis. So, we forecast the future trends of AI technology using the results of Bayesian interval estimation.
In this paper, we consider diverse prior distributions for Bayesian inference. The occurred frequency value of patent-keyword matrix can be distributed as Poisson probability density as follow [10].
P ( X = x ) = e λ λ x x ! ,   x = 0 , 1 , 2 ,
where, λ is the parameter of Poisson random variable representing average number of occurred frequency values. In this case, we can choose the following priors; Uniform, Jeffrey, Exponential, Gamma, and Chi-square. For example, we consider Chi-square prior distribution as follow [10],
P ( λ ) = λ d f / 2 1 × e λ / 2 Γ ( d f / 2 ) × 2 d f / 2 ,     λ > 0 ,       d f = 1 ,   2 ,  
By combining the prior of Chi-square with likelihood of Poisson, we can get the posterior distribution as follow [10].
( λ | x )     P ( λ ) × P ( x | λ )
  λ d f / 2 1 × e λ / 2 Γ ( d f / 2 ) × 2 d f / 2 × i = 1 n f ( x i | λ )
λ d f / 2 1 e λ / 2 ×   i = 1 n e λ λ x i x i !
  e ( n + 1 / 2 ) λ λ x i + d f / 2 1
So, the posterior probability density function is defined as follow [10].
G a m m a ( α = i = 1 n x i + d f 2 , β = n + 1 2 )
Table 2 represents the prior and posterior distributions used in our paper for Bayesian interval estimation [10].
In this paper, we use these five prior and posterior distributions to estimate Bayesian prediction intervals of patent keywords. Also, we estimate the Bayesian intervals of patent keywords over time. Because technology evolves over time [21]. Next, we make an experiment using AI patent keywords to illustrate the validity of our study.

4. Case Study Using AI Patent Data

Using the patent documents related to AI technology, we carried out a case study to show the validity of our research. This illustrates how our study could be applied to practical problems. We searched the AI patent documents from the patent databases [22,23]. The searching keywords equation is defined as follow;
Searching keywords equation = (big OR data) OR (neural OR network) OR ((Linear OR Logistic) AND Regression) OR (Support AND Vector) OR (Naive AND Bayesian) OR (Hidden AND Markov) OR (Conditional AND Random*) OR (Decision AND Tree) OR Cluster* OR ((Dimension OR feature) AND Reduction) OR (nearest AND neighbor) or ((((big OR data) OR mining) OR ((DB OR big) AND data)) OR knowledge OR (find OR detect OR discovery)))) or ((((ontology OR OWL OR (DAML AND OIL) OR SWRL OR (Semantic AND Web AND Rule AND Language) OR SEMANTIC) OR (represent* OR expres* OR design OR induct* OR deduct* OR reason* OR inferenc*)))) or ((((Boltzma* A/1 Machin*) OR ((deep* OR Convolution* OR Recurre* OR unsupervised* OR supervised* OR Reinforcement*) OR learning)) AND (KNN* OR k-nearest* OR (data OR mining) OR (big AND data*) OR neural OR network OR ((Linear OR Logistic) AND Regression) OR (Support AND Vector*) OR (Naive AND Bayesian*) OR (Hidden AND Markov*) OR (Conditional AND Random*) OR (Decision AND Tree) OR Cluster* OR ((Dimension OR feature) AND Reduction) OR (nearest AND neighbor*)))) or pattern* OR (aware* OR realiz* OR cognit* OR recogn* OR percept* OR understand* OR comprehens* OR estimat* OR assumpt* OR presump* OR anal*) OR (cognit* OR percept*) AND (computing OR process* OR application OR program) OR (humanlife OR “human life” OR living* OR livelihood OR lifelog OR “life log”) OR (aware* OR realiz* OR cognit* OR recogn* OR percept* OR understand* OR comprehens* OR estimat* OR assumpt* OR presump* OR anal*) or (emotion* OR sentiment* OR feel*) OR (aware* OR realiz* OR cognit* OR recogn* OR percept* OR understand* OR comprehens* or estimat* or assumpt* or presump* or anal*) OR space* OR (aware* OR realiz* OR cognit* OR recogn* or percept* or understand* or comprehens* or estimat* or assumpt* or presump* or anal*) or (collabor* or collect*) OR intel* OR ((((image* OR video* OR movie OR picture) OR (object* OR target* OR non-rigid* OR nonrigid*) OR (extract* OR awareness* OR realizat* OR cognit*OR capture*))))
From the collected patent documents, we removed the patent documents not related to AI or duplicated, and selected 13,858 valid patents from 1995 to 2016. In this case study, we use the R computing language and its text mining package to preprocess and analyze the collected patent data [19,20]. We also extracted 36 keywords representing AI technology from the valid patent documents by text mining technique. The keywords are as follow; analysis, awareness, behavior, cognitive, collaborative, computing, conversation, corpus, data, dialogue, feedback, figure, image, inference, interface, language, learning, mind. Morphological, natural, network, neuro, object, ontology, pattern, recognition, representation, sentence, sentiment, situation, spatial, speech, understanding, video, vision, voice. In this paper, we selected AI keywords in sufficient consultation with AI technology experts [24]. Using these keywords, we made the patent-keyword matrix as a structured data for Bayesian data analysis. The rows and columns of this matrix are 13,858 patents and 36 keywords respectively. Each element of this matrix is frequency value of occurred keyword in each patent. First, among the 36 AI technology keywords, we selected 8 keywords representing AI technology highly. Table 3 shows Bayesian intervals with 95% confidence level for the mean frequency of eight representative keywords.
In this case, we used five kinds of prior probability distributions, uniform, Jeffrey, exponential, and chi-square suitable for patent frequency data. We found that the differences in the widths of intervals were not large according to the prior probability distributions. From these results, we can see that the confidence interval for the keyword of data is the largest. Next was the speech and video keywords. So, we knew the technology based on data is most important to AI system. For more detailed AI technology forecasting, we estimated Bayesian prediction intervals for all 36 keywords. In addition, we show the intervals separated by time (1990s, 2000s, 2010s). Because technology is developed and evolves over time. Table 4 shows the Bayesian prediction interval with 95% confidence level for the 1990s.
From the intervals of AI keywords from 1990 to 1999, we can see the characteristics of AI technology during this period. We estimated Bayesian confidence intervals for all AI keywords using five representative prior distributions, uniform, Jeffrey, exponential, gamma, and chi-square, which are used for count data. We can see that the AI keywords with the highest average of their occurred frequencies in the estimated Bayesian intervals are object and data. In other words, we can confirm that the largest weight in the AI object is data. In addition, the keywords of speech, video, network, image, pattern, spatial, language, and analysis are also important in the AI technology field. During this period, AI technology was studied with a focus on data analysis, pattern recognition, speech and image processing, and understanding of natural language. Next, Table 5 illustrates the Bayesian prediction intervals with 95% confidence level for all AI keywords from 2000 to 2009.
As shown in the results of Table 4, the AI technology from 2000 to 2009 is affected by technological keywords such as data, object, speech, video, and image. However, this time period shows that the relative weight of feedback and learning keywords is increasing. In other words, we found that active researches related to the learning and behavior technologies for AI are being conducted at this time. Lastly, Table 6 represents the most recent results of Bayesian interval estimation for all AI keywords.
Similar to the results in Table 5 and Table 6, we can find that there is still a lot of research going on in data technology, imaging, and natural language processing technology at this time. Especially since 2010, the research demands for analysis, recognition, and interface for AI have increased. We also confirmed the emergence of technology related to AI situation. Combining the results in Table 4, Table 5 and Table 6, we show the technological change of AI over time in Figure 2.
This figure shows the top 20 AI technology keywords for each period. The impact of the keywords of object and data on AI technology has not changed over the years. However, the other technological keywords except object and data show a change in relative influence on AI technology over time. For example, the keyword of voice increased in the 2000s and then declined. In contrast, the keyword of interface tends to be the opposite of the keyword of voice. From the result of Figure 2, we conclude the technology related to data is very important to AI technology. In addition, the technology of video and image influences on the development of AI technology. Pattern recognition and natural language understanding are meaningful technologies for AI. Recently, the technology related to situation is needed to improve the performance of AI technology. This figure also provides the various results for researchers and developers in AI technology areas to build their R&D strategies for AI technology. To identify the trend of keywords over time, we show the ranking of all AI keywords by the width of Bayesian prediction intervals with 95% confidence level over time in Table 7.
The interval widths of most AI keywords have decreased, and this result is similar to the result of Figure 2. From the results, we confirmed that the detailed technologies for AI are expanding to various technological keywords without focusing on specific keywords. So, in the future, the AI technology is expected to have a large impact on most other technology fields as well as the society.

5. Conclusions

In this paper, we proposed a method for patent data analysis using Bayesian prediction interval estimation. Bayesian inference combines prior distribution with likelihood function to build posterior distribution for parameters. In general, the prior distribution represents the previous knowledge of parameter domain. In our technology analysis by patent data analysis, the prior distribution contains the experience and knowledge of AI experts. The likelihood function represents the likelihood for the parameters described by the observed data. In our method, we use the collected AI patent documents for the data in the likelihood function. We estimate the Bayesian prediction intervals of the parameters representing AI keywords using the posterior distribution combining the prior and the likelihood. Therefore, the proposed model is a statistical patent analysis model that can reflect the opinions of experts. This is an important issue in technology forecasting and management. This is because technology analysis should consider not only quantitative patent analysis but also expert knowledge in the relevant technological field. We tried to solve this problem in traditional patent data analysis using the learning approach of Bayesian statistics.
In addition, we applied the proposed method to AI technology analysis. This is because AI is a very interesting area in most technological fields. We collected the patent documents related to AI from the patent databases in the world. Using the preprocessing by text mining techniques, we made a structured data which is a matrix consists of patents (rows) and keywords (columns). Each element of this matrix represents the occurred frequency of a keyword in each patent. We estimated the Bayesian prediction intervals of the frequency mean for all AI keywords. From the results of Bayesian interval estimation, we extracted technological trends of AI in the future. So, we conclude that the data technology is core technology for AI system. In addition, the AI developers have studied on the video and image technology as well as the natural language processing. Recently, AI researchers are also interested in developing technology for situational awareness.
Our research is worthy of two perspectives in academic and practical points. On the academic aspects, the proposed method presented Bayesian inference method for patent keyword analysis considering the experience and knowledge of AI domain experts. Next, our research contributes to the technology management such as technology forecasting, research and development planning, technological innovation, etc., on the practical aspects. In our future study, we will study on more advanced Bayesian models for patent data analysis using the Markov Chain Monte Carlo (MCMC) computation, finite mixture or hierarchical models, Bayesian graph model, etc. In addition, we will consider the possibility of geographical patent distribution according to countries. So, we will compare the AI technology portfolios between the countries around the world. Lastly, we are to study on the way how this methodology could be applied to other practical scenarios with experts’ opinion.

Author Contributions

J.-B.R. and S.J. designed this research and collected the data set for the experiment. D.U. analyzed the data to show the validity of this paper and wrote the paper and performed all the research steps. In addition, all authors have cooperated with each other for revising the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2017R1D1A3B03031152).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Choi, J.; Jang, D.; Jun, S.; Park, S. A Predictive Model of Technology Transfer using Patent Analysis. Sustainability 2015, 7, 16175–16195. [Google Scholar] [CrossRef] [Green Version]
  2. Jun, S. A New Patent Analysis Using Association Rule Mining and Box-Jenkins Modeling for Technology Forecasting. Inf. Int. Interdiscip. J. 2013, 16, 555–562. [Google Scholar]
  3. Jun, S.; Park, S. Examining Technological Innovation of Apple Using Patent Analysis. Ind. Manag. Data Syst. 2013, 113, 890–907. [Google Scholar] [CrossRef]
  4. Kim, J.; Jun, S. Graphical causal inference and copula regression model for apple keywords by text mining. Adv. Eng. Inform. 2015, 29, 918–929. [Google Scholar] [CrossRef]
  5. Kim, J.; Jun, S.; Jang, D.; Park, S. An Integrated Social Network Mining for Product-based Technology Analysis of Apple. Ind. Manag. Data Syst. 2017, 117, 2417–2430. [Google Scholar] [CrossRef]
  6. Hunt, D.; Nguyen, L.; Rodgers, M. Patent Searching Tools & Techniques; Wiley: Hoboken, NJ, USA, 2007. [Google Scholar]
  7. Roper, A.T.; Cunningham, S.W.; Porter, A.L.; Mason, T.W.; Rossini, F.A.; Banks, J. Forecasting and Management of Technology; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  8. Kim, J.-M.; Kim, N.-K.; Jung, Y.; Jun, S. Patent data analysis using functional count data model. Soft Comput. 2019, 23, 8815–8826. [Google Scholar] [CrossRef]
  9. Uhm, D.; Ryu, J.; Jun, S. An Interval Estimation Method of Patent Keyword Data for Sustainable Technology Forecasting. Sustainability 2017, 9, 2025. [Google Scholar] [CrossRef] [Green Version]
  10. Gelman, A.; Carlin, J.B.; Stern, H.S.; Dunson, D.B.; Vehtari, A.; Rubin, D.B. Bayesian Data Analysis, 3rd ed.; Chapman & Hall/CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
  11. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  12. PwC. Artificial Intelligence Everywhere. Available online: http://www.pwc.com/ai (accessed on 2 January 2020).
  13. Deloitte. AI & Cognitive Technologies. Available online: https://www2.deloitte.com (accessed on 2 January 2020).
  14. Deloitte Insights. Future in the Valance? How Countries Are Pursuing an AI Advantage. Available online: https://www.deloitte.com/insights (accessed on 2 January 2020).
  15. Kim, J.-M.; Yoon, J.; Hwang, S.Y.; Jun, S. Patent Keyword Analysis Using Time Series and Copula Models. Appl. Sci. 2019, 9, 4071. [Google Scholar] [CrossRef] [Green Version]
  16. Kim, J.; Jun, S.; Jang, D.; Park, S. Sustainable Technology Analysis of Artificial Intelligence Using Bayesian and Social Network Models. Sustainability 2018, 10, 155. [Google Scholar] [CrossRef] [Green Version]
  17. Kruschke, J.K. Doing Bayesian Data Analysis, 2nd ed.; AP Elsevier: Waltham, MA, USA, 2015. [Google Scholar]
  18. Theodoridis, S. Machine Learning, a Bayesian and Optimization Perspective; Elsevier: London, UK, 2015. [Google Scholar]
  19. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2018. [Google Scholar]
  20. Feinerer, I.; Hornik, K. Package ‘tm’ Ver. 0.7-5, Text Mining Package, CRAN of R Project; R Foundation for Statistical Computing: Vienna, Austria, 2018. [Google Scholar]
  21. Dehghani, M.; Dangelico, R.M. Smart wearable technologies: State of the art and evolution over time through patent analysis and clustering. Int. J. Prod. Dev. 2018, 22, 293–313. [Google Scholar] [CrossRef]
  22. USPTO. The United States Patent and Trademark Office. Available online: http://www.uspto.gov (accessed on 10 July 2018).
  23. WIPSON. WIPS Corporation. Available online: http://www.wipson.com (accessed on 30 April 2018).
  24. KISTA. Korea Intellectual Property Strategy Agency. Available online: https://www.kista.re.kr/ (accessed on 1 January 2019).
Figure 1. Proposed patent data analysis.
Figure 1. Proposed patent data analysis.
Applsci 10 00570 g001
Figure 2. Technological changes of top 20 AI keywords over time.
Figure 2. Technological changes of top 20 AI keywords over time.
Applsci 10 00570 g002
Table 1. Bayesian inference and patent technology analysis.
Table 1. Bayesian inference and patent technology analysis.
Bayesian ComponentDensityPatent Technology Data
Prior P ( θ ) Domain knowledge of target technology
Likelihood P ( x | θ ) Collected patent documents
Posterior P ( θ | x ) Combine domain knowledge with patent data
Table 2. Prior and posterior distributions for Bayesian interval estimation.
Table 2. Prior and posterior distributions for Bayesian interval estimation.
Prior   Distribution ,   P ( λ ) Posterior   Distribution ,   P ( λ | x )
Uniform 1 G a m m a ( α = i = 1 n x i + 1 ,   β = n )
Jeffrey’s λ 1 / 2 G a m m a ( α = i = 1 n x i + 1 2 ,   β = n )
Exponential (a) a e a λ G a m m a ( α = i = 1 n x i + 1 ,   β = n + a )
Gamma (a, b) λ a 1 × e b λ Γ ( a ) × b a G a m m a ( α = i = 1 n x i + a ,   β = n + b )
Chi-square (df) λ d f / 2 1 × e λ / 2 Γ ( d f / 2 ) × 2 d f / 2 G a m m a ( α = i = 1 n x i + d f 2 ,   β = n + 1 2 )
Table 3. Bayesian intervals for eight representative keywords.
Table 3. Bayesian intervals for eight representative keywords.
KeywordUniformJeffreyExponential
(a = 1)
Gamma
(a = 3, b = 1)
Chi-Square
(DF = 5)
learning(0.0047, 0.0073)(0.0047, 0.0072)(0.0048, 0.0074)(0.0047, 0.0073)(0.0048, 0.0074)
analysis(0.0260, 0.0316)(0.0259, 0.0316)(0.0261, 0.0318)(0.0260, 0.0316)(0.0261, 0.0317)
data(0.8165, 0.8469)(0.8165, 0.8469)(0.8166, 0.8470)(0.8165, 0.8468)(0.8166, 0.8470)
image(0.2659, 0.2834)(0.2659, 0.2833)(0.2660, 0.2835)(0.2659, 0.2833)(0.2660, 0.2835)
network(0.2583, 0.2755)(0.2583, 0.2755)(0.2584, 0.2756)(0.2583, 0.2755)(0.2584, 0.2756)
pattern(0.1443, 0.1572)(0.1442, 0.1572)(0.1444, 0.1573)(0.1443, 0.1572)(0.1444, 0.1573)
speech(0.5405, 0.5653)(0.5405, 0.5652)(0.5406, 0.5654)(0.5405, 0.5652)(0.5406, 0.5654)
Table 4. Bayesian prediction intervals: 1990s.
Table 4. Bayesian prediction intervals: 1990s.
KeywordUniformJeffreyExponential
(a = 1)
Gamma
(a = 3, b = 1)
Chi-Square
(DF = 5)
analysis(0.0252, 0.0404)(0.0250, 0.0401)(0.0252, 0.0403)(0.0260, 0.0414)(0.0258, 0.0411)
awareness(0.0000, 0.0017)(0.0000, 0.0012)(0.0000, 0.0017)(0.0003, 0.0033)(0.0002, 0.0030)
behavior(0.0144, 0.0262)(0.0142, 0.0260)(0.0144, 0.0262)(0.0152, 0.0273)(0.0150, 0.0270)
cognitive(0.0000, 0.0017)(0.0000, 0.0012)(0.0000, 0.0017)(0.0003, 0.0033)(0.0002, 0.0030)
collaborative(0.0000, 0.0017)(0.0000, 0.0012)(0.0000, 0.0017)(0.0003, 0.0033)(0.0002, 0.0030)
computing(0.0001, 0.0026)(0.0000, 0.0022)(0.0001, 0.0026)(0.0005, 0.0040)(0.0004, 0.0037)
conversation(0.0022, 0.0079)(0.0021, 0.0076)(0.0022, 0.0079)(0.0029, 0.0091)(0.0027, 0.0088)
corpus(0.0094, 0.0192)(0.0092, 0.0190)(0.0093, 0.0192)(0.0101, 0.0203)(0.0099, 0.0201)
data(0.9117, 0.9940)(0.9115, 0.9937)(0.9113, 0.9935)(0.9122, 0.9944)(0.9122, 0.9944)
dialogue(0.0010, 0.0054)(0.0009, 0.0051)(0.0010, 0.0054)(0.0016, 0.0067)(0.0014, 0.0063)
feedback(0.0236, 0.0383)(0.0234, 0.0380)(0.0236, 0.0383)(0.0244, 0.0393)(0.0242, 0.0391)
figure(0.0013, 0.0060)(0.0012, 0.0057)(0.0013, 0.0060)(0.0019, 0.0073)(0.0017, 0.0070)
image(0.2179, 0.2590)(0.2177, 0.2587)(0.2178, 0.2589)(0.2187, 0.2598)(0.2185, 0.2596)
inference(0.0008, 0.0047)(0.0006, 0.0044)(0.0007, 0.0047)(0.0013, 0.0060)(0.0012, 0.0057)
interface(0.0136, 0.0252)(0.0134, 0.0249)(0.0136, 0.0252)(0.0144, 0.0262)(0.0142, 0.0260)
language(0.0389, 0.0572)(0.0386, 0.0570)(0.0388, 0.0572)(0.0397, 0.0582)(0.0395, 0.0580)
learning(0.0029, 0.0091)(0.0027, 0.0088)(0.0029, 0.0091)(0.0035, 0.0103)(0.0034, 0.0100)
mind(0.0001, 0.0026)(0.0000, 0.0022)(0.0001, 0.0026)(0.0005, 0.0040)(0.0004, 0.0037)
morphological(0.0000, 0.0017)(0.0000, 0.0012)(0.0000, 0.0017)(0.0003, 0.0033)(0.0002, 0.0030)
natural(0.0016, 0.0067)(0.0014, 0.0064)(0.0016, 0.0067)(0.0022, 0.0079)(0.0021, 0.0076)
network(0.3008, 0.3488)(0.3006, 0.3486)(0.3007, 0.3487)(0.3016, 0.3496)(0.3014, 0.3495)
neuro(0.0001, 0.0026)(0.0000, 0.0022)(0.0001, 0.0026)(0.0005, 0.0040)(0.0004, 0.0037)
object(1.1159, 1.2066)(1.1156, 1.2064)(1.1153, 1.2061)(1.1162, 1.2070)(1.1163, 1.2071)
ontology(0.0001, 0.0026)(0.0000, 0.0022)(0.0001, 0.0026)(0.0005, 0.0040)(0.0004, 0.0037)
pattern(0.2020, 0.2416)(0.2017, 0.2414)(0.2019, 0.2415)(0.2028, 0.2425)(0.2026, 0.2423)
recognition(0.0164, 0.0289)(0.0162, 0.0286)(0.0163, 0.0289)(0.0171, 0.0299)(0.0169, 0.0297)
representation(0.0035, 0.0103)(0.0034, 0.0100)(0.0035, 0.0103)(0.0042, 0.0114)(0.0041, 0.0111)
sentence(0.0128, 0.0241)(0.0126, 0.0238)(0.0128, 0.0241)(0.0136, 0.0252)(0.0134, 0.0249)
sentiment(0.0000, 0.0017)(0.0000, 0.0012)(0.0000, 0.0017)(0.0003, 0.0033)(0.0002, 0.0030)
situation(0.0029, 0.0091)(0.0027, 0.0088)(0.0029, 0.0091)(0.0035, 0.0103)(0.0034, 0.0100)
spatial(0.0676, 0.0913)(0.0674, 0.0910)(0.0676, 0.0913)(0.0684, 0.0922)(0.0682, 0.0920)
speech(0.7130, 0.7860)(0.7128, 0.7857)(0.7127, 0.7856)(0.7136, 0.7866)(0.7136, 0.7865)
understanding(0.0000, 0.0017)(0.0000, 0.0012)(0.0000, 0.0017)(0.0003, 0.0033)(0.0002, 0.0030)
video(0.5155, 0.5778)(0.5153, 0.5775)(0.5153, 0.5775)(0.5162, 0.5785)(0.5161, 0.5784)
vision(0.0124, 0.0236)(0.0122, 0.0233)(0.0124, 0.0236)(0.0132, 0.0246)(0.0130, 0.0244)
voice(0.0082, 0.0176)(0.0080, 0.0173)(0.0082, 0.0176)(0.0090, 0.0187)(0.0088, 0.0184)
Table 5. Bayesian prediction intervals: 2000s.
Table 5. Bayesian prediction intervals: 2000s.
KeywordUniformJeffreyExponential
(a = 1)
Gamma
(a = 3, b = 1)
Chi-Square
(DF = 5)
analysis(0.0238, 0.0316)(0.0237, 0.0315)(0.0238, 0.0316)(0.0241, 0.0319)(0.0240, 0.0318)
awareness(0.0002, 0.0015)(0.0002, 0.0014)(0.0002, 0.0015)(0.0004, 0.0019)(0.0004, 0.0018)
behavior(0.0239, 0.0318)(0.0239, 0.0317)(0.0239, 0.0318)(0.0242, 0.0321)(0.0241, 0.0320)
cognitive(0.0001, 0.0010)(0.0001, 0.0009)(0.0001, 0.0010)(0.0002, 0.0015)(0.0002, 0.0014)
collaborative(0.0000, 0.0008)(0.0000, 0.0007)(0.0000, 0.0008)(0.0002, 0.0013)(0.0001, 0.0012)
computing(0.0002, 0.0013)(0.0001, 0.0012)(0.0002, 0.0013)(0.0003, 0.0017)(0.0003, 0.0016)
conversation(0.0017, 0.0041)(0.0016, 0.0040)(0.0017, 0.0041)(0.0019, 0.0045)(0.0018, 0.0044)
corpus(0.0083, 0.0131)(0.0082, 0.0130)(0.0083, 0.0131)(0.0085, 0.0134)(0.0084, 0.0133)
data(0.8192, 0.8624)(0.8192, 0.8623)(0.8191, 0.8623)(0.8194, 0.8626)(0.8194, 0.8626)
dialogue(0.0004, 0.0019)(0.0004, 0.0018)(0.0004, 0.0019)(0.0006, 0.0023)(0.0005, 0.0022)
feedback(0.0250, 0.0330)(0.0249, 0.0329)(0.0250, 0.0330)(0.0253, 0.0333)(0.0252, 0.0332)
figure(0.0002, 0.0013)(0.0001, 0.0012)(0.0002, 0.0013)(0.0003, 0.0017)(0.0003, 0.0016)
image(0.2345, 0.2578)(0.2344, 0.2578)(0.2345, 0.2578)(0.2347, 0.2581)(0.2347, 0.2580)
inference(0.0008, 0.0027)(0.0007, 0.0026)(0.0008, 0.0027)(0.0010, 0.0030)(0.0009, 0.0029)
interface(0.0150, 0.0213)(0.0149, 0.0213)(0.0150, 0.0213)(0.0153, 0.0216)(0.0152, 0.0216)
language(0.0349, 0.0442)(0.0348, 0.0441)(0.0349, 0.0442)(0.0351, 0.0445)(0.0351, 0.0444)
learning(0.0054, 0.0093)(0.0053, 0.0093)(0.0054, 0.0093)(0.0056, 0.0097)(0.0055, 0.0096)
mind(0.0002, 0.0013)(0.0001, 0.0012)(0.0002, 0.0013)(0.0003, 0.0017)(0.0003, 0.0016)
morphological(0.0002, 0.0013)(0.0001, 0.0012)(0.0002, 0.0013)(0.0003, 0.0017)(0.0003, 0.0016)
natural(0.0003, 0.0017)(0.0003, 0.0016)(0.0003, 0.0017)(0.0005, 0.0021)(0.0005, 0.0020)
network(0.2187, 0.2413)(0.2186, 0.2412)(0.2187, 0.2413)(0.2190, 0.2416)(0.2189, 0.2415)
neuro(0.0017, 0.0041)(0.0016, 0.0040)(0.0017, 0.0041)(0.0019, 0.0045)(0.0018, 0.0044)
object(1.2712, 1.3248)(1.2711, 1.3247)(1.2710, 1.3246)(1.2713, 1.3249)(1.2713, 1.3249)
ontology(0.0041, 0.0077)(0.0041, 0.0076)(0.0041, 0.0077)(0.0044, 0.0080)(0.0043, 0.0079)
pattern(0.1514, 0.1703)(0.1514, 0.1703)(0.1514, 0.1703)(0.1517, 0.1706)(0.1516, 0.1705)
recognition(0.0162, 0.0227)(0.0161, 0.0227)(0.0162, 0.0227)(0.0165, 0.0231)(0.0164, 0.0230)
representation(0.0054, 0.0093)(0.0053, 0.0093)(0.0054, 0.0093)(0.0056, 0.0097)(0.0055, 0.0096)
sentence(0.0088, 0.0137)(0.0087, 0.0136)(0.0088, 0.0137)(0.0090, 0.0140)(0.0090, 0.0140)
sentiment(0.0000, 0.0008)(0.0000, 0.0007)(0.0000, 0.0008)(0.0002, 0.0013)(0.0001, 0.0012)
situation(0.0046, 0.0084)(0.0046, 0.0083)(0.0046, 0.0084)(0.0049, 0.0087)(0.0048, 0.0086)
spatial(0.0956, 0.1107)(0.0955, 0.1106)(0.0956, 0.1107)(0.0959, 0.1110)(0.0958, 0.1109)
speech(0.5729, 0.6091)(0.5729, 0.6091)(0.5729, 0.6091)(0.5731, 0.6093)(0.5731, 0.6093)
understanding(0.0004, 0.0019)(0.0004, 0.0018)(0.0004, 0.0019)(0.0006, 0.0023)(0.0005, 0.0022)
video(0.5034, 0.5373)(0.5033, 0.5372)(0.5033, 0.5372)(0.5036, 0.5375)(0.5035, 0.5375)
vision(0.0095, 0.0147)(0.0095, 0.0146)(0.0095, 0.0147)(0.0098, 0.0150)(0.0097, 0.0149)
voice(0.0171, 0.0238)(0.0171, 0.0238)(0.0171, 0.0238)(0.0174, 0.0241)(0.0173, 0.0241)
Table 6. Bayesian prediction intervals: 2010s.
Table 6. Bayesian prediction intervals: 2010s.
KeywordUniformJeffreyExponential
(a = 1)
Gamma
(a = 3, b = 1)
Chi-Square
(DF = 5)
analysis(0.0245, 0.0342)(0.0244, 0.0341)(0.0245, 0.0342)(0.0249, 0.0347)(0.0248, 0.0346)
awareness(0.0007, 0.0030)(0.0007, 0.0029)(0.0007, 0.0030)(0.0010, 0.0036)(0.0009, 0.0034)
behavior(0.0220, 0.0313)(0.0219, 0.0311)(0.0220, 0.0313)(0.0224, 0.0317)(0.0223, 0.0316)
cognitive(0.0000, 0.0008)(0.0000, 0.0005)(0.0000, 0.0008)(0.0001, 0.0015)(0.0001, 0.0013)
collaborative(0.0000, 0.0008)(0.0000, 0.0005)(0.0000, 0.0008)(0.0001, 0.0015)(0.0001, 0.0013)
computing(0.0012, 0.0039)(0.0011, 0.0037)(0.0012, 0.0039)(0.0015, 0.0044)(0.0014, 0.0043)
conversation(0.0029, 0.0067)(0.0028, 0.0066)(0.0029, 0.0067)(0.0032, 0.0072)(0.0031, 0.0071)
corpus(0.0116, 0.0186)(0.0115, 0.0184)(0.0116, 0.0186)(0.0120, 0.0190)(0.0119, 0.0189)
data(0.7394, 0.7891)(0.7393, 0.7890)(0.7393, 0.7889)(0.7397, 0.7893)(0.7397, 0.7893)
dialogue(0.0001, 0.0015)(0.0001, 0.0013)(0.0001, 0.0015)(0.0003, 0.0021)(0.0003, 0.0020)
feedback(0.0238, 0.0333)(0.0237, 0.0332)(0.0238, 0.0333)(0.0241, 0.0338)(0.0240, 0.0337)
figure(0.0001, 0.0012)(0.0000, 0.0010)(0.0001, 0.0012)(0.0002, 0.0018)(0.0002, 0.0017)
image(0.3169, 0.3497)(0.3168, 0.3496)(0.3169, 0.3496)(0.3173, 0.3501)(0.3172, 0.3500)
inference(0.0016, 0.0047)(0.0015, 0.0045)(0.0016, 0.0047)(0.0019, 0.0052)(0.0018, 0.0051)
interface(0.0118, 0.0188)(0.0117, 0.0187)(0.0118, 0.0188)(0.0122, 0.0193)(0.0121, 0.0191)
language(0.0397, 0.0518)(0.0396, 0.0517)(0.0397, 0.0518)(0.0401, 0.0522)(0.0400, 0.0521)
learning(0.0029, 0.0067)(0.0028, 0.0066)(0.0029, 0.0067)(0.0032, 0.0072)(0.0031, 0.0071)
mind(0.0001, 0.0015)(0.0001, 0.0013)(0.0001, 0.0015)(0.0003, 0.0021)(0.0003, 0.0020)
morphological(0.0002, 0.0018)(0.0002, 0.0017)(0.0002, 0.0018)(0.0005, 0.0024)(0.0004, 0.0023)
natural(0.0002, 0.0018)(0.0002, 0.0017)(0.0002, 0.0018)(0.0005, 0.0024)(0.0004, 0.0023)
network(0.2797, 0.3105)(0.2796, 0.3104)(0.2796, 0.3105)(0.2800, 0.3109)(0.2800, 0.3108)
neuro(0.0001, 0.0012)(0.0000, 0.0010)(0.0001, 0.0012)(0.0002, 0.0018)(0.0002, 0.0017)
object(1.4808, 1.5507)(1.4807, 1.5506)(1.4805, 1.5504)(1.4809, 1.5508)(1.4809, 1.5508)
ontology(0.0029, 0.0067)(0.0028, 0.0066)(0.0029, 0.0067)(0.0032, 0.0072)(0.0031, 0.0071)
pattern(0.0954, 0.1137)(0.0953, 0.1136)(0.0953, 0.1137)(0.0957, 0.1141)(0.0956, 0.1140)
recognition(0.0197, 0.0285)(0.0196, 0.0284)(0.0197, 0.0285)(0.0201, 0.0290)(0.0200, 0.0289)
representation(0.0034, 0.0075)(0.0033, 0.0074)(0.0034, 0.0075)(0.0037, 0.0080)(0.0036, 0.0079)
sentence(0.0042, 0.0087)(0.0042, 0.0086)(0.0042, 0.0087)(0.0046, 0.0092)(0.0045, 0.0091)
sentiment(0.0029, 0.0067)(0.0028, 0.0066)(0.0029, 0.0067)(0.0032, 0.0072)(0.0031, 0.0071)
situation(0.0042, 0.0087)(0.0042, 0.0086)(0.0042, 0.0087)(0.0046, 0.0092)(0.0045, 0.0091)
spatial(0.0871, 0.1047)(0.0870, 0.1046)(0.0871, 0.1047)(0.0875, 0.1051)(0.0874, 0.1050)
speech(0.3907, 0.4270)(0.3906, 0.4269)(0.3907, 0.4270)(0.3911, 0.4274)(0.3910, 0.4273)
understanding(0.0001, 0.0012)(0.0000, 0.0010)(0.0001, 0.0012)(0.0002, 0.0018)(0.0002, 0.0017)
video(0.4639, 0.5034)(0.4638, 0.5033)(0.4638, 0.5033)(0.4642, 0.5037)(0.4641, 0.5036)
vision(0.0083, 0.0143)(0.0082, 0.0142)(0.0083, 0.0143)(0.0087, 0.0148)(0.0086, 0.0147)
voice(0.0072, 0.0129)(0.0072, 0.0128)(0.0072, 0.0129)(0.0076, 0.0134)(0.0075, 0.0132)
Table 7. Width ranking of Bayesian prediction intervals: all AI keywords.
Table 7. Width ranking of Bayesian prediction intervals: all AI keywords.
Ranking1990s2000s2010s
KeywordWidthKeywordWidthKeywordWidth
1object0.0907object0.0536object0.0699
2data0.0823data0.0432data0.0497
3speech0.073speech0.0362video0.0395
4video0.0623video0.0339speech0.0363
5network0.048image0.0233image0.0328
6image0.0411network0.0226network0.0308
7pattern0.0396pattern0.0189pattern0.0183
8spatial0.0237spatial0.0151spatial0.0176
9language0.0183language0.0093language0.0121
10analysis0.0152feedback0.008analysis0.0097
11feedback0.0147behavior0.0079feedback0.0095
12recognition0.0125analysis0.0078behavior0.0093
13behavior0.0118voice0.0067recognition0.0088
14interface0.0116recognition0.0065interface0.007
15sentence0.0113interface0.0063corpus0.007
16vision0.0112vision0.0052vision0.006
17corpus0.0098sentence0.0049voice0.0057
18voice0.0094corpus0.0048sentence0.0045
19representation0.0068learning0.0039situation0.0045
20learning0.0062representation0.0039representation0.0041
21situation0.0062situation0.0038conversation0.0038
22conversation0.0057ontology0.0036learning0.0038
23natural0.0051conversation0.0024ontology0.0038
24figure0.0047neuro0.0024sentiment0.0038
25dialogue0.0044inference0.0019inference0.0031
26inference0.0039dialogue0.0015computing0.0027
27computing0.0025understanding0.0015awareness0.0023
28mind0.0025natural0.0014morphological0.0016
29neuro0.0025awareness0.0013natural0.0016
30ontology0.0025computing0.0011dialogue0.0014
31awareness0.0017figure0.0011mind0.0014
32cognitive0.0017mind0.0011figure0.0011
33collaborative0.0017morphological0.0011neuro0.0011
34morphological0.0017cognitive0.0009understanding0.0011
35sentiment0.0017collaborative0.0008cognitive0.0008
36understanding0.0017sentiment0.0008collaborative0.0008

Share and Cite

MDPI and ACS Style

Uhm, D.; Ryu, J.-B.; Jun, S. Patent Data Analysis of Artificial Intelligence Using Bayesian Interval Estimation. Appl. Sci. 2020, 10, 570. https://doi.org/10.3390/app10020570

AMA Style

Uhm D, Ryu J-B, Jun S. Patent Data Analysis of Artificial Intelligence Using Bayesian Interval Estimation. Applied Sciences. 2020; 10(2):570. https://doi.org/10.3390/app10020570

Chicago/Turabian Style

Uhm, Daiho, Jea-Bok Ryu, and Sunghae Jun. 2020. "Patent Data Analysis of Artificial Intelligence Using Bayesian Interval Estimation" Applied Sciences 10, no. 2: 570. https://doi.org/10.3390/app10020570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop