Next Article in Journal
Statistical Generalized Derivative Applied to the Profile Likelihood Estimation in a Mixture of Semiparametric Models
Next Article in Special Issue
A Note on Complexities by Means of Quantum Compound Systems
Previous Article in Journal
A Novel Hybrid Secure Image Encryption Based on Julia Set of Fractals and 3D Lorenz Chaotic Map
Previous Article in Special Issue
Some Notes on Counterfactuals in Quantum Mechanics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Applying the Bell’s Test to Chinese Texts

1
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
2
Saint Petersburg National Research, University of Information Technology Mechanics and Optics, St. Petersburg 197101, Russia
3
Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(3), 275; https://doi.org/10.3390/e22030275
Submission received: 24 January 2020 / Revised: 21 February 2020 / Accepted: 25 February 2020 / Published: 28 February 2020
(This article belongs to the Special Issue Quantum Information Revolution: Impact to Foundations)

Abstract

:
Search engines are able to find documents containing patterns from a query. This approach can be used for alphabetic languages such as English. However, Chinese is highly dependent on context. The significant problem of Chinese text processing is the missing blanks between words, so it is necessary to segment the text to words before any other action. Algorithms for Chinese text segmentation should consider context; that is, the word segmentation process depends on other ideograms. As the existing segmentation algorithms are imperfect, we have considered an approach to build the context from all possible n-grams surrounding the query words. This paper proposes a quantum-inspired approach to rank Chinese text documents by their relevancy to the query. Particularly, this approach uses Bell’s test, which measures the quantum entanglement of two words within the context. The contexts of words are built using the hyperspace analogue to language (HAL) algorithm. Experiments fulfilled in three domains demonstrated that the proposed approach provides acceptable results.

1. Introduction

The exponential growth in the number of Internet resources has led to problems in information retrieval. A simple search query can generate thousands or even millions of links. Traditional search methods are based on a selection from an index that links to documents that contain words from the query line [1]. The quality of a naive search using the same words from the query is insufficient for many reasons, including the synonymy and polysemy problems. For European languages, there are some techniques that improve the quality of information retrieval based on the index search. Latent semantic analysis (LSA) [2] allows finding words that have close meanings by using singular value decomposition. Hyperspace analogue to language (HAL) [3] analyzes word pairs by taking into account the frequency of each pair and the distance between words in the text. These models allow words to be represented by vectors of a reduced dimension, where semantically close words have similar vectors. Each row of the HAL matrix represents the context of a word that could be meant as a word sense. Information retrieval by word vectors instead of word patterns could involve word meanings into the search process and increase the search quality.
Searching by word vectors requires the documents to be vectorized. In this case, the information retrieval means that the document vectors should match the query vector. This matching is usually based on linear algebra. Nowadays, some researchers apply quantum mathematics to biology [4], sociology [5,6], and this task [7,8]. Document relevance to the query could be measured as a document vector’s projection to the query plane built by two query words. In [9,10], Bell’s test was used to assess the entanglement of words from a query in English texts as a measure of document relevance.
Chinese differs from alphabetic languages due to its very high context dependency because of the wide polysemy of ideograms. The coexistence of words in a text is not enough to assess their relevance to the query. Another feature of Chinese is the absence of blanks between words, which yields the nontrivial problem of word segmentation. The goal of this study is to estimate the applicability of quantum-inspired models to information retrieval in Chinese texts. This paper proposes an approach to estimate word pair entanglement using Bell’s test in CHSH form. In this work, an entanglement of two words means a correlation of these words’ contexts in documents. This measure is used for ranking Chinese documents by a pair of words from a user’s query.

2. Materials and Methods

The wide polysemy of Chinese words makes searching documents by patterns inefficient. From this point of view, it is interesting to apply to Chinese texts an approach that was validated for English in [10]. This approach is based on vector word representation and it incorporates the context of words into the search process. The authors of [10] used quantum mathematical formalism and Bell’s test as a measure of documents’ relevance to the query. Unlike in [10], in this work, an asymmetric HAL built on the n-gram segmentation principle was used.
Vector document representation operates with words but not with symbols. Chinese words consist of one or more characters, but a text in Chinese is a continuous sequence of ideograms without spaces between words. The standard preprocessing stage of Chinese texts is word segmentation. There are several Chinese word segmenting programs. The simplest ones, such as JIEBA (https://pypi.org/project/jieba/), do not use a dictionary, but the segmentation quality is not high. More powerful Chinese text analyzers such as CRIE implement multiphase processing, including word segmentation based on a large dictionary [11]. The authors of [12] showed that Chinese word segmentation is able to improve the performance of information retrieval in Chinese texts, but the improvement is not significant if the search is based on words as patterns in the documents (i.e., substrings in the sequence of ideograms).
In [13], an approach to Chinese document retrieval based on the matching of text strings was considered. Relevance was determined by how terms appeared in the text collection rather than by statistics of appearance, and the order of words was taken into account.
Unlike alphabetic languages, a separate syllable written as an ideogram in Chinese has a meaning. This allows the use of syllable-based statistical information in information retrieval [14]. Researchers have also had a positive experience in domain term extraction from Chinese texts without word segmentation [15]. The key idea is to split the sequence of symbols such as “abcdefg” into overlapping n-grams of “abc”, “bcd,” “cde,” “def,” and “efg.” Of course, among real words, this set will contain some bonded fragments of adjacent words. On the one hand, such pseudowords overload the document’s glossary; on the other hand, they indicate the joint occurrence of the words.
The “bag-of-words” paradigm limits the ability of a computer to understand natural language because it ignores the meaning of words and sentences. The vector model is an attempt to describe the sense of a word by the context [16]. Vector word representation means transforming a word w into the vector Vw = [v1 v2 … vL]T, which has dimension L in the space of real numbers. Thus, we need to build a projection a(w): W → RL, where W is a set of words, and w ∈W. According to the distributive hypothesis of Harris [7], vi contains the number of occurrences of the word vi from the dictionary W in a window around the word w. So, for the sentence “Alice likes Bob” and the dictionary [Alice, like, Bob], the context of the word “like” is defined by the vector v = [1 0 1]T. Of course, the source text should be lemmatized before processing. This context representation indirectly, via the window size, takes into account the distance between words. A HAL matrix [3] allows reflection in one matrix cell as a frequency of pairs vi, vj as the distance dp between these words in context p using the following formula:
H A L i , j = p = 1 P i j ( S d p + 1 ) ,     d i , j < S
where S is the size of the HAL window, and Pi,j is the number of contexts of width S in which the words vi, vj are contained. Each context of width S means a sequence of words of length S. Thus, if two words are contained in the same context of width S, it means that they are contained in some word sequence and the distance between them does not exceed S. Each row of the HAL matrix is the word’s vector and the sum of all rows builds the document’s vector.
The HAL matrix can be illustrated by a simple example. Let us consider the text “Alice likes Bob but Bob hates Alice.” The index after removal of the stop word “but” looks like this: [Alice:0, Likes:1, Bob:2, hates:3]. Assume the HAL window size is equal to 3. Then, the first row of the HAL matrix related to the word “Alice” is [0, 3, 2+1, 0], where 2+1 means two occurrences of pairs “Alice–Bob” with distances 2 and 3, respectively. Thus, the value in a cell of the HAL matrix reflects both the frequency of word pairs and the proximity of these words in the text.
Matrix representation of words makes it convenient to apply the quantum probability theory that operates with density matrices [17]. Moreover, modern studies in sociology demonstrate that quantum probability theory is able to explain the violation of laws of classical logic when interviewing people [18]. In [19], an analogy was discovered between the well-known double-slit experiment with elementary particles and the user’s state in information retrieval. Natural language is highly contextual; that is, an information stream can be represented by the quantum framework where each word or phrase could be modeled by density operators [20].
The aim of our research was to range documents {D} according to their relevance to a query consisting of two words: w1 and w2. First, we built a word index for the domain. Without word segmentation, the word index means all n-grams presented in the domain documents. Second, we built a HAL matrix for each HAL window size. The HAL matrix had the dimensions L × L, where L is the dictionary length. The jth coordinate of the document vector was
ψ j = i = 1 L A L i , j .
Then, we used the formula for the Bell’s test [7]:
S B e l l = | A ^ B ^ + ψ + A ^ x B ^ + ψ | + | A ^ B ^ ψ A ^ x B ^ ψ |
where
A ^ x = [ 0 1 1 0 ] ,     B ^ x = M ^ 1 · A ^ x · M ^ ,
  M ^ = p 1 p 2 1 p 2 p
  B ^ + = B ^ + B ^ x 2 ,      B ^ = B ^ B ^ x 2 ,
A ^ = [ 1 0 0 1 ] ,     B ^ = M ^ 1 · A ^ · M ^ ,
o p e r a t o r ψ = b r a | o p e r a t o r | k e t ,
b r a = [ α    β ] ,      k e t = [ α β ]
where = u B | u A , u A ,     u B are the normalized vectors of query words A and B, and α, β are projections of the document vector to vectors u A   ,   u B . Each addend in formula (3) can obtain values in {–1,1}, so the maximal value of Sbell = 4, but according to Tsirelson’s bound [20], it cannot exceed 2√2. Values between 2 and 2√2 mean a high correlation of the query words in the domain context.
A simple illustration of formula (3) is presented in the Figure 1, where W 1 ,   W 2 are the vectors of words. The case “a” can be considered as maximizing the expression (3). It shows the margin case for the formula S bell = | 2 2 + 2 2 | + | 2 2 ( 2 2 ) | = 2 2 . It violates Bell’s inequation and reflects a high semantic link of two words in the context. Cases “b” and “c” mean the negation and identity of words, respectively, and yield the value of the expression S bell = 2 . Any link missing between the words shows the “d” case and S bell tends to be less than two.

3. Results

For the experimental data, we used the textbook “Geology“ (www.baike.com/wiki/地质学基础), where each chapter was considered as a separate document. Thus, these documents corresponded to the same domain but to different topics. The task of the experiment was to range these topics according to their relevance to the query consisting of two words: 火山 (volcano) and 岩石 (rock). Each query word consisted of two syllables, so we split the documents to overlapping 2-grams, as shown in Section 2. The statistics of query word occurrence are listed in Table 1.
According to the occurrence statistics, chapters 3, 6, and 7 should be the most relevant to the query. However, the Bell’s test curves shown in Figure 2 demonstrate different results: chapters 3, 2, 1, and 8 are the most relevant.
The most closely related to the query was the chapter “Igneous rocks,” which corresponded to the frequency of query words in the document. The chapter “Minerals” only had two occurrences of the word “volcano,” but it contained many facts about volcanic rocks. Ranked third and fourth were the chapters “Introduction” and “Earthquake,” which also had few instances of the query words but higher values of the Bell’s test. “Introduction” has many phrases related to different types of rocks, and “Earthquake” discusses volcanic activity, which is closely related to the topic “volcanic rocks.” On the contrary, the TF-IDF analysis (Figure 3), along with “Earthquake” and “Introduction,” highlighted as relevant the chapter “Metamorphic Rock,” which was not actually relevant. The TF-IDF measure is a numerical statistic that reflects how important a word is to a document in a corpus. This measure is often used as a weighting factor of words. It increases proportionally to the number of times a word appears in the document and decreases if the word is a general term (like conjunctions or stop-words). It this paper the TF-IDF statistic is used for context vectors with cosine distance for them as an alternative to the measure based on HAL and Bell’s test.
In the second experiment, we used the textbook “History of Natural Science” (book.douban.com/subject/3428495/), which was also divided into eight topics. The occurrences of the query words 实验 (experiment) and 科学 (science) are shown in Table 2. Based on the query word frequency, documents 7, 5, and 4 were expected to be the most relevant.
Like the first experiment, the Bell’s test results (Figure 4) differed from the ones based on the term frequency: the most relevant documents were found to be 1, 6, and 5. Similarly, the chapter “Introduction” contains many sentences related to scientific methods, including experiments. The high level of the Bell’s test for the chapter “Medicine” reflected the empirical nature of early medical research. The chapter “Descartes’s Math and Philosophy” appeared irrelevant to the query, even though it contains a description of some philosophical mental experiments. Chapters 2 and 3 have many facts about experiments, but they are not explicitly indicated as scientific ones.
The TF-IDF analysis of the domain “History of Science” demonstrated significant nonmonotonic behavior, as shown in Figure 5. These data cohered with the term occurrences listed in Table 2 but did not reflect the sense of the chapters. Particularly, according to the TF-IDF analysis, the most relevant chapter is “Orient. and Europe Middle Ages,” which contains only a few references to scientific experiments.
The third domain considered by the experiments was “Psychology” and the source of the texts was the Chinese version of the book “Evolutionary Psychology” by David Buss [21]. Figure 6 demonstrates the results of the Bell’s test for the query “natural selection.” One can see that the highest violation of the Bell’s test was for the book chapters “Women’s Long-Term Partnership Strategy,” "Short-Term Sexual Strategy," "Men’s Long-Term Partnership Strategy," and “Introduction.” The chapters "Conflicts and Wars," "Parent Problems," "Overcoming Bad Environment—Human Survival," and "Cooperation–Alliance" had the lowest relevance. This ranking was satisfactory according to the expert assessment.
According to the TF-IDF analysis within the window (Figure 7), the most relevant chapter to the query “natural selection” is “Introduction.” Other chapters had quite close TF-IDF values. The minimal TF-IDF value was for the chapter “Men’s Long-Term Partnership Strategy.”
The comparison of the Bell’s test and TF-IDF analysis allowed us to conclude that Bell’s test distinguishes relevant and nonrelevant topics more clearly and properly than TF-IDF. The table below shows TF-IDF and Bell’s test comparison to summarize this observation. Table 3 contains the computed discounted cumulative gain (DCG) measure for each domain of the experiment. This measure shows the ranking quality of an algorithm. The formula for the DCG measure is
  D C G = i = 1 p r e l i l o g 2 ( i + 1 )
where p is the number of documents (or positions), i is position of a document in order of algorithm estimation, and reli is the estimation of relevance for the ith position (it can be 0 for an irrelevant document or 1 for a relevant document).
In accordance with our experiment, Bell’s test as an algorithm for relevance estimation showed a greater value of the DCG measure than the TF-IDF algorithm. This experiment demonstrates the validity of using Bell’s test for document ranking. A comparison of Bell’s test as a ranking algorithm and other algorithms (not only TF-IDF) with bigger test samples can be considered as a direction for further research.

4. Discussion

The Bell’s test applied to a pair of words in a document is a measure of the “entanglement” of two words in the text. The results of our experiments for Chinese texts demonstrate that the Bell’s test, in general, allows ranking the documents according to their relevance to the query. Unlike the results found for English texts [6], where the values of Bell’s test increased by expanding the HAL window, the Bell’s test for Chinese texts demonstrated the opposite behavior. In a narrow context, all the documents had the value Sbell > 2. This phenomenon could be explained by the wide polysemy of ideograms where narrow contexts (short HAL window) look similar. This effect is strengthened because of the avoidance of word segmentation where fragments of words are considered as separate words. With a wider context, the values of Bell’s test for the considered documents differed. Thus, the criterion of the relevance of a document to a query could be the point of crossing the Bell’s curve at the line Sbell = 2. The later this point, the more relevant the document.

5. Conclusions and Future Work

The experimental study on two subject areas demonstrated the validity of Bell’s test for Chinese texts without word segmentation with one restriction: Bell’s test is not valid for a narrow context. The Bell’s test is informative for HAL window size values greater than 20-30. The advantage of Bell’s test for Chinese texts is its ability to find relevance for a query even for documents containing a few instances of the query words. The disadvantage of the proposed approach is the necessity of using, as a query, only a couple of n-grams of the same length.
Future work might improve the method of calculation of the document vector. Presently, it is implemented as the sum of rows of the HAL matrix (2). Further research should also emphasize searches using more than two-word queries. The current approach is based on the quantum-like entanglement of two query words, similar to how quantum mechanics deals only with two particles. Another problem that is not discussed here is the search speed [20]. The time it takes to build the document’s vector is several seconds and this is critical if one were to process many documents.

Author Contributions

Conceptualization, I.A.B.; data curation, X.H.; investigation, A.V.P.; software, A.V.P.; validation, J.A.K. and C.Y.; writing—review and editing, J.A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This work was supported in part by a grant from ITMO University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dong, H.; Hussain, F.; Chang, E. A survey in traditional information retrieval models. In Proceedings of the 2008 2nd IEEE International Conference on Digital Ecosystems and Technologies, Phitsanulok, Thailand, 26–29 February 2008. [Google Scholar] [CrossRef]
  2. Dumais, S.T. Latent Semantic Analysis. Ann. Rev. Inf. Sci. Technol. 2005, 38, 188–230. [Google Scholar] [CrossRef]
  3. Jones, M.N.; Willits, J.; Dennis, S. Models of Semantic Memory. The Oxford Handbook of Computational and Mathematical Psychology; Oxford University Press: Oxford, UK, 2015; pp. 232–254. [Google Scholar] [CrossRef] [Green Version]
  4. Arndt, M.; Juffmann, T.; Vedral, V. Quantum Physics Meets Biology. HFSP J. 2009, 3, 386–400. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Arfi, B. Challenges to a Quantum-Theoretic Social Theory. Millenn. J. Int. Studies 2018, 47, 99–113. [Google Scholar] [CrossRef]
  6. Haven, E.; Khrennikov, A. Quantum Social Science; Cambridge University Press: Cambridge, UK, 2013; ISBN 9781139003261. [Google Scholar]
  7. Busemeyer, J.R.; Bruza, P.D. Quantum Models of Cognition and Decision; Cambridge University Press: Cambridge, UK, 2012; ISBN 9780511997716. [Google Scholar]
  8. Khrennikov, A. Basics of Quantum Theory for Quantum-Like Modeling Information Retrieval; Quantum-Like Models for Information Retrieval and Decision-Making, STEAM-H: Science, Technology, Engineering, Agriculture, Mathematics & Health; Aerts, D., Khrennikov, A., Melucci, M., Toni, B., Eds.; Springer: Cham, Switzerland, 2019; pp. 51–82. [Google Scholar] [CrossRef]
  9. Barros, J.; Toffano, Z.; Meguebli, Y.; Doan, B.L. Contextual Query Using Bell Tests; Quantum Interaction, QI 2013, Lecture Notes in Computer Science; Atmanspacher, H., Haven, E., Kitto, K., Raine, D., Eds.; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8369. [Google Scholar] [CrossRef] [Green Version]
  10. Platonov, A.V.; Bessmertny, I.A.; Semenenko, E.K.; Alodjants, A.P. Non-Separability Effects in Cognitive Semantic Retrieving; Quantum-Like Models for Information Retrieval and Decision-Making, STEAM-H: Science, Technology, Engineering, Agriculture, Mathematics & Health; Aerts, D., Khrennikov, A., Melucci, M., Toni, B., Eds.; Springer: Cham, Switzerland, 2019; pp. 35–40. [Google Scholar] [CrossRef]
  11. Sung, Y.T.; Chang, T.H.; Lin, W.C. CRIE: An automated analyzer for Chinese texts. Behav. Res. 2016, 48, 1238. [Google Scholar] [CrossRef] [PubMed]
  12. Zhihan, L.; Xu, Y.; Geva, S. A Hybrid Chinese Information Retrieval Model. In Proceedings of the 6th International Conference on Active Media Technology, Toronto, ON, Canada, 28–30 August 2010; pp. 267–276. [Google Scholar] [CrossRef] [Green Version]
  13. Zeng, B.; Yao, L.; Wang, R. A Chinese Document Retrieval Method Considering Text Order Information. In Proceedings of the 2017 International Conference on Computer, Electronics and Communication Engineering, Sanya, China, 25 June 2017; pp. 605–610. [Google Scholar] [CrossRef] [Green Version]
  14. Bai, B.R.; Chen, B.; Wang, H.M. Syllable-based chinese text/spoken document retrieval using text/speech queries. Int. J. Pattern Recognit. Artif. Intell. 2000, 14, 603–616. [Google Scholar] [CrossRef] [Green Version]
  15. Yu, C.; Bessmertny, I.A. Contrastive Domain Term Extraction from Chinese Texts without Word Segmentation. In Proceedings of the International Conference on Advanced Education and Management Sciences (AEMS-2017), Bangkok, Thailand, 26–27 March 2017; pp. 50–55. [Google Scholar] [CrossRef] [Green Version]
  16. Turney, P.D.; Pantel, P. From Frequency to Meaning: Vector Space Models of Semantics. J. Artif. Intell. Res. 2010, 37, 29–34. [Google Scholar] [CrossRef]
  17. Wittek, P.; Darányi, S. Spectral Composition of Semantic Spaces; Quantum Interaction, QI 2011, Lecture Notes in Computer Science; Song, D., Melucci, M., Frommholz, I., Zhang, P., Wang, L., Arafat, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; Volume 7052. [Google Scholar] [CrossRef] [Green Version]
  18. Meguebli, Y.; Popineau, F.; Doan, B.L. A Novel Architecture for a Smart Information Retrieval System Based on Opinions Engineering. In Proceedings of the Symposium on Human-Computer Interaction and Information Retrieval, Cambridge, MA, USA, 2012. [Google Scholar]
  19. Zuccon, G. An analogy between the double slit experiment and document ranking. Proceedings of The 3rd IRSG Symposium: Future Directions in Information Access, Padua, Italy, 31 August–1 September 2009; p. 9. [Google Scholar] [CrossRef]
  20. Melucci, M. An Efficient Algorithm to Compute a Quantum Probability Space. IEEE Trans. Knowl. Data Eng. 2019, 31, 1452–1463. [Google Scholar] [CrossRef]
  21. David, B. Evolutionary Psychology: The New Science of the Mind, 4th ed.; Pearson: London, UK, 2011; ISBN 978-0205015627. [Google Scholar]
Figure 1. Graphic interpretation of Bell’s test.
Figure 1. Graphic interpretation of Bell’s test.
Entropy 22 00275 g001
Figure 2. Bell’s test for the domain “Geology.”
Figure 2. Bell’s test for the domain “Geology.”
Entropy 22 00275 g002
Figure 3. TF-IDF analysis for the domain “Geology.”
Figure 3. TF-IDF analysis for the domain “Geology.”
Entropy 22 00275 g003
Figure 4. Bell’s test for the domain “History of Science.”
Figure 4. Bell’s test for the domain “History of Science.”
Entropy 22 00275 g004
Figure 5. TF-IDF analysis for the domain “History of Science.”
Figure 5. TF-IDF analysis for the domain “History of Science.”
Entropy 22 00275 g005
Figure 6. Bell’s test for the domain “Psychology.”
Figure 6. Bell’s test for the domain “Psychology.”
Entropy 22 00275 g006
Figure 7. TF- IDF analysis for the domain “Psychology.”
Figure 7. TF- IDF analysis for the domain “Psychology.”
Entropy 22 00275 g007
Table 1. Term occurrence in the documents of the domain “Geology.”
Table 1. Term occurrence in the documents of the domain “Geology.”
No.Document火山 (volcano)岩石 (rocks)
1Introduction751
2Minerals226
3Igneous rock238122
4Sedimetary rock38193
5Metamorphic rock4148
6Mineral deposit4448
7Structural movement and structural changes4344
8Earthquake1831
Table 2. Term occurrence in the documents of the domain “History of science.”
Table 2. Term occurrence in the documents of the domain “History of science.”
NoDocument实验 (experiment)科学(science)
1Introduction651
2Orient. and Europe Middle ages1973
316th and 17th Cent. Scientific revolution1946
4Galileo and mechanics3027
5Descartes’s math and philosophy35110
6Medicine1956
717th Cent. science58189
818th Cent. mechanical sciences644
Table 3. Computed discounted cumulative gain for each domain.
Table 3. Computed discounted cumulative gain for each domain.
Domain NameTF-IDF DCGBell’s Test DCG
Geology1.081.70
History of science1.242.13
Psychology1.672.38

Share and Cite

MDPI and ACS Style

Bessmertny, I.A.; Huang, X.; Platonov, A.V.; Yu, C.; Koroleva, J.A. Applying the Bell’s Test to Chinese Texts. Entropy 2020, 22, 275. https://doi.org/10.3390/e22030275

AMA Style

Bessmertny IA, Huang X, Platonov AV, Yu C, Koroleva JA. Applying the Bell’s Test to Chinese Texts. Entropy. 2020; 22(3):275. https://doi.org/10.3390/e22030275

Chicago/Turabian Style

Bessmertny, Igor A., Xiaoxi Huang, Aleksei V. Platonov, Chuqiao Yu, and Julia A. Koroleva. 2020. "Applying the Bell’s Test to Chinese Texts" Entropy 22, no. 3: 275. https://doi.org/10.3390/e22030275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop