Next Article in Journal
SAM-Based Input Augmentations and Ensemble Strategies for Image Segmentation
Previous Article in Journal
AI-Based Digital Twins of Students: A New Paradigm for Competency-Oriented Learning Transformation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Revealing Short-Term Memory Communication Channels Embedded in Alphabetical Texts: Theory and Experiments

by
Emilio Matricciani
Dipartimento di Elettronica, Informazione e Bioingegneria (DEIB), Politecnico di Milano, 20133 Milan, Italy
Information 2025, 16(10), 847; https://doi.org/10.3390/info16100847
Submission received: 20 August 2025 / Revised: 24 September 2025 / Accepted: 29 September 2025 / Published: 30 September 2025

Abstract

The aim of the present paper is to further develop a theory on the flow of linguistic variables making a sentence, namely, the transformation of (a) characters into words; (b) words into word intervals; and (c) word intervals into sentences. The relationship between two linguistic variables is studied as a communication channel whose performance is determined by the slope of their regression line and by their correlation coefficient. The mathematical theory is applicable to any field/specialty in which a linear relationship holds between two variables. The signal-to-noise ratio Γ is a figure of merit of a channel being “deterministic”, i.e., a channel in which the scattering of the data around the regression line is negligible. The larger Γ is, the more the channel is “deterministic”. In conclusion, humans have invented codes whose sequences of symbols that make words cannot vary very much when indicating single physical or mental objects of their experience (larger Γ ) . On the contrary, large variability (smaller Γ ) is achieved by introducing interpunctions to make word intervals, and word intervals make sentences that communicate concepts. This theory can inspire new research lines in cognitive science research.

1. Introducing an Equivalent Input–Output Model of Short-Term Memory

Humans can communicate and extract meaning both from spoken and written languages. Whereas the sensory processing pathways for listening and reading are distinct, listeners and readers appear to extract very similar information about the meaning of a narrative story—heard or read—because the brain assimilates a written text like the corresponding spoken/heard text [1]. In the following, therefore, we consider the processing of reading or writing a text—a writer is also a reader of his/her own text—the same due to the same brain activity. In other words, the human brain represents semantic information in a modal form, independently of input modality.
How the human brain analyzes parts of a sentence (parsing) and describes their syntactic roles is still a major question in cognitive neuroscience. In references [2,3], we proposed that a sentence is elaborated by short-term memory (STM) with two independent processing units in series (equivalent surface processors) of similar size. The clues for conjecturing this input–output model emerged by considering many novels belonging to the Italian and English literature. In reference [3], we showed that there are no significant mathematical/statistical differences between the two literary corpora, according to the so-called surface deep language parameters, suitably defined.
The model conjectures that the mathematical structure of alphabetical languages—digital codes created by the human mind for communication—seems to be deeply rooted in humans, independently of the particular language used or historical epoch. The complex and inaccessible mental process lying beneath communication—still largely unknown—can be studied by looking at the input–output functioning revealed by the structure of alphabetical languages.
The first processor is linked to the number of words between two contiguous interpunctions, denoted by the variable   I p —termed word interval (Appendix A lists the mathematical symbols used in the present article)—approximately ranging in Miller’s 7 ± 2 law range [4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]. The second processor is linked to the number M F   of   I p ’s contained in a sentence, referred to as the extended short-term memory (E–STM), ranging approximately from 1 to 6. These two units can process sentences containing approximately 8.3 to 61.2 words, values that can be converted into time by assuming a reading speed. This conversion gives 2.6 ~ 19.5 s for a fast reader [14] and 5.3 ~ 30.1 s for a reader of novels, values well supported by experiments [15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30].
The E–STM must not be confused with the intermediate memory [31,32]. It is not modeled by studying neuronal activity, but by studying only the surface aspects of human communication, due, of course, to neuronal activity, such as words and interpunctions, whose effects writers and readers have experienced since the invention of writing. In other words, the model proposed in references [2,3] describes the “input–output” characteristics of STM. In reference [33], we further developed the theory by including an equivalent first processor that memorizes syllables and characters to produce a word.
In conclusion, in references [2,3,33], we have proposed an input–output model of STM, made of three equivalent linear processors in series which independently process (1) syllables and characters to make a word, (2) words and interpunctions to make a word interval, and (3) word intervals to make a sentence. This is a simple but useful approach because the multiple brain process regarding speech/texts is not yet fully understood; characters, words, and interpunctions—the latter are needed to distinguish word intervals and sentences—can be easily studied [34,35,36]. Moreover, the theory can inspire new research lines in cognitive science.
In other words, the model conjectures that the mathematical structure of alphabetical languages is deeply rooted in humans, independently of the particular language used or historical epoch. The complex and inaccessible mental process lying beneath communication—still largely unknown—is revealed by looking at the input–output functioning built in in alphabetical languages of any historical epoch.
The literature on STM and its various aspects is immense and multidisciplinary; we have recalled above only few references, but nobody—as far as we know—has considered the connections we found and discussed in references [2,3,33]. Our modeling of STM processing by three units in series is new.
A sentence conveys meaning, of course; therefore, the theory we have developed might be one of the necessary starting points to arrive at the Information Theory that will finally include meaning.
Today, many scholars are trying to arrive at a “semantic communication” theory or “semantic information” theory, but the results are still, in our opinion, in their infancy [37,38,39,40,41,42,43,44,45]. These theories, i.e., those concerning STM, have not considered the main “ingredients” of our theory, namely the number of characters per word, C P , I P , and M F , parameters that anybody understands and can calculate in any alphabetical language [34,35,36], as a starting point for including meaning. This is still a very open issue.
The aim of the present paper is twofold: (a) to further develop the theory proposed in references [2,3,33], and (b) to apply it to the flow of linguistic variables making a sentence. This “signal” flow is built in in the model proposed in reference [33], namely, the transformation of (a) characters into words, (b) words into word intervals, and (c) word intervals into sentences, according to Figure 1. Since the connection between these linguistic variables is described by regression lines [34,35,36], in the present article, we analyze experimental scatterplots between these variables.
The article is divided in two parts. In the first part—from Section 2, Section 3 and Section 4—we recall and further develop the theory of linear channels [2,3,33]; in the second part—from Section 5, Section 6, Section 7 and Section 8—we apply it to a significant database of literary texts.
The database of literary texts considered is a large set of the New Testament (NT) books, namely the Gospels according to Matthew, Mark, Luke, John, the Book of Acts, the Epistle to the Romans, and the Apocalypse—155 chapters in total, according to the traditional subdivision of these texts. We have considered the original Greek texts and their translation to Latin and to 35 modern languages, texts partially studied in reference [35]. Notice that in this paper, “translation” is indistinguishable from “language” because we deal only with one translation per language.
We consider the NT books and their modern translations for two reasons: (a) they tell the same story, and therefore, it is meaningful to compare the translations in different languages; (b) they use common words—not the words of scientific/academic disciplines—therefore, they can give some clues on how most humans communicate.
After this introductory section, Section 2 presents the theory of linear regression lines and associated communication channels; Section 3 presents the connection of single linear channels; Section 4 proposes and discusses the theory of series connection of single channels affected by noise; Section 5 reports an exploratory data analysis of the NT texts; Section 6 reports findings concerning single channels; Section 7 concerns series connection of channels; Section 8 concerns cross channels; and finally, Section 9 summarizes the main findings and indicates future studies.

2. Theory of Linear Regression Lines and Associated Communication Channels

In this section, we recall and further expand the general theory of stochastic variables linearly connected, originally developed for linguistic channels [35,36] but applicable to any other field/specialty in which a linear relationship holds between two variables.
Let x (independent variable) and y (dependent variable) be linked by the following line:
y = m x + b
Notice that Equation (1) models a deterministic relationship through the slope m and the intercept b . Since in most scatterplots between linguistic variables b = 0 , in the following we assume
b = 0 .
However, notice that if b 0 , the theory can be fully applied by defining a new dependent variable y ˇ = y b .
In general, the relationship between x and y is not deterministic, i.e., given by Equation (1), but stochastic (random). Equation (1) models, in fact, two variables perfectly correlated—correlation coefficient r = 1 —characterized by a multiplicative “bias” m . In general, however, these conditions do not hold. Therefore, Equation (1) can be written as follows:
y = m x + n
In Equation (3), n is an additive Gaussian stochastic variable with zero mean value [34,35,36]; therefore, Equation (3) models a noisy linear channel. Notice that n must not be confused with the intercept b .
Figure 2 shows the flow chart describing Equations (1) and (3) with a system/channel representation. The black box indicated with m represents the deterministic channel, i.e., Equation (1); the black box indicated with r represents the parallel channel due to the scattering of y around the regression line. The additive noise n is a Gaussian stochastic variable with zero mean that makes the linear channel partially stochastic, namely “noisy”.
Now, let us consider the following:
(a)
The variance of the difference between the values calculated with Equation (1) m 1 and those calculated with y = x m = 1 (45° line) at a given x value as the “regression noise” power N m [35]. This “noise” is due to the multiplicative bias between the two variables.
(b)
The variance of the difference between the values not lying on Equation (1) ( r 1 ) and those lying on it ( r = 1 ) as the “correlation noise” power N r [35]. This “noise” is due to the spread of y around the line given by Equation (1), modeled by n .
(c)
Let s x 2 and s y 2 be the variances of x and y
In case (a), we obtain the difference m 1 x ; therefore, the variance (or power) of the values lying on the regression line, regression noise, is given by the following:
N m = m 1 2 s x 2
Now, we define the regression noise-to-signal power ratio (NSR), R m , as follows:
R m = N m s x 2 = m 1 2
In case (b), the fraction of the variance s y 2 due to the values of y not lying on the regression line (correlation noise power,   N r ) is given by the following [46]:
N r = 1 r 2 s y 2
The parameter r 2 is called the coefficient of determination and it is proportional to the variance of y explained by the regression line [46]. However, this variance is correlated with the slope m because the fraction of the variance s y 2 due to the regression line, namely r 2 s y 2 , is related to m according to the following [46]:
r 2 s y 2 = m 2 s x 2 .
Figure 3 shows the flow chart of variances.
Therefore, inserting Equation (7) in Equation (6), we obtain the correlation NSR, R r :
R r = N r s x 2 = ( 1 r 2 ) r 2   m 2
Now, since the two noise sources are disjoint, the total NSR R of the channel shown in Figure 2 and Figure 3 is given by the following:
R = R m + R r
Therefore, R depends only on the two parameters m and r of the regression line:
R = m 1 2 + ( 1 r 2 ) r 2   m 2
Finally, the signal-to-noise ratio (SNR) γ is given by the following:
γ = 1 R = 1 ( m 1 ) 2 + 1 r 2 r 2 m 2
In decibels, this is written as follows:
Γ = 10 × l o g 10 γ
Of course, we expect that no channel can yield r = 1   and m = 1 ; therefore, γ = . In empirical scatterplots, it is very likely that r < 1 ,   m 1 .
In conclusion, the slope m measures the multiplicative “bias” of the dependent variable y compared to the independent variable x in the deterministic channel; the correlation coefficient r   measures how “precise” the linear best fit is.
Finally, notice the more direct and insightful analysis that can be achieved by using the NSR instead of the more common SNR because, in Equation (9), the single-channel NSRs simply add together. This makes it easy to study, for example, which addend determines R , and thus Γ , while this is far less easy with Equation (11). Moreover, this choice also leads to a useful graphical representation of Equation (10) that can guide analysis and design [11], as shown in Section 8.
In the next sections, we apply the theory of linear channel modeling to specific cases.

3. Connection of Single Linear Channels

We first study how the output variable y k of channel k relates to the output variable y j of another similar channel j for the same input x . This channel is termed “cross channel” and it is fundamental in studying language translation [35]. Secondly, we study how the output of a deterministic channel, modeled by Equation (1), relates to the output of its stochastic version, Equation (3).

3.1. Cross Channels

Let us consider a scatterplot k and a scatterplot j in which the independent variable x and the dependent variable y are linked by linear regression lines:
y k = m k x k
y j = m j x j
As discussed in Section 2, Equations (13) and (14) do not give the full relationship between the two variables because they link only conditional average values, measured by the slopes m k and m j in the deterministic channels. According to Equation (3), we can write more general linear relationships by considering the scattering of the data, always present in experiments, modeled by additive Gaussian zero mean noise sources n k and n j .
y k = m k x k + n k ,
y j = m j x j + n j .
Now, we can develop a series of interesting investigations about these equations. By eliminating x , we can compare the dependent variable y j of Equation (16) to the dependent variable y k of Equation (15) for x k = x j = x . In doing so, we can find the regression line and the correlation coefficient of the new scatterplot linking y j to y k without the availability of the scatterplot itself.
By eliminating x between Equations (15) and (16), we obtain the following:
y j = m j m k y k m j m k n k + n j
Compared to the new independent variable y k , the slope m k j of the regression line is given by the following:
m k j = m j / m k
Because the two Gaussian noise sources are independent and additive, the total noise is given by the following:
n k j = m j m k n k + n j = m k j n k + n j
Figure 4 shows the flow chart describing the cross channel.
Now, from Equation (18), the R m of the new channel is
R m = ( m k j 1 ) 2 .
The unknown correlation coefficient r k j between y j and y k is given by the following [35]:
r k j = c o s a r c o s ( r j ) a r c o s ( r k )
Therefore, the R r of the new channel is
R r = 1 r k j 2 r k j 2 m k j 2 .
In conclusion, in the new channel connecting y j to y k , we can determine the slope and the correlation coefficient of the scatterplot between y j and y k for the same value of the independent variable x . Now, the availability of this scatterplot is experimentally very rare because we are unlikely to find values of y k and y j for exactly the same value of x ; therefore, cross channels can reveal relationships very difficult to discover experimentally.
In the next sections, we further develop the theory of linear channels, originally established in reference [35], for cross channels.

3.2. Stochastic Versus Deterministic Channel

We compare a deterministic channel k with a stochastic channel j derived from channel k by adding noise. In other words, we start from the regression line given by Equation (1) and then add noise n due to the correlation coefficient r 1 . Therefore, from the theory of stochastic channels discussed in Section 3.1, we obtain
y k = m k x k ,
y j = m k x k + n k ,
m k j = m k / m k = 1 ,
R m = ( m k j 1 ) 2 = 0 ,
r k j = c o s a r c o s r j a r c o s 1 = c o s a r c o s ( r j ) = r j = r ,
R = R r = 1 r 2 r 2 .
In conclusion, in transforming a deterministic channel into a stochastic channel, only the correlation noise is present; therefore, the SNR is given by
γ = r 2 1 r 2 .
Equation (29) coincides with the ratio between the variance explained by the regression line (proportional to the coefficient of determination r 2 ) and the variance due to the scattering (correlation noise), proportional to 1 r 2 [46].
So far, we have considered single channels. In the next section, we consider the series connection of single channels to determine the SNR of the overall channel.

4. Series Connection of Single Channels Affected by Correlation Noise

In this section, we consider a channel made of a series of single channels. We consider this case because it can be found in many specialties, and because in Section 7 we apply it to specific linguistic channels.
Figure 5 shows the flow chart of three single channels in series. These channels can be characterized as in Section 3.2, i.e., only with the correlation noise; therefore, the overall channel is compared to the deterministic channel in which
m = m 1 m 2 m 3 .
From Figure 5, it is evident that the output noise of a preceding channel produces additive noise at the output of the next channel in series. The purpose of this section is to calculate R at the output of the series of channels.
Theorem 1.
The NSR  R  of  n  linear channels in series, each characterized by the correlation noise-to-signal ratio  R i , is given by
R = i = 1 n R i .
Proof. 
Let the three linear relationships of the isolated channels of Figure 5 (i.e., before connecting them in series) be given by
y = m 1 x + n 1 ,
z = m 2 y + n 2 ,
t = m 3 z + n 3 .
Let s x 2 , s y 2 , s z 2 be the variances (power) of the variables, and let N 1 = N 1 r , N 2 = N 2 r , N 3 = N 3 r be the variances (power) of the Gaussian zero mean noise n 1 , n 2 , n 3 ; then, the NSRs of the isolated channels are given by
R 1 = N 1 s y 2 = N 1 m 1 2 s x 2 ,
R 2 = N 2 s z 2 = N 2 m 2 2 s y 2 ,
R 3 = N 3 s t 2 = N 3 m 3 2 s z 2 .
When the first two blocks are connected in series, the input to the second block must also include the output noise of the first block; therefore, from Equations (31) and (33), we obtain the modified output variable z ˘ :
z ˘ = m 2 y = m 2 m 1 x + n 1 + n 2 = m 2 m 1 x + m 2 n 1 + n 2 .
In Equation (37), m 2 m 1 x is the output “signal” and m 2 n 1 + n 2 is the output noise; therefore, the NSR at the output of the second block is
R = m 2 2 N 1 + N 2 m 2 2 m 1 2 s x 2 = m 2 2 N 1 m 2 2 m 1 2 s x 2 + N 2 m 2 2 m 1 2 s x 2 = N 1 m 1 2 s x 2 + N 2 m 2 2 s x 2 = R 1 + R 2 .
Now, for three channels in series, it is sufficient to consider R given by Equation (38) as the input NSR to the third single channel to obtain the final NSR and prove Equation (31):
R = R 1 + R 2 + R 3 .
Finally, notice that R of Equation (30) is proportional to the mean < R i > :
R = n i = 1 n R i / n = n < R i > .
In other words, the series channel averages the single R i .
In conclusion, Equations (30)–(40) allow us to study channels made by the series of several single channels affected by correlation noise by simply adding together their single NSRs.
In the next sections, we apply the theory to linguistic channels suitably defined, after exploring the database on the NT mentioned in Section 1.

5. Exploratory Data Analysis

In this second part, we explore the linear relationships between characters, words, interpunctions, and sentences, according to the flow chart shown in Figure 1, of the New Testament books considered (Matthew, Mark, Luke, John, Acts, Epistle to the Romans, Apocalypse). This is the database of our experimental analysis and application of the theory of linear channels discussed in the previous sections.
Table 1 lists the language of translation and language family, including the total number of characters ( C ) , words ( W ) , sentences ( S ), and interpunctions ( I ).
Figure 6 shows the scatterplots in the original Greek texts between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) characters and sentences. Figure 7 shows these scatterplots in the English translation. Appendix B shows example of scatterplots in other languages. Table 2 reports the slope m and correlation coefficient r of the indicated scatterplots (155 samples for each scatterplot) for each translation, namely the input parameters of our theory on communication channels. The differences between languages are due to the large “domestication” of the original Greek texts discussed in reference [47].
The four scatterplots define fundamental linear channels and they are connected with important linguistic parameters previously studied [34,35,36], namely the following:
(a)
The number of characters per word, C P , given by the ratio between characters (abscissa) and words (ordinate) in Figure 6a.
(b)
The number of words between two successive interpunctions, I P —called the word interval—given by the ratio between interpunctions (abscissa) and words (ordinate) in Figure 6b.
(c)
The number of word intervals in sentences, M F , given by the ratio between sentences (abscissa) and interpunctions (interpunctions) in Figure 6c.
Figure 6d shows the scatterplot between characters and sentences, which will be discussed in Section 7.
In the next section, we study the channels corresponding to these scatterplots.
Figure 8 shows the probability distributions of the correlation coefficient r and the coefficient of determination r 2 for the scatterplots: words versus characters (green line); interpunctions versus words (cyan); and sentences versus interpunctions (magenta). The black line refers to the scatterplot sentences versus characters; the red line refers to the series channel considered in Section 7, which links characters to sentences.
For correlation coefficients—and consequently for the coefficient of determination, which determines the SNR—we notice the following remarkable findings:
(a)
In any language, the largest correlation coefficient is found in the scatterplot between characters and words. The communication digital codes invented by humans show remarkable strict relationships between digital symbols (characters) and their sequences (words) to indicate items of their experience, material or immaterial. Languages do not differ from each other very much, with r in the range 0.9753–0.9983 (Armenian, Cebuano) and overall r = 0.9925 ± 0.0038
(b)
The smallest correlation coefficient is found in the scatterplot between characters and sentences, being overall 0.0140 ± 0.0027 . This relationship must be, of course, the most unpredictable and variable because the many digital symbols that make a sentence can create an extremely large number of combinations, each delivering a different concept.
(c)
The correlation coefficient (and also the coefficient of determination r 2 ) decreases as characters combine to create words, as words combine to create word intervals, and as word intervals combine to create sentences.
The path just mentioned in item (c) describes an increasing creativity and variety of meaning different from that of the deterministic channel.
The characters-to-words channel shows the smallest r 2 ; therefore, this channel is the nearest to being purely deterministic. It does not tend to be typical of a particular text/writer but more of a language because a writer has very little freedom in using words of very different length [34], if we exclude specialized words belonging to scientific and academic disciplines.
On the contrary, the channels words-to-interpunctions and interpunctions-to-sentences are less deterministic, as a writer can exercise his/her creativity of expression more freely; therefore, these channels depend more on the writer/text than on the language. Finally, the big “jump” from characters to sentences gives the greatest freedom.
In conclusion, humans have invented codes whose sequences of symbols that make words cannot vary very much when indicating single physical or mental objects of their experience. To communicate concepts, on the contrary, a large variability can be achieved by introducing interpunctions to form word intervals and word intervals to form sentences, the final depositary of human basic concepts.
Figure 9 shows the probability distributions of the slope m . The black line (only partially visible because it is superposed to the red line) refers to the scatterplot sentences versus characters; the red line refers to the series channel that connects sentences to characters, as discussed in Section 7.
On the slopes, we notice the following important findings:
(a)
The slope of the scatterplot between interpunctions and sentences (magenta line) is the largest in any language—overall 0.3795 ± 0.0755 —and determines the number of word intervals, M F , contained in a sentence in its deterministic channel.
(b)
The slope of the scatterplot between interpunctions and words (cyan line) determines the length of the word interval, I P , in its deterministic channel.
(c)
The slope of the scatterplot between words and characters (green line) determines the number of characters per word, C P , in its deterministic channel. As discussed below, this channel is the most “universal” channel because, from language to language, C P varies little compared to other linguistic variables.
(d)
The smallest slopes are in the scatterplots between characters and sentences, being overall 0.0140 ± 0.0027 . For example, in English there are 519,043 characters and 6590 sentences (Table 1); now, according to Table 1, the deterministic channel predicts 0.0128 × 519,043 6644 sentences, just +0.8% difference from the true value.
As reiterated above, the slopes describe deterministic channels. As discussed in Section 6, a deterministic channel is not “deterministic” when concerning the number of concepts, because the same number of sentences can communicate different meanings by just changing words and interpunctions. What is “deterministic” is the size of the ensemble.
In the next section, we model single linguistic channels, i.e., channels not yet connected in series, from the linear relationships shown above.

6. Single Linguistic Channels

In this section, we apply the theory developed in Section 3.2 to the scatterplots of Section 5 and therefore to the following single channels:
(a)
Characters-to-words.
(b)
Words-to-interpunctions.
(c)
Interpunctions-to-sentences.
These single channels are modeled like in Figure 2 and Figure 3; they are affected only by the regression noise. Γ is obtained from Equation (29) and drawn in Figure 10. Table 3 reports the mean and standard deviation of Γ in each channel.
From Figure 10 and Table 3, we notice the following interesting facts:
(a)
Languages show different Γ values due to the large degree of domestication of the original Greek texts [47].
(b)
Γ decreases steadily in this order: characters-to-words, words-to-interpunctions, and interpunctions-to-sentences. A decreasing Γ says, to a certain extent, how much less deterministic a channel is.
(c)
Words-to-interpunction and interpunctions-to-sentences have close values, and therefore, they show similar deterministic channels.
(d)
Most languages have a Γ greater than that in Greek. This agrees with the finding that in modern translations of the Greek texts, domestication prevails over foreignization [47].
(e)
Finally, we can consider Γ as a figure of merit of a linguistic channel being deterministic: the larger Γ is, the more the channel is deterministic.
Figure 11 shows histograms (37 samples) of Γ for each channel. The probability density function of Γ can be modeled with a Gaussian model (therefore, γ is a lognormal stochastic variable), with the mean and standard deviation reported in Table 3. Figure 12 shows the probability functions of Γ which show, again, the differences and similarities of the channels.
In conclusion, the large Γ of the characters-to-words channel, in any language, indicates that the transformation of characters into words is the most deterministic.
In the next section, we connect the single channels to obtain the series channels modeled in Figure 5 and study them according to the theory of Section 4.

7. Series Connection of Linguistic Channels Affected by Correlation Noise

Let us connect the three single channels to obtain the series channel shown in Figure 5 and apply the theory of Section 4. We first show the results concerning the theory of series channel, and then we compare the single channel characters-to-sentences to that obtained with the series of single channels.
Figure 13a shows the single NSRs and the series NSRs in linear units for each language; Figure 13b shows the corresponding Γ (dB), partially already reported in Figure 10. We can notice that in the sum indicated in Equation (31), the NSR of the characters-to-words channel is negligible compared to the other two NSRs. For example, in English (language no. 10), R = 0.0152 + 0.1059 + 0.1120 = 0.2331 0.1059 + 0.1120 = 0.2179 ; therefore, R 0.22 against R 0.23 . In general, R 1 R 2 , R 3 , so the characters-to-words channel can be ignored for a first approximation, because it is about 1 / 10 of the other two addends in Equation (31).
For the characters-to-words channel, Figure 14a shows the slope calculated from the scatterplot between characters and sentences (Table 2) and the slope given by Equation (30). The agreement is excellent; in practice, the two values coincide (correlation coefficient 0.9998). Figure 14b shows the scatterplot between the correlation coefficient calculated from the scatterplot between characters and sentences (Table 2) and that calculated by solving Equation (29) for r after calculating γ from Equation (40). In this case, the two values are poorly correlated (correlation coefficient 0.3929). Finally, notice the difference between the probability of Γ calculated by solving Equation (29) for r the red line in Figure 12—and that calculated from the available scatterplots and regression line (Table 2), i.e., the black line. The smoother red curve models more accurately the relationship between characters and sentences than the available scatterplot shown in Figure 6d, because R is proportional to the mean value of the single channels (see Equation (41)).
In conclusion, the Γ calculated in a series channel linking two variables is more reliable than that calculated from a single channel/scatterplot between the two variables.
In the next section, we apply the theory of cross channels of Section 3.1.

8. Cross Channels: Language Translations

In cross channels, we study how the output variable y k of channel k relates to the output variable y j of another similar channel j for the same input x ; therefore, we apply the theory of Section 3.1. In this new channel, we can determine the slope and the correlation coefficient of the scatterplot between y k and y j for the same value of the independent variable x ; therefore, cross channels can reveal relationships more difficult to discover experimentally.
From the database of the NT texts and the scatterplots of Figure 6, we can study at least three cross channels:
(a)
The words-to-words channel, by eliminating characters; therefore, the number of words is compared for the same number of characters.
(b)
The interpunctions-to-interpunctions channel, by eliminating words; therefore, the number of word intervals is compared for the same number of words.
(c)
The sentences-to-sentences channel, by eliminating interpunctions; therefore, the number of sentences is compared for the same number of word intervals.
Now, since these channels connect one independent variable in one language to the same (dependent) variable in another language, they describe very important linguistic channels, namely translation channels, and they can be studied from this particular perspective. Therefore, cross channels in alphabetical texts describe the mathematics/statistics of translation, as we first studied in reference [35].
Figure 15 shows the slope m k j and the correlation coefficient r k j by assuming Greek as language k , namely the reference language, for the three cross channels. We can notice the following:
(a)
For most languages, m k j > 1 in any cross channel; therefore, most modern languages tend to use more words for the same number of characters; more word intervals for the same number of words; and more sentences for the same number of word intervals than Greek. In other words, the corresponding deterministic channel (the channel characterized by a multiplicative slope) is significantly biased compared to the original Greek texts.
(b)
The correlation coefficient r k j is always very near unity. Therefore, the scattering of the data around the regression line is similar in all three cross channels.
Figure 16 shows the findings assuming English as the reference language. In this case, we consider the “translation” from English into the other languages [35]. Clear differences are noticeable:
(a)
Words-to-words channel: For most languages, m k j 1 . The multiplicative bias is small, as languages tend to use the same number of words as in English. This was not the case for Greek. The correlation coefficient r k j   is practically the same for all languages. In other words, modern languages tend to use the same number of words as in English for the same number of characters; therefore, domestication of the alleged translation of English to the other languages is moderate, compared to Greek or Latin (see languages 1 and 2 in Figure 16a).
(b)
Interpunctions-to-interpunctions channel: The multiplicative bias m k j is strong, as in Greek; therefore, the deterministic cross channels are different from language to language. The correlation coefficient r k j is more scattered than in Figure 15 and different from language to language. Curiously, in the channel English-to-Greek, m k j 1 , and there is no bias. The correlation coefficient r k j is similar to that of the sentences-to-sentences channel.
(c)
Sentences-to-sentences: m k j 1 for most languages, and r k j is similar to that of the interpunctions-to-interpunctions channel.
Since similar diagrams can be shown when other modern languages are considered as the independent language, we can conclude that the translation from Greek to modern languages shows a high degree of domestication, due especially to the multiplicative bias, namely to the deterministic channels rather than the stochastic part of the channel. In conclusion, the translation from a modern language into another modern language is mainly achieved through deterministic channels. Therefore, the SNR is mainly determined by R m .
This conclusion is visually evident in the scatterplot between X = R m and Y = R r shown in Figure 17, where a constant value of Γ traces the arch of the circle [34]. It is clear that R m R r ; in other words, in the three cross channels, Γ is dominated by R m , in agreement with what is shown in Figure 15 and Figure 16.
Finally, Figure 18 shows the mean value and standard deviation of Γ in the three channels by assuming the language indicated in abscissa as an independent language/translation.
Notice that, overall, the probability distribution of Γ can be modeled as Gaussian, with the mean value and standard deviation reported in Table 4 (for its calculation, see Appendix C). Notice that cross channels have larger Γ values than the series channels (Table 3), because “translation” between two modern languages uses mostly deterministic channels.
Figure 19 shows, as an example, the modeling of the words-to-words overall channel.
Now, we conjecture the characteristics of the three channels for an indistinct human being by merging all values, as done with words in Figure 19.
Figure 20 shows the Gaussian probability density functions and their probability distributions of the overall Γ in the three channels calculated with the values in Table 4. These distributions refer, therefore, to channels in which all languages merge into a single digital code. In other words, we might consider these probability distributions as “universal”, typical of humans using plain text.
From Figure 18, Figure 19 and Figure 20 and Table 4, the following “universal” characteristics clearly emerge.
(a)
The words-to-words channel is distinguished from the other two channels, with a larger Γ . This channel is the most deterministic.
(b)
The interpunctions-to-interpunctions and sentences-to-sentences channels are very similar both in the mean value and standard deviation of Γ , therefore indicating a similar freedom in creating variations with respect to their deterministic channels.

9. Summary and Conclusions

How the human brain analyzes parts of a sentence (parsing) and describes their syntactic roles is still a major question in cognitive neuroscience. In references [2,3,33], we proposed that a sentence is elaborated by short-term memory with three independent processing units in series: (1) syllables and characters to make a word, (2) words and interpunctions to make a word interval, and (3) word intervals to make a sentence.
This approach is simple but useful, because the multiple processing of the brain regarding speech/text is not yet fully understood but characters, words, and interpunctions—the latter are needed to distinguish word intervals and sentences—can be easily studied in any alphabetical language and epoch. Our conjecture, therefore, is that we can find clues on the performance of the mind, at a high cognitive level, by studying the most abstract human invention, namely alphabetical texts.
The aim of the present paper was to further develop and complete the theory proposed in references [2,3,33] and then apply it to the flow of linguistic variables making a sentence, namely, the transformation of (a) characters into words; (b) words into word intervals; and (c) word intervals into sentences. Since the connection between these linguistic variables is described by regression lines, we have analyzed experimental scatterplots between the variables.
In the first part of the article, we have recalled and further developed the theory of linear channels, which models stochastic variables linearly connected. The theory is applicable to any field/specialty in which a linear relationship holds between two variables.
We have first studied how the output variable y k of channel k relates to the output variable y j of another similar channel j for the same input x . These channels can be termed as “cross channels” and are fundamental in studying language translation.
Secondly, we have studied how the output of a deterministic channel relates to the output of its noisy version. A deterministic channel is not “deterministic” when concerning the number of concepts, because, for example, the same number of sentences can communicate different meanings by just changing words and interpunctions. What is “deterministic” is the size of the ensemble.
Then, we have studied a channel made of a series of single channels and have established that its noise-to-signal ratio Γ is proportional to the average of the single-channel noise-to-signal ratios.
In the second part of the article, we have explored, experimentally, the linear relationships between characters, words, interpunctions, and sentences in a large set of the New Testament books. We have considered the original Greek texts and their translation to Latin and to 35 modern languages because, in any language, they tell the same story; therefore, it is meaningful to compare their translations. Moreover, they use common words; therefore, they can give some clues on how most humans communicate.
The characters-to-words channel is the nearest to being purely deterministic. It does not tend to be typical of a particular text/writer but more of a language because a writer has very little freedom in using words of very different length.
On the contrary, the channels words-to-interpunctions and interpunctions-to-sentences are less deterministic, as they depend more on writer/text than on language.
The signal-to-noise ratio Γ is a figure of merit of the deterministic channel. The larger Γ is, the more the channel is deterministic.
In conclusion, humans have invented codes whose sequences of symbols that make words cannot vary very much when indicating single physical or mental objects of their experience. On the contrary, to communicate concepts, large variability is achieved by introducing interpunctions to make word intervals and word intervals to make sentences, the final depositary of human basic concepts. Future work should be devoted to non-alphabetical languages. Finally, notice that the theory can inspire new research lines in cognitive science research.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The author thanks Lucia Matricciani for drawing Figure 1, Figure 2, Figure 3, Figure 4 and Figure 5.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A

Table A1. List of mathematical symbols.
Table A1. List of mathematical symbols.
SymbolDefinition
m Slope of regression line
m k j Slope in cross channel
m k j Correlation coefficient in cross channel
n C Number of characters per chapter
n W Number of words per chapter
n S Number of sentences per chapter
n I Number of interpunctions per chapter
r Correlation coefficient of linear variables
r 2 Coefficient of determination
s Standard deviation
s 2 Variance
C P Characters per word
  I p Word interval
M F   Word intervals per sentence
N m   Regression noise power
N r   Correlation noise power
R m   Regression noise-to-signal power ratio
R r   Correlation noise-to-signal power ratio
P F Words per sentence
γ Signal-to-noise ratio (linear)
Γ Signal-to-noise ratio (dB)
< > Mean value

Appendix B. Scatterplots in Different Languages

Figure A1. Scatterplots for the French translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure A1. Scatterplots for the French translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g0a1
Figure A2. Scatterplots for the Italian translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure A2. Scatterplots for the Italian translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g0a2
Figure A3. Scatterplots for the Portuguese translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure A3. Scatterplots for the Portuguese translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g0a3
Figure A4. Scatterplots for the Spanish translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure A4. Scatterplots for the Spanish translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g0a4
Figure A5. Scatterplots for the German translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure A5. Scatterplots for the German translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g0a5
Figure A6. Scatterplots for the Russian translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure A6. Scatterplots for the Russian translation between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g0a6

Appendix C

Let m k and s k be the (conditional) mean value and standard deviation of samples belonging to set k -th out of N sets of the ensemble, e.g., the values shown in Figure 18. From statistical theory [46,47], the unconditional mean (ensemble mean) m is given by the mean of means:
m = 1 N k = 1 N m k .
The unconditional variance (ensemble variance) s 2 ( s is the unconditional standard deviation) is given by
s 2 = v a r m k + 1 N k = 1 N s k 2 ,
v a r m k = 1 N k = 1 N m k 2 m 2 .
From Equations (A1)–(A3), we obtain the overall values reported in Table 4.

References

  1. Deniz, F.; Nunez–Elizalde, A.O.; Huth, A.G.; Gallant Jack, L. The Representation of Semantic Information Across Human Cerebral Cortex During Listening Versus Reading Is Invariant to Stimulus Modality. J. Neurosci. 2019, 39, 7722–7736. [Google Scholar] [CrossRef] [PubMed]
  2. Matricciani, E. A Mathematical Structure Underlying Sentences and Its Connection with Short–Term Memory. AppliedMath 2024, 4, 120–142. [Google Scholar] [CrossRef]
  3. Matricciani, E. Is Short–Term Memory Made of Two Processing Units? Clues from Italian and English Literatures down Several Centuries. Information 2024, 15, 6. [Google Scholar] [CrossRef]
  4. Miller, G.A. The Magical Number Seven, Plus or Minus Two. Some Limits on Our Capacity for Processing Information. Psychol. Rev. 1956, 63, 343–352. [Google Scholar] [CrossRef]
  5. Crowder, R.G. Short–term memory: Where do we stand? Mem. Cogn. 1993, 21, 142–145. [Google Scholar] [CrossRef]
  6. Lisman, J.E.; Idiart, M.A.P. Storage of 7 ± 2 Short–Term Memories in Oscillatory Subcycles. Science 1995, 267, 1512–1515. [Google Scholar] [CrossRef]
  7. Cowan, N. The magical number 4 in short–term memory: A reconsideration of mental storage capacity. Behav. Brain Sci. 2000, 24, 87–114. [Google Scholar] [CrossRef]
  8. Bachelder, B.L. The Magical Number 7 ± 2: Span Theory on Capacity Limitations. Behav. Brain Sci. 2001, 24, 116–117. [Google Scholar] [CrossRef]
  9. Saaty, T.L.; Ozdemir, M.S. Why the Magic Number Seven Plus or Minus Two. Math. Comput. Model. 2003, 38, 233–244. [Google Scholar] [CrossRef]
  10. Burgess, N.; Hitch, G.J. A revised model of short–term memory and long–term learning of verbal sequences. J. Mem. Lang. 2006, 55, 627–652. [Google Scholar] [CrossRef]
  11. Richardson, J.T.E. Measures of short–term memory: A historical review. Cortex 2007, 43, 635–650. [Google Scholar] [CrossRef] [PubMed]
  12. Mathy, F.; Feldman, J. What’s magic about magic numbers? Chunking and data compression in short–term memory. Cognition 2012, 122, 346–362. [Google Scholar] [CrossRef] [PubMed]
  13. Gignac, G.E. The Magical Numbers 7 and 4 Are Resistant to the Flynn Effect: No Evidence for Increases in Forward or Backward Recall across 85 Years of Data. Intelligence 2015, 48, 85–95. [Google Scholar] [CrossRef]
  14. Trauzettel-Klosinski, S.; Dietz, K. Standardized Assessment of Reading Performance: The New International Reading Speed Texts IreST. IOVS 2012, 53, 5452–5461. [Google Scholar] [CrossRef]
  15. Melton, A.W. Implications of Short–Term Memory for a General Theory of Memory. J. Verbal Learn. Verbal Behav. 1963, 2, 1–21. [Google Scholar] [CrossRef]
  16. Atkinson, R.C.; Shiffrin, R.M. The Control of Short–Term Memory. Sci. Am. 1971, 225, 82–91. [Google Scholar] [CrossRef]
  17. Murdock, B.B. Short–Term Memory. Psychol. Learn. Motiv. 1972, 5, 67–127. [Google Scholar]
  18. Baddeley, A.D.; Thomson, N.; Buchanan, M. Word Length and the Structure of Short–Term Memory. J. Verbal Learn. Verbal Behav. 1975, 14, 575–589. [Google Scholar] [CrossRef]
  19. Case, R.; Midian Kurland, D.; Goldberg, J. Operational efficiency and the growth of short–term memory span. J. Exp. Child Psychol. 1982, 33, 386–404. [Google Scholar] [CrossRef]
  20. Grondin, S. A temporal account of the limited processing capacity. Behav. Brain Sci. 2000, 24, 122–123. [Google Scholar] [CrossRef]
  21. Pothos, E.M.; Joula, P. Linguistic structure and short–term memory. Behav. Brain Sci. 2000, 24, 138–139. [Google Scholar] [CrossRef]
  22. Conway, A.R.A.; Cowan, N.; Michael, F.; Bunting, M.F.; Therriaulta, D.J.; Minkoff, S.R.B. A latent variable analysis of working memory capacity, short–term memory capacity, processing speed, and general fluid intelligence. Intelligence 2002, 30, 163–183. [Google Scholar] [CrossRef]
  23. Jonides, J.; Lewis, R.L.; Nee, D.E.; Lustig, C.A.; Berman, M.G.; Moore, K.S. The Mind and Brain of Short–Term Memory. Annu. Rev. Psychol. 2008, 69, 193–224. [Google Scholar] [CrossRef] [PubMed]
  24. Barrouillest, P.; Camos, V. As Time Goes By: Temporal Constraints in Working Memory. Curr. Dir. Psychol. Sci. 2012, 21, 413–419. [Google Scholar]
  25. Potter, M.C. Conceptual short–term memory in perception and thought. Front. Psychol. 2012, 3, 113. [Google Scholar] [CrossRef]
  26. Jones, G.; Macken, B. Questioning short–term memory and its measurements: Why digit span measures long–term associative learning. Cognition 2015, 144, 1–13. [Google Scholar] [CrossRef]
  27. Chekaf, M.; Cowan, N.; Mathy, F. Chunk formation in immediate memory and how it relates to data compression. Cognition 2016, 155, 96–107. [Google Scholar] [CrossRef]
  28. Norris, D. Short–Term Memory and Long–Term Memory Are Still Different. Psychol. Bull. 2017, 143, 992–1009. [Google Scholar] [CrossRef]
  29. Houdt, G.V.; Mosquera, C.; Napoles, G. A review on the long short–term memory model. Artif. Intell. Rev. 2020, 53, 5929–5955. [Google Scholar] [CrossRef]
  30. Islam, M.; Sarkar, A.; Hossain, M.; Ahmed, M.; Ferdous, A. Prediction of Attention and Short–Term Memory Loss by EEG Workload Estimation. J. Biosci. Med. 2023, 11, 304–318. [Google Scholar] [CrossRef]
  31. Rosenzweig, M.R.; Bennett, E.L.; Colombo, P.J.; Lee, P.D.W. Short–term, intermediate–term and Long–term memories. Behav. Brain Res. 1993, 57, 193–198. [Google Scholar] [CrossRef]
  32. Kaminski, J. Intermediate–Term Memory as a Bridge between Working and Long–Term Memory. J. Neurosci. 2017, 37, 5045–5047. [Google Scholar] [CrossRef] [PubMed]
  33. Matricciani, E. Equivalent Processors Modelling the Short–Term Memory. Preprints 2025. [Google Scholar] [CrossRef]
  34. Matricciani, E. Deep Language Statistics of Italian throughout Seven Centuries of Literature and Empirical Connections with Miller’s 7 ∓ 2 Law and Short–Term Memory. Open J. Stat. 2019, 9, 373–406. [Google Scholar] [CrossRef]
  35. Matricciani, E. A Statistical Theory of Language Translation Based on Communication Theory. Open J. Stat. 2020, 10, 936–997. [Google Scholar] [CrossRef]
  36. Matricciani, E. Multiple Communication Channels in Literary Texts. Open J. Stat. 2022, 12, 486–520. [Google Scholar] [CrossRef]
  37. Strinati, E.C.; Barbarossa, S. 6G Networks: Beyond Shannon Towards Semantic and Goal–Oriented Communications. Comput. Netw. 2021, 190, 107930. [Google Scholar] [CrossRef]
  38. Shi, G.; Xiao, Y.; Li, Y.; Xie, X. From semantic communication to semantic–aware networking: Model, architecture, and open problems. IEEE Commun. Mag. 2021, 59, 44–50. [Google Scholar] [CrossRef]
  39. Xie, H.; Qin, Z.; Li, G.Y.; Juang, B.H. Deep learning enabled semantic communication systems. IEEE Trans. Signal Process. 2021, 69, 2663–2675. [Google Scholar] [CrossRef]
  40. Luo, X.; Chen, H.H.; Guo, Q. Semantic communications: Overview, open issues, and future research directions. IEEE Wirel. Commun. 2022, 29, 210–219. [Google Scholar] [CrossRef]
  41. Wanting, Y.; Hongyang, D.; Liew, Z.Q.; Lim, W.Y.B.; Xiong, Z.; Niyato, D.; Chi, X.; Shen, X.; Miao, C. Semantic Communications for Future Internet: Fundamentals, Applications, and Challenges. IEEE Commun. Surv. Tutor. 2023, 25, 213–250. [Google Scholar]
  42. Bao, J. Towards a theory of semantic communication. In Proceedings of the Network Science Workshop, West Point, NY, USA, 22–24 June 2011; pp. 110–117. [Google Scholar]
  43. Bellegarda, J.R. Exploiting Latent Semantic Information in Statistical Language Modeling. Proc. IEEE 2000, 88, 1279–1296. [Google Scholar] [CrossRef]
  44. D’Alfonso, S. On Quantifying Semantic Information. Information 2011, 2, 61–101. [Google Scholar] [CrossRef]
  45. Zhong, Y. A Theory of Semantic Information. China Commun. 2017, 14, 1–17. [Google Scholar] [CrossRef]
  46. Papoulis Papoulis, A. Probability & Statistics; Prentice Hall: Hoboken, NJ, USA, 1990. [Google Scholar]
  47. Matricciani, E. Domestication of Source Text in Literary Translation Prevails over Foreignization. Analytics 2025, 4, 17. [Google Scholar] [CrossRef]
Figure 1. A flow chart of linguistic variables. The output variable of each block is connected to its input variable by a regression line.
Figure 1. A flow chart of linguistic variables. The output variable of each block is connected to its input variable by a regression line.
Information 16 00847 g001
Figure 2. Flow chart of linear systems. Upper panel: deterministic channel with multiplicative bias m , Equation (1). Lower panel: noisy deterministic channel with multiplicative bias and Gaussian noise source, Equation (3).
Figure 2. Flow chart of linear systems. Upper panel: deterministic channel with multiplicative bias m , Equation (1). Lower panel: noisy deterministic channel with multiplicative bias and Gaussian noise source, Equation (3).
Information 16 00847 g002
Figure 3. A flow chart of variances: r 2 s y 2 is the output variance of the values lying on the regression line, Equation (7); 1 r 2 s y 2 is the output variance due to the values of y not lying on the regression line, Equation (6).
Figure 3. A flow chart of variances: r 2 s y 2 is the output variance of the values lying on the regression line, Equation (7); 1 r 2 s y 2 is the output variance due to the values of y not lying on the regression line, Equation (6).
Information 16 00847 g003
Figure 4. Flow chart describing cross channel.
Figure 4. Flow chart describing cross channel.
Information 16 00847 g004
Figure 5. Flow chart of noisy single channels connected in series.
Figure 5. Flow chart of noisy single channels connected in series.
Information 16 00847 g005
Figure 6. Scatterplots for the original Greek texts between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Figure 6. Scatterplots for the original Greek texts between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. The continuous black line is the regression line.
Information 16 00847 g006
Figure 7. Scatterplots for the English texts between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. In this case, English is the language to be translated. The continuous black line is the regression line.
Figure 7. Scatterplots for the English texts between (a) characters and words; (b) words and interpunctions; (c) interpunctions and sentences; and (d) between characters and sentences. In this case, English is the language to be translated. The continuous black line is the regression line.
Information 16 00847 g007
Figure 8. (a) Cumulative probability distribution that the abscissa is not exceeded, correlation coefficient r ; (b) cumulative probability distribution that the abscissa is not exceeded, coefficient of determination r 2 . Both refer to the following scatterplots: words versus characters, green; interpunctions versus words: cyan; sentences versus interpunctions, magenta. The black line refers to the scatterplot sentences versus characters; the red line refers to the series channel considered in Section 7.
Figure 8. (a) Cumulative probability distribution that the abscissa is not exceeded, correlation coefficient r ; (b) cumulative probability distribution that the abscissa is not exceeded, coefficient of determination r 2 . Both refer to the following scatterplots: words versus characters, green; interpunctions versus words: cyan; sentences versus interpunctions, magenta. The black line refers to the scatterplot sentences versus characters; the red line refers to the series channel considered in Section 7.
Information 16 00847 g008
Figure 9. Cumulative probability distribution that the abscissa is not exceeded for the regression line slope m in the following scatterplots: words versus characters, green; interpunctions versus words: cyan; sentences versus interpunctions, magenta. The black line (not visible because superposed by the red cline) refers to the scatterplot sentences versus characters; the red line refers to the series channel considered in Section 7.
Figure 9. Cumulative probability distribution that the abscissa is not exceeded for the regression line slope m in the following scatterplots: words versus characters, green; interpunctions versus words: cyan; sentences versus interpunctions, magenta. The black line (not visible because superposed by the red cline) refers to the scatterplot sentences versus characters; the red line refers to the series channel considered in Section 7.
Information 16 00847 g009
Figure 10. (a) Signal-to-noise ratio SNR Γ (dB) versus language (see order number in Table 1); (b) theoretical relationship between Γ and coefficient of determination. Characters-to-words, green; words-to-interpunctions, cyan; interpunctions-to-sentences, magenta. Horizontal lines in (a) draw mean values.
Figure 10. (a) Signal-to-noise ratio SNR Γ (dB) versus language (see order number in Table 1); (b) theoretical relationship between Γ and coefficient of determination. Characters-to-words, green; words-to-interpunctions, cyan; interpunctions-to-sentences, magenta. Horizontal lines in (a) draw mean values.
Information 16 00847 g010
Figure 11. Histograms (37 samples) of the signal-to-noise ratio (SNR Γ ) for each channel: (a) characters-to-interpunctions; (b) words-to-interpunctions; (c) interpunctions-to-sentences.
Figure 11. Histograms (37 samples) of the signal-to-noise ratio (SNR Γ ) for each channel: (a) characters-to-interpunctions; (b) words-to-interpunctions; (c) interpunctions-to-sentences.
Information 16 00847 g011
Figure 12. Cumulative probability distribution that the abscissa is not exceeded, for the signal-to-noise ratio, SNR Γ , in the following channels: characters-to-words, green; words-to-interpunctions, cyan; interpunctions-to-sentences, magenta. The black line refers to the channel characters-to-sentences estimated from the scatterplot of Figure 6d; the red line refers to the series channel considered in Section 7.
Figure 12. Cumulative probability distribution that the abscissa is not exceeded, for the signal-to-noise ratio, SNR Γ , in the following channels: characters-to-words, green; words-to-interpunctions, cyan; interpunctions-to-sentences, magenta. The black line refers to the channel characters-to-sentences estimated from the scatterplot of Figure 6d; the red line refers to the series channel considered in Section 7.
Information 16 00847 g012
Figure 13. (a) Single-channel NSR and series-channel NSR in linear units; (b) the signal-to-noise ratio SNR Γ (dB). The horizontal lines draw mean values. Channels: characters-to-words, green; words-to-interpunctions, cyan; interpunctions-to-sentences, magenta; series channel, red.
Figure 13. (a) Single-channel NSR and series-channel NSR in linear units; (b) the signal-to-noise ratio SNR Γ (dB). The horizontal lines draw mean values. Channels: characters-to-words, green; words-to-interpunctions, cyan; interpunctions-to-sentences, magenta; series channel, red.
Information 16 00847 g013
Figure 14. (a) A scatterplot between the slope calculated from the scatterplot between characters and sentences (Table 2) and the slope given by Equation (31); (b) a scatterplot between the correlation coefficient calculated from the scatterplot between characters and sentences (Table 2) and that calculated by solving Equation (29) for r . The continuous black line is the regression line.
Figure 14. (a) A scatterplot between the slope calculated from the scatterplot between characters and sentences (Table 2) and the slope given by Equation (31); (b) a scatterplot between the correlation coefficient calculated from the scatterplot between characters and sentences (Table 2) and that calculated by solving Equation (29) for r . The continuous black line is the regression line.
Information 16 00847 g014
Figure 15. The mean value (upper panel) and correlation coefficient (lower panel) in the indicated languages, assuming Greek as the reference language (to be translated) in the channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences.
Figure 15. The mean value (upper panel) and correlation coefficient (lower panel) in the indicated languages, assuming Greek as the reference language (to be translated) in the channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences.
Information 16 00847 g015
Figure 16. The mean value (upper panel) and correlation coefficient (lower panel) in the indicated languages, assuming English as the reference language (to be translated) in the channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences.
Figure 16. The mean value (upper panel) and correlation coefficient (lower panel) in the indicated languages, assuming English as the reference language (to be translated) in the channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences.
Information 16 00847 g016
Figure 17. A scatterplot between X = R m and Y = R r in the indicated channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences. Red circles indicate the coordinates < R m > , < R r > of the barycenter.
Figure 17. A scatterplot between X = R m and Y = R r in the indicated channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences. Red circles indicate the coordinates < R m > , < R r > of the barycenter.
Information 16 00847 g017
Figure 18. The mean value (upper panel) and standard deviation lower of SNR Γ (dB) in the indicated language (see Table 1) in the indicated channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences. Black lines indicate the overall means. The mean and standard deviations are calculated from the mean of variances.
Figure 18. The mean value (upper panel) and standard deviation lower of SNR Γ (dB) in the indicated language (see Table 1) in the indicated channels: (a) words-to-words; (b) interpunctions-to-interpunctions; (c) sentences-to-sentences. Black lines indicate the overall means. The mean and standard deviations are calculated from the mean of variances.
Information 16 00847 g018
Figure 19. A histogram of the signal-to-noise ratio SNR Γ (dB) in the words-to-words channel ( 37 × 37 37 = 1332 samples), shown with blue circles. The continuous black line models the histogram with a Gaussian density function.
Figure 19. A histogram of the signal-to-noise ratio SNR Γ (dB) in the words-to-words channel ( 37 × 37 37 = 1332 samples), shown with blue circles. The continuous black line models the histogram with a Gaussian density function.
Information 16 00847 g019
Figure 20. “Universal” Gaussian probability density function (upper panel) and probability distribution function (that the abscissa is not exceeded) of Γ (dB) in the following channels: words-to-words, black; interpunctions-to-interpunctions, blue; sentences-to-sentences, red. The horizontal black line indicates the mean value.
Figure 20. “Universal” Gaussian probability density function (upper panel) and probability distribution function (that the abscissa is not exceeded) of Γ (dB) in the following channels: words-to-words, black; interpunctions-to-interpunctions, blue; sentences-to-sentences, red. The horizontal black line indicates the mean value.
Information 16 00847 g020
Table 1. The language of translation and language family of the New Testament books (Matthew, Mark, Luke, John, Acts, Epistle to the Romans, Apocalypse), including the total number of characters ( C ) , words ( W ) , sentences ( S ), and interpunctions ( I ). The list concerning the genealogy of Jesus of Nazareth reported in Matthew 1.1–1.17 and in Luke 3.23–3.38 was deleted so as to not bias the statistics of linguistic variables [35]. The source of the texts considered is reported in reference [35].
Table 1. The language of translation and language family of the New Testament books (Matthew, Mark, Luke, John, Acts, Epistle to the Romans, Apocalypse), including the total number of characters ( C ) , words ( W ) , sentences ( S ), and interpunctions ( I ). The list concerning the genealogy of Jesus of Nazareth reported in Matthew 1.1–1.17 and in Luke 3.23–3.38 was deleted so as to not bias the statistics of linguistic variables [35]. The source of the texts considered is reported in reference [35].
LanguageOrderAbbreviationLanguage Family C W S I
Greek1GrHellenic486,520100,145475913,698
Latin2LtItalic467,02590,799537018,380
Esperanto3EsConstructed492,603111,259548322,552
French4FrRomance557,764133,050725817,904
Italian5ItRomance505,535112,943639618,284
Portuguese6PtRomance486,005109,468708020,105
Romanian7RmRomance513,876118,744702118,587
Spanish8SpRomance505,610117,537651818,410
Danish9DnGermanic541,675131,021876222,196
English10EnGermanic519,043122,641659016,666
Finnish11FnGermanic563,65095,879589319,725
German12GeGermanic547,982117,269706920,233
Icelandic13IcGermanic472,441109,170719319,577
Norwegian14NrGermanic572,863140,844930218,370
Swedish15SwGermanic501,352118,833766815,139
Bulgarian16BgBalto-Slavic490,381111,444772720,093
Czech17CzBalto-Slavic416,44792,533751419,465
Croatian18CrBalto-Slavic425,90597,336675017,698
Polish19PlBalto-Slavic506,66399,592818121,560
Russian20RsBalto-Slavic431,91392,736559422,083
Serbian21SrBalto-Slavic441,998104,585753218,251
Slovak22SlBalto-Slavic465,280100,151802319,690
Ukrainian23UkBalto-Slavic488,845107,047804322,761
Estonian24EtUralic495,382101,657631019,029
Hungarian25HnUralic508,77695,837597122,970
Albanian26AlAlbanian502,514 123,625580719,352
Armenian27ArArmenian472,196100,604659518,086
Welsh28WlCeltic527,008130,698567622,585
Basque29BsIsolate588,76294,898559119,312
Hebrew30HbSemitic372,03188,478759715,806
Cebuano31CbAustronesian681,407146,481922116,788
Tagalog32TgAustronesian618,714128,209794416,405
Chichewa33ChNiger–Congo575,45494,817756015,817
Luganda34LgNiger–Congo570,73891,819707316,401
Somali35SmAfro-Asiatic584,135109,686612717,765
Haitian36HtFrench Creole514,579152,82310,42923,813
Nahuatl37NhUto-Aztecan816,108121,600926319,271
Table 2. Slope m and correlation coefficient r of indicated regression lines in each language/translation.
Table 2. Slope m and correlation coefficient r of indicated regression lines in each language/translation.
LanguageWords vs. CharactersInterpunctions vs. WordsSentences vs. InterpunctionsSentences vs. Characters.
m 1 r 1 m 2 r 2 m 3 r 3 m r
Greek0.20540.98930.13690.92980.35410.93820.00990.8733
Latin0.19440.98900.20380.95150.29570.93660.01170.8646
Esperanto0.22560.99200.20450.96680.24610.95450.01130.8998
French0.23860.99450.13470.94830.40450.95090.01310.9339
Italian0.22330.99210.16360.94760.34890.95370.01270.8856
Portuguese0.22460.99240.18450.96200.35320.94840.01460.9106
Romanian0.23120.99330.15680.95890.38230.93840.01380.8820
Spanish0.23200.99190.15800.96190.35650.95810.01300.9047
Danish0.24170.99450.16940.95740.39610.95510.01630.9257
English0.23640.99250.13650.95090.39620.94830.01280.8916
Finnish0.17020.99040.20670.96210.30290.94640.01070.9131
German0.21420.99380.17310.96370.35110.95550.01300.9325
Icelandic0.23150.99370.18050.96000.36720.95270.01540.9296
Norwegian0.24600.99560.13050.95810.50180.96210.01620.9626
Swedish0.23710.99180.12770.92180.50410.94990.01540.9423
Bulgarian0.22710.99260.18090.95900.38610.94820.01590.9203
Czech0.22230.99270.21250.94960.38790.92820.01840.9034
Croatian0.22870.99150.18250.95040.38530.96050.01610.9095
Polish0.19680.99390.21590.96500.37680.92450.01600.9049
Russian0.21480.98890.23970.97120.25660.92740.01320.8728
Serbian0.23700.99250.17450.95130.41540.94360.01720.9111
Slovak0.21490.99110.19730.95320.40850.95440.01730.9092
Ukrainian0.21810.98930.21220.97300.35560.94480.01660.9545
Estonian0.20540.99120.18810.95590.33420.94670.01290.8995
Hungarian0.18820.98850.24120.97190.26320.94820.01200.9282
Albanian0.24580.98960.15730.96070.30400.95820.01170.9106
Armenian0.21400.97530.18020.96990.36980.96350.01420.8868
Welsh0.24820.99530.17340.98180.25430.94930.01090.9336
Basque0.16140.99390.20450.96730.29250.95060.00970.9210
Hebrew0.23800.99450.17840.96150.48690.96350.02060.9144
Cebuano0.21490.99830.11450.94650.54910.95780.01360.9670
Tagalog0.20720.99570.12810.95550.48790.93630.01300.9411
Chichewa0.16490.99640.16850.94200.47330.95960.01320.9381
Luganda0.16100.99510.17970.94880.43140.95010.01250.9235
Somali0.18760.99650.16280.93000.35050.93990.01070.8773
Haitian0.29720.99590.15710.96720.43380.95670.02030.9288
Nahuatl0.14890.99550.15930.93040.47590.95820.01140.9435
Overall 0.2161
± 0.0296
0.9925
± 0.0038
0.1750
± 0.0308
0.9558
± 0.0131
0.3795
± 0.0755
0.9492
± 0.0100
0.0140
± 0.0027
0.9149
± 0.0252
Table 3. The mean and standard deviation of the signal-to-noise ratio Γ (dB) in the indicated channel. The probability density function of each channel is modeled as Gaussian.
Table 3. The mean and standard deviation of the signal-to-noise ratio Γ (dB) in the indicated channel. The probability density function of each channel is modeled as Gaussian.
Channel Mean   ±   Standard   Deviation   of   Γ (dB)
Characters-to-Words 18.60 ± 2.00
Words-to-Interpunctions 10.42 ± 1.38
Interpunctions-to-Sentences 9.66 ± 0.89
Table 4. The mean and standard deviation of the signal-to-noise ratio Γ (dB) in the indicated cross channels. The probability density function of each channel is modeled as Gaussian.
Table 4. The mean and standard deviation of the signal-to-noise ratio Γ (dB) in the indicated cross channels. The probability density function of each channel is modeled as Gaussian.
Channel Mean   Γ (dB)Standard Deviation (dB)
Words-to-Words18.939.21
Interpunctions-to-Interpunctions15.607.99
Sentences-to-Sentences14.948.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Matricciani, E. Revealing Short-Term Memory Communication Channels Embedded in Alphabetical Texts: Theory and Experiments. Information 2025, 16, 847. https://doi.org/10.3390/info16100847

AMA Style

Matricciani E. Revealing Short-Term Memory Communication Channels Embedded in Alphabetical Texts: Theory and Experiments. Information. 2025; 16(10):847. https://doi.org/10.3390/info16100847

Chicago/Turabian Style

Matricciani, Emilio. 2025. "Revealing Short-Term Memory Communication Channels Embedded in Alphabetical Texts: Theory and Experiments" Information 16, no. 10: 847. https://doi.org/10.3390/info16100847

APA Style

Matricciani, E. (2025). Revealing Short-Term Memory Communication Channels Embedded in Alphabetical Texts: Theory and Experiments. Information, 16(10), 847. https://doi.org/10.3390/info16100847

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop