Abstract
The purpose of the present paper is to further investigate the mathematical structure of sentences—proposed in a recent paper—and its connections with human short–term memory. This structure is defined by two independent variables which apparently engage two short–term memory buffers in a series. The first buffer is modelled according to the number of words between two consecutive interpunctions—variable referred to as the word interval, —which follows Miller’s law; the second buffer is modelled by the number of word intervals contained in a sentence, , ranging approximately for one to seven. These values result from studying a large number of literary texts belonging to ancient and modern alphabetical languages. After studying the numerical patterns (combinations of and ) that determine the number of sentences that theoretically can be recorded in the two memory buffers—which increases with the use of and —we compare the theoretical results with those that are actually found in novels from Italian and English literature. We have found that most writers, in both languages, write for readers with small memory buffers and, consequently, are forced to reuse sentence patterns to convey multiple meanings.
1. Does the Short–Term Memory Process Words with Two Independent Buffers in Series?
Recently [1], we proposed a well–grounded conjecture that a sentence—read or pronounced as the two activities are similarly processed by the brain [2]—is elaborated by the short–term memory (STM), with two independent processing units in series that have similar buffer size. The clues for conjecturing this model have emerged from considering many novels belonging to Italian and English literature. In [1], we have shown that there are no significant mathematical/statistical differences between the two literary corpora, according to surface deep–language variables. In other words, the mathematical surface structure of alphabetical languages—a creation of the human mind—seems to be deeply rooted in humans, independent of the particular language used.
A two–unit STM processing can be justified according to how a human mind seems to memorize “chunks” of information written in a sentence. Although simple and related to the surface of language, the model seems to describe mathematically the input–output characteristics of a complex mental process largely unknown.
According to [1], the first processing unit is linked to the number of words between two contiguous interpunctions, the variable for which is indicated by —termed the word interval (Appendix A lists the mathematical symbols used in the present paper)—approximately ranging within Miller’s law range [3,4,5,6,7,8,9,10,11,12]. The second unit is linked to the number of s contained in a sentence, referred to as the extended STM, or E–STM, ranging approximately from one to six. We have shown that the capacity (expressed in words) required to process a sentence ranges from to words, values that can be converted into time by assuming a reading speed. This conversion gives the range s for a fast reader [13], and s for an average reader of novels, values that are well-supported by the experiments reported in the literature [14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29].
The E–STM must not be confused with the intermediate memory [30,31]. It is not modelled by studying neuronal activity, but by studying the surface aspects of human communication, such as words and interpunctions, whose effects writers and readers have experienced since the invention of writing.
The modeling of the STM processing by two units in a series has never been considered in the literature before [1,32]. The reader is very likely aware that the literature on the STM and its various aspects is very large and multidisciplinary, but nobody—as far as we know—has never considered the connections we have found and discussed in [1,32]. Moreover, a sentence conveys meaning; therefore, the theory we are further developing in the present paper might be a starting point to arrive at the information theory that includes meaning.
Currently, some attempts are being made by many scholars to arrive at a “semantic communication” theory or a “semantic information” theory, but the results are still, in our opinion, in their infancies [33,34,35,36,37,38,39,40,41]. These theories, as those concerning the STM, have not considered the main “ingredients” of our theory—namely and —as a starting point for including meaning, which is still a very open issue.
Figure 1 sketches the flowchart of the two processing units [1]. The words , ,… are stored in the first buffer up to items—approximately in Miller’s range—until an interpunction is introduced to fix the length of . The word interval is then stored in the second buffer up to items, from about one to six, until the sentence ends. The process is then repeated for the next sentence.
Figure 1.
Flowchart of the two processing units of a sentence. The words , ,… are stored in the first buffer up to items to complete a word interval , which is approximately in Miller’s range, when an interpunction is introduced. is then stored in the E–STM buffer, up to items, i.e., in cells, approximately one to six, until the sentence ends.
The purpose of the present paper is to further investigate the mathematical structure underlying sentences, both theoretically and experimentally, by considering the novels previously mentioned [1] listed in Table A1 for Italian literature and in Table A2 for English literature.
After this introduction, in Section 2, we study the probability distribution function (PDF) of sentence size—measured in words—that is recordable by an E–STM buffer made of cells (this parameter plays the role of ). In other words, in this section, we study and discuss the length of sentences that humans can possibly conceive with an E–STM made of memory cells.
In Section 3, we study the number of sentences, with the same number of words, that cells can process. In this section, we study and discuss the complementary issue of Section 2, namely, how many sentences with a constant number of words humans can conceive, based solely on the E–STM of cells.
In Section 4, we compare the number of sentences that authors of Italian and English literature actually wrote for their novels to the number of sentences theoretically available to them, by defining a multiplicity factor. In Section 5, we define a mismatch index, which synthetically measures to what extent a writer uses the number of sentences that are theoretically available. In Section 6, we show that the parameters studied increase with the year of novel publication. Finally, in Section 7, we summarize the main results and propose future work.
2. Probability Distribution of Sentence Length versus E–STM Buffer Size
First, we study the conditional PDF of sentence length, measured in words —i.e., the parameter which in long texts, such as chapters, gives of each chapter—recordable in an E–STM buffer made of cells, i.e., the parameter which gives in chapters. Second, we study the overlap of the PDFs because this overlap gives interesting indications.
2.1. Probability Distribution of Sentence Length
To estimate the PDF of sentence length, we run a Monte Carlo simulation based on the PDF of obtained in [1] by merging the two literatures listed in Section 1.
In [1], we have shown that the PDF of , and —as previously mentioned, these averages refer to single chapters of the novels—can be modelled with a three-parameter log–normal density function [42] (natural logs):
In Equation (1), and are, respectively, the mean value and the standard deviation the log–normal PDF. Table 1 reports these values for the three deep-language variables.
Table 1.
Mean value and standard deviation of the log–normal PDF of the indicated variable [1].
The Monte Carlo simulation steps are as follows:
- Consider a buffer made of cells. The sentence contains word intervals: for example, if , the sentence contains two interpunctions followed by a full stop, a question mark, or an exclamation mark.
- Generate independent values of according to the log–normal model given by Equation (1) and Table 1. The independence of from a cell to another cell is reasonable [1]. In detail, from a random number generator of standard Gaussian density variables (zero mean and unit standard deviation), we get the relationship ; therefore, the three-parameter log–normal variable is then given by .
- Add the number of words contained in the cells to obtain :
- Repeat steps one through three many times (we repeated these steps 100,000 times, i.e., we simulated 100,000 sentences of different length) to obtain a stable conditional PDF of .
- Repeat steps one through four for another and obtain another PDF.
Figure 2 shows the conditional PDF for several values of . Each PDF can be very well-modelled by a Gaussian PDF because the probability of getting unacceptable negative values is negligible in any of the PDFs shown in Figure 2. For example, for , the mean value and the standard deviation are, respectively, words and words.
Figure 2.
Conditional PDFs of words per sentence versus an E–STM buffer of cells from two to eight. Each PDF can be modelled with a Gaussian PDF with a mean value proportional to and a standard deviation proportional to .
In general terms [42], the mean value of Equation (2) is given by:
Therefore, is proportional to . As for the standard deviation of , if the ’s are independent—as we assume—then the variance of is given by:
Therefore, the standard deviation is proportional to . Finally, according to the central limit theory [42], the PDF can be modelled as Gaussian in a significant range about the mean.
In conclusion, the Monte Carlo simulation produces a Gaussian PDF with a mean value proportional to and a standard deviation proportional to . These findings are clearly evident in the PDFs shown in Figure 2, in which and increase as theoretically expected; therefore, the mean values and standard deviations of the other PDFs can be calculated by scaling the values found for . For example, for , words and words.
Figure 3 shows the histograms corresponding to Figure 2. The number of samples for each conditional PDF, out of 100,000 considered in the Monte Carlo simulation, is obtained by distributing the samples according to the PDF of given by Equation (1) and Table 1. The case gives the largest sample size.
Figure 3.
Conditional histograms of words per sentence versus an E–STM buffer of cells from two to eight, obtained from Figure 1, by simulating 100,000 sentences weighted with the PDF of .
The results shown above have an experimental basis because the relationship between , the average of words per sentence for the entire novel which is calculated by averaging the of single chapters via weighting single chapters with the fraction of novel total word, as discussed in [32], versus , the average of of a novel calculated as is linear, as Figure 4 shows by drawing versus concerning the Italian and English novels mentioned above.
Figure 4.
Scatterplot of versus of Italian novels (blue circles) and English novels (red circles).
2.2. Overlap of the Conditional Probability Distributions
Figure 2 shows that the conditional PDFs overlap; therefore, some sentences can be processed by buffers of diverse size, either larger or smaller. Let us define the probability of these overlaps.
Let be the intersection of two contiguous Gaussian PDFs, for example, and ; therefore, the probability that a sentence length can be found in the nearest lower Gaussian PDF (going from ) is given by [42]:
Similarly, the probability that a sentence length can be found in the nearest higher Gaussian PDF (going from ) is given by:
For example, the threshold value between and is words and , while .
Figure 5 draws these probabilities (%) versus (lower ). Because increases with , therefore . However, this is not only a mathematically obvious result, but it also meaningful because it indicates that: (a) a human mind can process sentences of lengths belonging to the contiguous lower or higher (the probability of going to more distant PDFs is negligible) and (b) the number of these sentences is larger in the case , which simply means that an E–STM buffer can process to a larger extent data matched to a smaller capacity buffer than data matched to a larger capacity buffer.
Figure 5.
Overlap probability (%) versus (lower ); and are given by Equations (5) and (6).
Finally, notice that each sentence conveys meaning—theoretically, any sequence of words might be meaningful, although this may not always be the case, but we do not know the proportion—therefore, the PDFs found above are also the PDFs associated with meaning. Moreover, the same numerical sequence of words can carry different meanings, according to the words used. Multiplicity of meaning, therefore, is “built in” in a sequence of words. We will further explore this issue in the next sections by considering the number of sentences that authors of Italian and English literature actually wrote.
So far, we have explored the processing of the words of a sentence by simulating sentences of diverse length that are conditioned to the E–STM buffer size. In the next section, we explore the complementary processing concerning the number of sentences that contain the same number of words.
3. Theoretical Number of Sentences Recordable in Cells
We study the number of sentences of words that an E–STM buffer, made of cells, can theoretically process. In summary, we ask the following question: how many sentences containing the same number of words (Equation (2)) can be theoretically written in cells?
Table 2 reports these numbers as a function of and . We calculated these data first by running a code and then by finding the mathematical recursive formula that generates them, given by the following:
Table 2.
Theoretical number of sentences (columns) recordable in an E–STM buffer made of cells with the same number of words (items) indicated in the first column.
For example, if words and , we read and ; therefore, .
Figure 6 draws the data reported in some lines of Table 2 for a quick overview. We see how fast the number of sentences changes with for constant . For example, if words, then ranges from 1 ( to 52,698 sentences (. Maxima are clearly visible for and words at and , respectively. Values become fantastically large for larger and , well beyond the ability and creativity of single writers, as we will show in Section 4.
Figure 6.
Number of sentences made of words versus an E–STM buffer capacity of .
Figure 7 draws the data reported in some columns of Table 2, i.e., the number of sentences versus , for fixed . In this case, it is useful to adopt an efficiency factor , which is defined as the ratio between and for a given :
Figure 7.
Number of sentences recordable in an E–STM buffer capacity of versus words per sentence.
This factor explains, summarily, how a buffer of cells is efficient in providing sentences with a given number of words, its units being sentences per word.
Figure 8 shows versus . It is interesting to note that for words, the buffer can be more efficient than the others. Beyond , the larger buffers become very efficient with very large .
Figure 8.
Efficiency , Equation (8), of an E–STM buffer of cells versus words per sentence .
If a writer uses short buffers—e.g., deliberately because of his/her style, or necessarily because of the reader’s E–STM memory size—then he/she has to repeat the same numerical sequence of words many times, according to the number of meanings conveyed. For example, if and the writer has only nine different choices, or patterns, of two numbers whose sum is 10 (Table 2). Therefore, Table 2 gives the minimum number of meanings that can be conveyed. The larger the , the larger is the variety of sentences that can be written with words.
The following question naturally arises: How many sentences authors do write in their texts as compared to the theoretical number available to them? In the next section, we will compare these two sets of data by studying the novels taken from the Italian and English literature listed in Appendix B, by assuming their average values and and by defining a multiplicity factor.
4. Experimental Multiplicity Factor of Sentences
We compare the number of sentences that authors of Italian and English literature actually wrote for each novel to the number of sentences theoretically available to them, according to the and of each novel. In this analysis, we do not consider the values of and of each chapter of a novel because the detail would be so fine as to miss the general trend given by the average values , of the complete novel.
As is well known, the average value and the standard deviation of integers very likely are not integers, as is always the case for the linguistic parameters; therefore, to apply the mathematical theory of the previous sections, we must do some interpolations and only at the end of the calculation consider the integers.
Let us compare the experimental number of sentences in a novel, as reported in Table A1 and Table A2, to the theoretical number available to the author, according to the experimental values (which plays the role of ) and (which plays the role of ) of the novel.
By referring to Figure 7, the interpolation between the integers of Table 2 to find the curve of constant —given by the real number —is linear along both axes. At the intersection of the vertical line (corresponding to the real number ) and the new curve (corresponding to the real number ), we find the theoretical by rounding the value to the nearest integer toward zero. For example, for David Copperfield, in Table A2 we read , and the interpolation gives . Figure 9 shows the result of this exercise. We see that increases rapidly with . The most displaced (red) circle is due to Robinson Crusoe.
Figure 9.
Theoretical number of sentences versus for Italian (blue circles) and English (red circles) novels. The most displaced (red) circle is due to Robinson Crusoe.
The comparison between and is performed by defining a multiplicity factor , defined as the ratio between (experimental value) and (theoretical value):
The values of for each novel are reported in Table A1 and Table A2. For example, for David Copperfield, . Figure 10 shows versus . We notice a fairly significant increasing trend of with .
Figure 10.
Multiplicity factor versus for Italian (blue circles) and English (red circles) novels.
Figure 11.
Multiplicity factor versus theoretical number of words for Italian (blue circles and blue line) and English (red circles and red line) novels.
The correlation coefficient of log values is for Italian and for English.
Based on Equations (10) and (11), when for Italian novels and for English novels; therefore, novels with sentences in the range use, on average, the number of sentences theoretically available for their averages and .
Figure 12.
Multiplicity factor versus for Italian (blue circles and blue line) and English (red circles and red line) novels.
For the Italian literature in question, (correlation coefficient of linear–log values is ) when ; for the English literature (correlation coefficient of linear–log values is ), when . Therefore, novels with sentences in the range use, on average, the same E–STM buffer size of cells.
From Figure 10, Figure 11 and Figure 12, we can draw the following conclusion: in general, is more likely than and often . When , the writer reuses the same pattern of number of words many times. The multiplicity factor, therefore, indicates also the minimum multiplicity of meaning conveyed by an E–STM besides, of course, the many diverse meanings conveyed by the same sequence of obtainable by only changing words. Few novels show . In these cases, the writer has enough diverse patterns to convey meaning, but most of them are not used.
Finally, it is interesting to relate to a universal readability factor , which is a function of both and [43].
The universal readability index, as compared to the current readability indices for the few languages for which they are available (mainly for English [43]), considers also the reader’s short-term memory processing capacity. It can be used to assess the readability of texts written in any alphabetical language, as described in [43].
Figure 13 shows versus . Because the readability of a text increases as increases, we can see that the novels with tend to be less readable than those with . The less-readable novels have, in general, large values of and therefore may contain more E–STM cells (large ).
Figure 13.
Multiplicity factor versus readability index for Italian (blue circles) and English (red circles) novels.
In conclusion, if a writer does use the full variety of sentence patterns available, or even overuses them, then he/she writes texts that are easier to read. On the other hand, if a writer does not use the full variety of sentence patterns available, then he/she tends to write texts that are more difficult to read. In the next section, we define a useful index, the mismatch index, which describes these cases.
5. Mismatch Index
We define a useful index, the mismatch index, which measures to what extent a writer uses the number of sentences that are theoretically available according to the averages and of the novel. For this purpose, we define the mismatch index:
According to Equation (14), when ; hence, and in this case the experiment and theory are perfectly matched. They are overmatched when ( and undermatched when (.
Figure 14 shows the scatterplot of versus . The mathematical models drawn are calculated by substituting Equations (12) and (13) in Equation (14). We can reiterate that when (overmatching, ), the writer repeats sentence patterns because there are not enough diverse patterns to convey all the meanings. The texts are easier to read. When (undermatching, ), the writer has theoretically many sentence patterns to choose from, but he/she uses only a few or very few of them. The texts are more difficult to read.
Figure 14.
Mismatch index versus for Italian (blue circles and blue line) and English (red circles and red line) novels.
Figure 15 shows the scatterplot of versus . The mathematical models drawn were calculated by substituting Equations (10) and (11) in Equation (14). Overmatching was found for for Italian and for English.
Figure 15.
Mismatch index versus the theoretical number of sentences for Italian (blue circles and blue line) and English (red circles and red line) novels.
Finally, Figure 16 shows versus , Equation (14), a picture that summarizes the entire analysis of mismatch.
Figure 16.
Mismatch index versus the multiplicity factor for Italian (blue circles) and English (red circles) novels.
6. Time Dependence
The novels considered in Table A1 and Table A2 were published in a period spanning several centuries. We show that the multiplicity factor and the mismatch index do depend on time.
Figure 17 shows the multiplicity factors versus the years of publication of the novels since 1800. It is evident that writers tend to use larger values of —therefore the E–STM buffers are of smaller sizes—as we approach the present epoch and a possible saturation at . The English literature shows a stable increasing pattern while the Italian literature seems to contain samples that come from two diverse sets of data, one of which evolved in agreement with English literature, the other (given by the novels labelled with “*” in Table A1) is always increasing with time but with a diverse slope.
Figure 17.
Multiplicity factor versus the year of novel publication for Italian (blue circles) and English (red circles) novels.
Figure 18 shows the mismatch index versus the year of novel publication.
Figure 18.
Mismatch factor versus the year of novel publication for Italian (blue circles) and English (red circles) novels.
Figure 19 shows the universal readability index versus time. In both Figure 18 and Figure 19, we can observe the same trends shown in Figure 17, which therefore reinforces the conjecture that: (a) the writers are partially changing their style with time by making their novels more readable, i.e., more matched to less-educated readers according to the relationship between and the schooling years in the Italian school system, as discussed in [43]; (b) a saturation seems to occur in all parameters in the novels written in the second half of the XX century, at least according to the novels of Appendix B.
Figure 19.
Universal readability index versus the year of novel publication for Italian (blue circles) and English (red circles) novels.
7. Summary and Future Work
In the present paper, we have further investigated the mathematical structure of sentences and its connections with human short–term memory. This structure is defined by two independent variables which apparently engage two short-term memory buffers in series. The first buffer is modelled according to the number of words between two consecutive interpunctions—variable-termed word interval —which follows Miller’s law; the second buffer is modelled by the number of word intervals contained in a sentence, , ranging approximately from one to seven. These values arise from an extensive analysis of alphabetical texts [44].
We have studied the numerical patterns (combinations of and ) that determine the number of sentences that theoretically can be recorded in the two memory buffers—which increases with and —and we have compared the theoretical results with those that are actually found in novels from Italian and English literature. We have found that most writers, in both languages, write for readers with small memory buffers and, consequently, are forced to reuse sentence patterns to convey multiple meanings. In this case, texts are easier to read, according to the universal readability index.
Future work should consider other literatures to confirm what, in our opinion, is general because the topic is connected to the human mind. The same analysis performed on ancient languages, such as Greek and Latin—for which there are large literary corpora—would show whether these ancient writers/readers displayed similar short–term memory buffers.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data are contained within the article.
Acknowledgments
The author wishes to thank the many scholars who, with great care and love, maintain digital texts to be available to readers and scholars of different academic disciplines, such as Perseus Digital Library and Project Gutenberg.
Conflicts of Interest
The author declares no conflicts of interest.
Appendix A. List of Mathematical Symbols
| Symbol | Definition |
| Cells of E–STM buffer | |
| Universal readability index | |
| Mismatch index | |
| Word interval | |
| Word intervals in a sentence, chapter average | |
| Word intervals in a sentence, novel average | |
| Words in a sentence, chapter average | |
| Words in a sentence, novel average | |
| Experimental sentences | |
| Theoretical sentences written in cells | |
| Words in a sentence | |
| Three–parameter log–normal density function | |
| Gaussian PDF | |
| Mean value of Gaussian PDF | |
| standard deviation of Gaussian PDF | |
| Multiplicity factor | |
| Efficiency factor | |
| Mean value of log–normal PDF | |
| standard deviation of log–normal PDF |
Appendix B. List of the Novels Considered from Italian and English Literature
Table A1 and Table A2 list the authors, the titles of the novels, and their years of publication in either Italian and English literature as considered in the paper, with deep–language average statistics, multiplicity factor , and mismatch index . The averages have been calculated by weighting each chapter value with its fraction of the total number of words in the novel, as described in [32].
Table A1.
Authors of novels of Italian literature. Number of total sentences (sentences ending with full stops, question marks, or exclamation marks); average number of characters per word, ; average number of words pr sentence, ; average number of word intervals, ; average number word intervals per sentence, ; multiplicity factor ; and mismatch index .
Table A1.
Authors of novels of Italian literature. Number of total sentences (sentences ending with full stops, question marks, or exclamation marks); average number of characters per word, ; average number of words pr sentence, ; average number of word intervals, ; average number word intervals per sentence, ; multiplicity factor ; and mismatch index .
| Author (Literary Work, Year) | Sentences | ||||||
|---|---|---|---|---|---|---|---|
| Anonymous (I Fioretti di San Francesco, 1476) | 1064 | 4.65 | 37.70 | 8.24 | 4.56 | 0.004 | −0.99 |
| Bembo Pietro (Prose, 1525) | 1925 | 4.37 | 37.91 | 6.42 | 5.92 | 0.001 | −1.00 |
| Boccaccio Giovanni (Decameron, 1353) | 6147 | 4.48 | 44.27 | 7.79 | 5.69 | 0.001 | −1.00 |
| Buzzati Dino (Il deserto dei tartari, 1940) | 3311 | 5.10 | 17.75 | 6.63 | 2.67 | 6.90 | 0.75 |
| Buzzati Dino (La boutique del mistero, 1968 *) | 4219 | 4.82 | 15.45 | 6.37 | 2.41 | 18.83 | 0.90 |
| Calvino (Il barone rampante, 1957 *) | 3864 | 4.63 | 19.87 | 6.73 | 2.91 | 4.37 | 0.63 |
| Calvino Italo (Marcovaldo, 1963 *) | 2000 | 4.74 | 17.60 | 6.59 | 2.67 | 4.28 | 0.62 |
| Cassola Carlo (La ragazza di Bube, 1960 *) | 5873 | 4.48 | 11.93 | 5.64 | 2.11 | 87.66 | 0.98 |
| Collodi Carlo (Pinocchio, 1883) | 2512 | 4.60 | 16.92 | 6.19 | 2.72 | 5.74 | 0.70 |
| Da Ponte Lorenzo (Vita, 1823) | 5459 | 4.71 | 26.15 | 6.91 | 3.78 | 0.51 | −0.32 |
| Deledda Grazia (Canne al vento, 1913, Nobel Prize 1926) | 4184 | 4.51 | 15.08 | 6.06 | 2.48 | 18.35 | 0.90 |
| D’Azeglio Massimo (Ettore Fieramosca, 1833) | 3182 | 4.64 | 29.77 | 7.36 | 4.03 | 0.12 | −0.78 |
| De Amicis Edmondo (Cuore, 1886) | 4775 | 4.55 | 19.43 | 5.61 | 3.41 | 2.48 | 0.42 |
| De Marchi Emilio (Demetrio Panelli, 1890) | 5363 | 4.70 | 18.95 | 7.06 | 2.68 | 8.95 | 0.80 |
| D’Annunzio Gabriele (Le novelle delle Pescara, 1902) | 3027 | 4.91 | 17.99 | 6.38 | 2.79 | 5.35 | 0.68 |
| Eco Umberto (Il nome della rosa, 1980 *) | 8490 | 4.81 | 21.08 | 7.46 | 2.81 | 8.70 | 0.79 |
| Fogazzaro (Il santo, 1905) | 6637 | 4.79 | 14.84 | 6.33 | 2.34 | 37.08 | 0.95 |
| Fogazzaro (Piccolo mondo antico, 1895) | 7069 | 4.79 | 16.08 | 6.10 | 2.64 | 20.98 | 0.91 |
| Gadda (Quer pasticciaccio brutto… 1957 *) | 5596 | 4.76 | 18.43 | 4.98 | 3.68 | 2.69 | 0.46 |
| Grossi Tommaso (Marco Visconti, 1834) | 5301 | 4.59 | 28.07 | 6.56 | 4.23 | 0.16 | −0.72 |
| Leopardi Giacomo (Operette morali, 1827) | 2694 | 4.70 | 31.78 | 6.90 | 4.54 | 0.03 | −0.95 |
| Levi Primo (Cristo si è fermato a Eboli, 1945 *) | 3611 | 4.73 | 22.94 | 5.70 | 4.02 | 0.47 | −0.36 |
| Machiavelli Niccolò (Il principe, 1532) | 702 | 4.71 | 40.17 | 6.45 | 6.23 | 0.0001 | −1.00 |
| Manzoni Alessandro (I promessi sposi, 1840) | 9766 | 4.60 | 24.83 | 5.30 | 4.63 | 0.33 | −0.51 |
| Manzoni Alessandro (Fermo e Lucia, 1821) | 7496 | 4.75 | 30.98 | 7.17 | 4.30 | 0.12 | −0.78 |
| Moravia Alberto (Gli indifferenti, 1929 *) | 2830 | 4.81 | 36.00 | 6.74 | 5.34 | 0.003 | −0.99 |
| Moravia Alberto (La ciociara, 1957 *) | 4271 | 4.56 | 29.93 | 7.28 | 4.12 | 0.12 | −0.78 |
| Pavese Cesare (La bella estate, 1940) | 2121 | 4.54 | 12.37 | 5.97 | 2.06 | 31.19 | 0.94 |
| Pavese Cesare (La luna e i falò, 1949 *) | 2544 | 4.47 | 17.83 | 6.83 | 2.60 | 5.64 | 0.70 |
| Pellico Silvio (Le mie prigioni, 1832) | 3148 | 4.80 | 17.27 | 6.50 | 2.69 | 7.00 | 0.75 |
| Pirandello Luigi (Il fu Mattia Pascal, 1904, Nobel Prize 1934) | 5284 | 4.63 | 14.57 | 4.94 | 2.93 | 16.72 | 0.89 |
| Sacchetti Franco (Trecentonovelle, 1392) | 8060 | 4.37 | 22.43 | 5.82 | 3.83 | 1.41 | 0.17 |
| Salernitano Masuccio (Il Novellino, 1525) | 1965 | 4.40 | 19.20 | 5.14 | 3.68 | 0.79 | −0.12 |
| Salgari Emilio (Il corsaro nero, 1899) | 6686 | 4.99 | 15.09 | 6.36 | 2.36 | 34.46 | 0.94 |
| Salgari Emilio (I minatori dell’Alaska, 1900) | 6094 | 5.01 | 15.24 | 6.25 | 2.44 | 27.21 | 0.93 |
| Svevo Italo (Senilità, 1898) | 4236 | 4.86 | 16.04 | 7.75 | 2.07 | 32.34 | 0.94 |
| Tomasi di Lampedusa (Il gattopardo, 1958 *) | 2893 | 4.99 | 26.42 | 7.90 | 3.33 | 0.47 | −0.36 |
| Verga (I Malavoglia, 1881) | 4401 | 4.46 | 20.45 | 6.82 | 3.00 | 4.21 | 0.62 |
Table A2.
Authors of the novels of English literature. Number of total sentences; average number of characters per word, ; average number of words pr sentence, ; average number of word intervals, ; average number word intervals per sentence, ; multiplicity factor ; and mismatch index . Notice that for Dickens’ novels, Table 1 of [45] reported the number of sentences ending only with full stops; sentences ending with question marks and exclamation marks were not reported, contrarily to all other literary texts there reported. Moreover, the analysis conducted in [45] was performed by considering only the sentences ending with full stops; this is why the values of and there reported are larger (upper bounds) than those listed below.
Table A2.
Authors of the novels of English literature. Number of total sentences; average number of characters per word, ; average number of words pr sentence, ; average number of word intervals, ; average number word intervals per sentence, ; multiplicity factor ; and mismatch index . Notice that for Dickens’ novels, Table 1 of [45] reported the number of sentences ending only with full stops; sentences ending with question marks and exclamation marks were not reported, contrarily to all other literary texts there reported. Moreover, the analysis conducted in [45] was performed by considering only the sentences ending with full stops; this is why the values of and there reported are larger (upper bounds) than those listed below.
| Literary Work (Author, Year) | Sentences | ||||||
|---|---|---|---|---|---|---|---|
| The Adventures of Oliver Twist (C. Dickens, 1837–1839) | 9121 | 4.23 | 18.04 | 5.70 | 3.16 | 9.46 | 0.81 |
| David Copperfield (C. Dickens, 1849–1850) | 19,610 | 4.04 | 18.83 | 5.61 | 3.35 | 12.63 | 0.85 |
| Bleak House (C. Dickens, 1852–1853) | 20,967 | 4.23 | 16.95 | 6.59 | 2.57 | 56.98 | 0.97 |
| A Tale of Two Cities (C. Dickens, 1859) | 8098 | 4.26 | 18.27 | 6.19 | 2.93 | 11.89 | 0.84 |
| Our Mutual Friend (C. Dickens, 1864–1865) | 17,409 | 4.22 | 16.46 | 6.03 | 2.73 | 43.41 | 0.95 |
| Matthew King James (1611) | 1040 | 4.27 | 22.96 | 5.90 | 3.90 | 0.15 | −0.73 |
| Robinson Crusoe (D. Defoe, 1719) | 2393 | 3.94 | 52.90 | 7.12 | 7.40 | 0.00002 | −1.00 |
| Pride and Prejudice (J. Austen, 1813) | 6013 | 4.40 | 21.31 | 7.16 | 2.95 | 5.20 | 0.68 |
| Wuthering Heights (E. Brontë, 1845–1846) | 6352 | 4.27 | 17.78 | 5.97 | 2.97 | 9.83 | 0.82 |
| Vanity Fair (W. Thackeray, 1847–1848) | 13,007 | 4.63 | 21.95 | 6.73 | 3.25 | 5.26 | 0.68 |
| Moby Dick (H. Melville, 1851) | 9582 | 4.52 | 23.82 | 6.45 | 3.64 | 1.56 | 0.22 |
| The Mill On The Floss (G. Eliot, 1860) | 9018 | 4.29 | 23.84 | 7.09 | 3.35 | 2.17 | 0.37 |
| Alice’s Adventures in Wonderland (L. Carroll, 1865) | 1629 | 3.96 | 17.19 | 5.79 | 2.95 | 2.90 | 0.49 |
| Little Women (L.M. Alcott, 1868–1869) | 10,593 | 4.18 | 18.09 | 6.30 | 2.85 | 17.34 | 0.89 |
| Treasure Island (R. L. Stevenson, 1881–1882) | 3824 | 4.02 | 18.93 | 6.05 | 3.09 | 3.79 | 0.58 |
| Adventures of Huckleberry Finn (M. Twain, 1884) | 5887 | 3.85 | 19.39 | 6.63 | 2.94 | 7.05 | 0.75 |
| Three Men in a Boat (J.K. Jerome, 1889) | 5341 | 4.25 | 10.55 | 6.14 | 1.72 | 130.27 | 0.98 |
| The Picture of Dorian Gray (O. Wilde, 1890) | 4292 | 4.19 | 14.30 | 6.29 | 2.21 | 33.02 | 0.94 |
| The Jungle Book (R. Kipling, 1894) | 3214 | 4.11 | 16.46 | 7.14 | 2.29 | 14.10 | 0.87 |
| The War of the Worlds (H.G. Wells, 1897) | 3306 | 4.38 | 19.22 | 7.67 | 2.48 | 6.72 | 0.74 |
| The Wonderful Wizard of Oz (L.F. Baum, 1900) | 2219 | 4.017 | 17.90 | 7.63 | 2.34 | 7.02 | 0.75 |
| The Hound of The Baskervilles (A.C. Doyle, 1901–1902) | 4080 | 4.15 | 15.07 | 7.83 | 1.91 | 43.87 | 0.96 |
| Peter Pan (J.M. Barrie, 1902) | 3177 | 4.12 | 15.65 | 6.35 | 2.44 | 13.07 | 0.86 |
| A Little Princess (F.H. Burnett, 1902–1905) | 4838 | 4.18 | 14.26 | 6.79 | 2.09 | 46.97 | 0.96 |
| Martin Eden (J. London, 1908–1909) | 9173 | 4.32 | 15.61 | 6.76 | 2.30 | 46.33 | 0.96 |
| Women in love (D.H. Lawrence, 1920) | 16,048 | 4.26 | 11.62 | 5.22 | 2.22 | 216.86 | 0.99 |
| The Secret Adversary (A. Christie, 1922) | 8536 | 4.28 | 8.97 | 5.52 | 1.62 | 294.34 | 0.99 |
| The Sun Also Rises (E. Hemingway, 1926) | 7614 | 3.92 | 9.43 | 6.02 | 1.56 | 237.94 | 0.99 |
| A Farewell to Arms (H. Hemingway,1929) | 10,324 | 3.94 | 9.05 | 6.80 | 1.32 | 356.00 | 0.99 |
| Of Mice and Men (J. Steinbeck, 1937) | 3463 | 4.02 | 8.63 | 5.61 | 1.54 | 133.19 | 0.99 |
References
- Matricciani, E. Is Short-Term Memory Made of Two Processing Units? Clues from Italian and English Literatures down Several Centuries. Information 2024, 15, 6. [Google Scholar] [CrossRef]
- Deniz, F.; Nunez-Elizalde, A.O.; Huth, A.G.; Gallant Jack, L. The Representation of Semantic Information Across Human Cerebral Cortex During Listening Versus Reading Is Invariant to Stimulus Modality. J. Neurosci. 2019, 39, 7722–7736. [Google Scholar] [CrossRef]
- Miller, G.A. The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychol. Rev. 1956, 63, 81–97. [Google Scholar] [CrossRef]
- Crowder, R.G. Short-term memory: Where do we stand? Mem. Cogn. 1993, 21, 142–145. [Google Scholar] [CrossRef]
- Lisman, J.E.; Idiart, M.A.P. Storage of 7 ± 2 Short-Term Memories in Oscillatory Subcycles. Science 1995, 267, 1512–1515. [Google Scholar] [CrossRef]
- Cowan, N. The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behav. Brain Sci. 2001, 24, 87–114. [Google Scholar] [CrossRef]
- Bachelder, B.L. The Magical Number 7 ± 2: Span Theory on Capacity Limitations. Behav. Brain Sci. 2001, 24, 116–117. [Google Scholar] [CrossRef]
- Saaty, T.L.; Ozdemir, M.S. Why the Magic Number Seven Plus or Minus Two. Math. Comput. Model. 2003, 38, 233–244. [Google Scholar] [CrossRef]
- Burgess, N.; Hitch, G.J. A revised model of short-term memory and long-term learning of verbal sequences. J. Mem. Lang. 2006, 55, 627–652. [Google Scholar] [CrossRef]
- Richardson, J.T.E. Measures of short-term memory: A historical review. Cortex 2007, 43, 635–650. [Google Scholar] [CrossRef]
- Mathy, F.; Feldman, J. What’s magic about magic numbers? Chunking and data compression in short-term memory. Cognition 2012, 122, 346–362. [Google Scholar] [CrossRef] [PubMed]
- Gignac, G.E. The Magical Numbers 7 and 4 Are Resistant to the Flynn Effect: No Evidence for Increases in Forward or Backward Recall across 85 Years of Data. Intelligence 2015, 48, 85–95. [Google Scholar] [CrossRef]
- Trauzettel-Klosinski, S.; Dietz, K. Standardized Assessment of Reading Performance: The New International Reading Speed Texts IreST. Investig. Ophthalmol. Vis. Sci. 2012, 53, 5452–5461. [Google Scholar] [CrossRef] [PubMed]
- Melton, A.W. Implications of Short-Term Memory for a General Theory of Memory. J. Verbal Learn. Verbal Behav. 1963, 2, 1–21. [Google Scholar] [CrossRef]
- Atkinson, R.C.; Shiffrin, R.M. The Control of Short-Term Memory. Sci. Am. 1971, 225, 82–91. [Google Scholar] [CrossRef]
- Murdock, B.B. Short-Term Memory. Psychol. Learn. Motiv. 1972, 5, 67–127. [Google Scholar]
- Baddeley, A.D.; Thomson, N.; Buchanan, M. Word Length and the Structure of Short-Term Memory. J. Verbal Learn. Verbal Behav. 1975, 14, 575–589. [Google Scholar] [CrossRef]
- Case, R.; Midian Kurland, D.; Goldberg, J. Operational efficiency and the growth of short-term memory span. J. Exp. Child Psychol. 1982, 33, 386–404. [Google Scholar] [CrossRef]
- Grondin, S. A temporal account of the limited processing capacity. Behav. Brain Sci. 2000, 24, 122–123. [Google Scholar] [CrossRef]
- Pothos, E.M.; Joula, P. Linguistic structure and short-term memory. Behav. Brain Sci. 2000, 138–139. [Google Scholar] [CrossRef]
- Conway, A.R.A.; Cowan, N.; Michael, F.; Bunting, M.F.; Therriaulta, D.J.; Minkoff, S.R.B. A latent variable analysis of working memory capacity, short-term memory capacity, processing speed, and general fluid intelligence. Intelligence 2002, 30, 163–183. [Google Scholar] [CrossRef]
- Jonides, J.; Lewis, R.L.; Nee, D.E.; Lustig, C.A.; Berman, M.G.; Moore, K.S. The Mind and Brain of Short-Term Memory. Annu. Rev. Psychol. 2008, 69, 193–224. [Google Scholar] [CrossRef] [PubMed]
- Barrouillest, P.; Camos, V. As Time Goes by: Temporal Constraints in Working Memory. Curr. Dir. Psychol. Sci. 2012, 21, 413–419. [Google Scholar] [CrossRef]
- Potter, M.C. Conceptual short term memory in perception and thought. Front. Psychol. 2012, 3, 113. [Google Scholar] [CrossRef]
- Jones, G.; Macken, B. Questioning short-term memory and its measurements: Why digit span measures long-term associative learning. Cognition 2015, 144, 1–13. [Google Scholar] [CrossRef] [PubMed]
- Chekaf, M.; Cowan, N.; Mathy, F. Chunk formation in immediate memory and how it relates to data compression. Cognition 2016, 155, 96–107. [Google Scholar] [CrossRef]
- Norris, D. Short-Term Memory and Long-Term Memory Are Still Different. Psychol. Bull. 2017, 143, 992–1009. [Google Scholar] [CrossRef]
- Houdt, G.V.; Mosquera, C.; Napoles, G. A review on the long short-term memory model. Artif. Intell. Rev. 2020, 53, 5929–5955. [Google Scholar] [CrossRef]
- Islam, M.; Sarkar, A.; Hossain, M.; Ahmed, M.; Ferdous, A. Prediction of Attention and Short-Term Memory Loss by EEG Workload Estimation. J. Biosci. Med. 2023, 11, 304–318. [Google Scholar] [CrossRef]
- Rosenzweig, M.R.; Bennett, E.L.; Colombo, P.J.; Lee, P.D.W. Short-term, intermediate-term and Long-term memories. Behav. Brain Res. 1993, 57, 193–198. [Google Scholar] [CrossRef]
- Kaminski, J. Intermediate-Term Memory as a Bridge between Working and Long-Term Memory. J. Neurosci. 2017, 37, 5045–5047. [Google Scholar] [CrossRef][Green Version]
- Matricciani, E. Deep Language Statistics of Italian throughout Seven Centuries of Literature and Empirical Connections with Miller’s 7 ∓ 2 Law and Short-Term Memory. Open J. Stat. 2019, 9, 373–406. [Google Scholar] [CrossRef]
- Strinati, E.C.; Barbarossa, S. 6G Networks: Beyond Shannon towards Semantic and Goal-Oriented Communications. Comput. Netw. 2021, 190, 107930. [Google Scholar] [CrossRef]
- Shi, G.; Xiao, Y.; Li, Y.; Xie, X. From semantic communication to semantic-aware networking: Model, architecture, and open problems. IEEE Commun. Mag. 2021, 59, 44–50. [Google Scholar] [CrossRef]
- Xie, H.; Qin, Z.; Li, G.Y.; Juang, B.H. Deep learning enabled semantic communication systems. IEEE Trans. Signal Process. 2021, 69, 2663–2675. [Google Scholar] [CrossRef]
- Luo, X.; Chen, H.H.; Guo, Q. Semantic communications: Overview, open issues, and future research directions. IEEE Wirel. Commun. 2022, 29, 210–219. [Google Scholar] [CrossRef]
- Yang, W.; Du, H.; Liew, Z.Q.; Lim, W.Y.B.; Xiong, Z.; Niyato, D.; Chi, X.; Shen, X.; Miao, C. Semantic Communications for Future Internet: Fundamentals, Applications, and Challenges. IEEE Commun. Surv. Tutor. 2023, 25, 213–250. [Google Scholar] [CrossRef]
- Xie, H.; Qin, A. A lite distributed semantic communication system for internet of things. IEEE J. Sel. Areas Commun. 2021, 39, 142–153. [Google Scholar]
- Bellegarda, J.R. Exploiting Latent Semantic Information in Statistical Language Modeling. Proc. IEEE 2000, 88, 1279–1296. [Google Scholar] [CrossRef]
- D’Alfonso, S. On Quantifying Semantic Information. Information 2011, 2, 61–101. [Google Scholar] [CrossRef]
- Zhong, Y. A Theory of Semantic Information. China Commun. 2017, 14, 1–17. [Google Scholar] [CrossRef]
- Papoulis Papoulis, A. Probability & Statistics; Prentice Hall: Hoboken, NJ, USA, 1990. [Google Scholar]
- Matricciani, E. Readability Indices Do Not Say It All on a Text Readability. Analytics 2023, 2, 296–314. [Google Scholar] [CrossRef]
- Matricciani, E. The Theory of Linguistic Channels in Alphabetical Texts; Cambridge Scholars Publishing: Newcastle upon Tyne, UK, 2024. [Google Scholar]
- Matricciani, E. Capacity of Linguistic Communication Channels in Literary Texts: Application to Charles Dickens’ Novels. Information 2023, 14, 68. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).


















