Abstract
In the first part of the article, we recall our general theory of linguistic channels—based on regression lines between deep language parameters—and study their capacity and interdependence. In the second part, we apply the theory to novels written by Charles Dickens and other authors of English literature, including the Gospels in the King James version of the Bible. In literary works (or in any long texts), there are multiple communication channels. The theory considers not only averages but also correlation coefficients. The capacity of linguistic channels is a Gaussian stochastic variable. The similarity between two channels is measured by the likeness index. Dickens’ novels show striking and unexpected mathematical/statistical similarity to the synoptic Gospels. The Pythagorean distance, defined in a suitable Cartesian plane involving deep language parameters, and the likeness index correlate with an inverse proportional relationship. A similar approach can be applied to any literary corpus written in any alphabetical language.
1. Linguistic Communication Channels in Literary Texts
In recent papers [,,,], we have developed a new and general statistical theory on the deep mathematical structure of literary texts (or any long text) written in alphabetical languages—including translations—based on Shannon’s communication theory [], which involves linguistic stochastic variables and communication channels suitably defined. In the theory, “translation” means not only the conversion of a text from one language to another—which is properly understood, of course, as translation—but also how some linguistic parameters of a text are related to those of another text, either in the same language or in another language. “Translation”, therefore, in the general theory, refers also to the case in which a text is compared (metaphorically “translated” into) to another text, regardless of the language of the two texts.
The theory, whose features are further developed in the present article, has important limitations because it gives no clues as to the correct use of words and grammar, the variety and richness of the literary expression, or its beauty or efficacy. It does not measure the quality and clearness of ideas. The comprehension of a text is the result of many other factors, the most important being the reader’s culture and reading habits, besides the obvious understanding of the language. In spite of these limitations, the theory can be very useful, because it can be applied to any alphabetical language, such as those studied in [], because it deals with the underlying mathematical structure of texts, which can be very similar from language to language, therefore defeating the apparent scattering due to the mythical Babel Tower event.
The theory does not follow the actual paradigm of linguistic studies. Most studies on the relationships between texts concern translation because of the importance of automatic (i.e., machine) translation. Translation transfers meaning from one set of sequential symbols into another set of sequential symbols and was studied as a language learning methodology, or as part of comparative literature, with theories and models imported from other disciplines [,]. References [,,,,,,] report results not based on the mathematical analysis of texts, as the theory here further developed does. However, when a mathematical approach is used, as in References [,,,,,,,,,,,,], most of these studies neither consider Shannon’s communication theory nor the fundamental connection that some linguistic variables seem to have with the reading ability and short-term memory capacity of readers [,,,]. In fact, these studies are mainly concerned with automatic translations, not with the high-level direct response of human readers. Very often, they refer only to one very limited linguistic variable, e.g., phrases [], and not to sentences, which convey a completely developed thought, rather than deep language parameters, as our theory does.
As stated in [], statistical automatic translation is a process in which the text to be translated is “decoded” by eliminating the noise by adjusting lexical and syntactic divergences to reveal the intended message. In our theory, what is defined as “noise”—given by quantitative differences between the source text (input) and translated text (output)—must not be eliminated because it makes the translation readable and matched to the reader’s short-term memory capacity [], a connection never considered in [,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,], references that represent only a small part of the vast literature on machine translation.
Besides the total numbers of characters, words, sentences, and interpunctions (punctuation marks), the theory considers the number of words , the number of sentences per chapter, or any chosen subdivision of a literary text, large enough to provide reliable statistics, e.g., a few hundred words. Moreover, it also considers what we have termed the deep language variables, namely the number of characters per word , words per sentence , words per interpunctions (this parameter, called also the “word interval” [], is linked to the short-term memory capacity of readers), and interpunctions per sentence (this parameter gives also the number of contained in a sentence).
To study the apparently chaotic data that emerge from literary texts in any language, the theory compares a text (the reference, or input text) to another text (output), with a complex communication channel—composed of several parallel channels [], one of which is explicitly considered in the present article—in which both the input and output are affected by “noise”, i.e., by the different scattering of the data around an average relationship, namely a regression line.
In [], we have shown how much the mutual mathematical relationships of a literary work written in a language are saved or lost in translating it into another language. To make objective comparisons, we have defined the likeness index , based on the probability and communication theory of noisy digital channels.
We have shown (see Section 4 of []) that two linguistic variables, e.g., the variables and , can be linearly linked by regression lines. This is a general feature of texts. For example, if we consider the regression line linking to in a reference text and that found in another text, it is possible to link the of the first text to the of the second text with another regression line without explicitly calculating its parameters (slope and correlation coefficient) from the samples, because the mathematical problem has the same structure of the theory developed in Section 1 of [].
In [], we have applied the theory developed in [,,] to compare how a literary character speaks to different audiences by diversifying and adjusting (“fine tuning”) two important linguistic communication channels, namely the “sentences channel”—this channel links the sentences of the input text to the sentences of the output text for the same number of words—and the “interpunctions channel”—this channel links the word intervals of the two texts for the same number of sentences. We have shown that the theory can “measure” how an author shapes a character’s speaking to different audiences by modulating the deep language parameters.
In the present article, we have developed the theory of linguistic channels further. The article is structured in two parts. In the first part, we study the capacity of linguistic channels and show their interdependence. In the second part, to show some features and usefulness of the theory, we apply it to novels written by Charles Dickens (1812–1870) and compare their statistical/mathematical features to those of a few novels of English literature. Moreover, a comparison with the King James version of the Gospels shows a striking and unexpected similarity to Dickens’ novels.
After this introduction, Section 2 deals with the fundamental relationships in linguistic channels; Section 3 deals with their experimental signal-to-noise ratios (Monte Carlo simulations) and recalls the meaning of self- and cross-channels; Section 4 deals with the Shannon capacity and its probability distribution. In the second part of the article, paralleling the theoretical Section 2, Section 3, Section 4 and Section 5 deals with Charles Dickens’ novels and their deep language variables; Section 6 reports the experimental signal-to-noise ratio of the self- and cross-channels in Dickens’ novels and in the Gospel of Matthew; Section 7 deals with the Shannon capacity of the self- and cross-channels and the likeness index; Section 8 deals with the likely influence of the Gospels on Dickens’ novels; Section 9 reports some final remarks, and Section 10 concludes. Appendix A reports some numerical tables on the channels involving the Gospels.
2. Fundamental Relationships in Linguistic Communication Channels
In this section, we recall the general theory of linguistic channels [,,,]. In a literary work, an independent (reference) variable . (e.g., the number of words per chapter and a dependent variable (e.g., the number of sentences in the same chapter can be related by the regression line passing through the origin of the Cartesian coordinates:
In Equation (1), is the slope of the line.
Let us consider two different text blocks and , e.g., the chapters of work and work . Equation (1) does not give the full dependence of the two variables because it links only average conditional values. We can write more general linear relationships, which consider the scattering of the data—measured by the correlation coefficients and , respectively, not considered in Equation (1)—around the average values (measured by the slopes and ):
The linear model Equation (1) connects and only on the average, while the linear model Equation (2) introduces additive “noise” through the stochastic variables and , with zero mean value [,,]. The noise is due to the correlation coefficient , not considered in Equation (1).
We can compare two literary works by eliminating . In other words, we compare the output variable for the same number of the input variable . In the example previously mentioned, we can compare the number of sentences in two works—for an equal number of words—by considering not only the average relationship, Equation (1), but also the scattering of the data, measured by their correlation; see Equation (2). We refer to this communication channel as the “sentences channel” and to this processing as “fine tuning” because it deepens the analysis of the data and can provide more insight into the relationship between two literary works, or more general texts.
By eliminating from Equation (2), we obtain the linear relationship between, now, the input number of sentences in work (now, the reference, input work) and the number of sentences in text (now, the output work):
Compared to the new reference work , the slope is given by
The noise source that produces the correlation coefficient between and is given by
The “regression noise-to-signal ratio”, due to , of the new channel is given by []
The unknown correlation coefficient between and is given by []
The “correlation noise-to-signal ratio”, , due to , of the new channel from text to text is given by []
Because the two noise sources are disjoint and additive, the total noise-to-signal ratio of the channel connecting text to text is given by []
Notice that Equation (9) can be represented graphically []. Finally, the total signal-to-noise ratio is given by
Of course, we expect, and it is so in the following, that no channel can yield and ; therefore, , a case referred to as the ideal channel, unless a text is compared with itself (self-comparison, self-channel). In practice, we always find and . The slope . measures the multiplicative “bias” of the dependent variable compared to the independent variable; the correlation coefficient measures how “precise” the linear best fit is.
In conclusion, the slope is the source of the regression noise; the correlation coefficient is the source of the correlation noise of the channel.
3. Experimental Signal-to-Noise Ratios in Linguistic Channels
Because of the different sample size used in calculating a regression line, its slope and correlation coefficient —being stochastic variables—are characterized by average values and standard deviations, which depend on the sample size []. Obviously, the theory would yield more precise estimates of —see Equation (10)—for a larger sample size. With a small sample size—as is the case with the number of chapters of a literary text—the standard deviations of and can give too large a variation in (see the sensitivity of this parameter to the slope and the correlation coefficient in []). To avoid this inaccuracy—due to a small sample size, not to the theory of Section 2—we have defined [] and discussed [,] a “renormalization” based on Monte Carlo simulations, whose results can be considered “experimental”. Therefore, the results of the simulation can replace, as discussed in [], the theoretical values.
Let us recall the steps of the Monte Carlo simulation by explicitly considering the sentences channel [].
Let the literary work be the “output”, of which we consider disjoint block texts (e.g., chapters), and let us compare it with a particular input literary work characterized by a regression line, as detailed in Section 2. The steps of the Monte Carlo simulation are the following:
- Generate independent numbers (the number of disjoint block texts, e.g., chapters) from a discrete uniform probability distribution in the range 1 to , with replacement, i.e., a text can be selected more than once.
- “Write” another possible “work ” with new disjoint texts, e.g., the sequence 2; 1; ; ; hence, take text 2, followed by text 1, text text up to texts. A block text can appear twice (with probability ), three times (with probability ), etc., and the new “work ” can contain a number of words greater or smaller than the original work, on average (the differences are small and do not affect the final statistical results and analysis).
- Calculate the parameters and of the regression line between words (independent variable) and sentences (dependent variable) in the new “work ”, namely Equation (1).
- Compare and of the new “work ” (output, dependent work) with any other work (input, independent work, and ), in the “cross-channels” so defined, including the original work (a particular case referred to as the “self-channel”).
- Calculate , , and of the cross-channels (linking sentences to sentences), according to the theory of Section 2.
- Consider the values of so obtained, in Equation (10), as “experimental” results .
- Repeat Steps 1 to 6 many times to obtain reliable results (we have done so 5000 times because this number of simulations ensures reliable results down to two decimal digits in ).
In conclusion, the Monte Carlo simulation should eliminate the inaccuracy in estimating the slope and correlation coefficient due to a small sample size. However, besides the usefulness of the simulation as a “renormalization” tool to avoid small sample size inaccuracy, as shown in [,], there is another property—very likely more interesting—of the new generated literary works. In fact, as the mathematical theory does not consider meaning, the new works obtained in Step 2 might have been “written” by the author, because they maintain the main statistical properties of the deep language parameters of the original text. In other words, they are “literary works” that the author might have written at the time that he wrote the original work.
4. Capacity of Self- and Cross-Channels and Its Probability Distribution
In Reference [] (see Figure 7 of []), we have shown that the probability density function of in both self- and cross-channels can be approximately modeled as Gaussian, with average value (dB) and standard deviation (dB), i.e., the values reported, for example, in Tables 4 and 5 of [], or below.
In this section, we determine the probability density function of the Shannon capacity of self- and cross-channels, starting from the Gaussian probability density function of . For this calculation, we need to apply the theory of variable transformation [].
First, it can be shown that the probability density function of the linear signal-to-noise ratio
is given by the log-normal probability density function with average value and standard deviation :
Now, each channel has capacity (bits per symbol), which can be conservatively (see the discussion in []) calculated according to Shannon []:
Therefore, the capacity of linguistic self- and cross-channels, as those relating to the sentences channel, can be calculated from Equation (13), in which has a probability density function to be determined from the log-normal probability density function, as in Equation (12).
By setting , the theory of variable transformation applied to Equations (12) and (13) gives the following probability density function of (natural logs):
Now, if , (as , )—a condition that applies to all cases studied below—it can be approximated, in a large range of , with
Finally, by setting
the probability density function is given by
In other words, if then the probability distribution function of channel capacity is Gaussian in a large range (, of course), with average value and standard deviation given by Equation (16).
In [], we explored a means of comparing the signal-to-noise ratios of self- and cross-channels objectively, and possibly also obtaining more insight into texts’ mathematical likeness. In comparing a self-channel with a cross-channel, the probability of mistaking one work for another is a binary problem because a decision must be taken between two alternatives. The problem is classical in binary digital communication channels affected by noise, as recalled in []. In digital communication, “error” means that bit 1 is mistaken for bit 0 or vice versa; therefore, the channel performance worsens as the error frequency (i.e., the probability of error) increases. However, in linguistics, self- and cross-channel “error” means that a text can be more or less mistaken, or confused, with another text; consequently, two texts are more similar as the probability of error increases. Therefore, as in [], a large error probability means that two literary works are mathematically similar in the considered channel.
As with the likeness index defined in [] for the of self- and cross-channels, we could define also the “capacity likeness index” . Again, ; means totally independent texts, and means totally dependent texts. However, if Equation (16) holds—as is the case in the literary works here considered and shown below—then the capacity likeness index of the self- and cross-channels and the likeness index concerning , coincide, because the two Gaussian densities of are obtained from those of
by rigidly shifting them to the left in the of the same quantity. Therefore, in the following we do not distinguish between the two indices.
In the following second part of the article, we apply the theory to literary works belonging to the English literature by mainly studying Dickens’ novels, including the Gospels of the classical King James translation.
5. Charles Dickens’ Novels and Deep Language Variables
The novels of Charles Dickens that are studied are listed in Table 1. They range from one of the earliest ones, The Adventures of Oliver Twist (1837–1839), to the last one, Our Mutual Friend (1864–1865). This particular choice may be useful to study the possible time dependence of their mathematical characteristics.
Table 1.
Charles Dickens’ novels. Number of chapters (i.e., the number of samples considered in calculating the regression lines of the theory), total number of characters contained in the words, total number of words and sentences, deep language parameters , , , , with standard deviation reported in the second line.
Table 2 and Table 3 list the other English literary works—including the Gospel according to Matthew in the King James version of the Bible—studied and compared to Dickens’ novels. The novels belong to the XIX and XX centuries and have been chosen because their texts are freely available in digital format in the internet.
Table 2.
English literature. Literary works are ordered according to publication years. Number of chapters (i.e., the number of samples considered in calculating the regression lines of the theory), total number of characters related only to words, total number of words and sentences. The order number is useful to identify the single literary works in Figure 2.
Table 3.
English literature. Literary works are ordered according to publication years. Deep language parameters , , , , with standard deviation reported in the second line. The order number is useful to identify the single literary works in Figure 2.
Table 1 and Table 2 report also the number of chapters (i.e., the number of samples considered in calculating the regression lines of the theory), the total number of characters contained in the words, and the total number of words and sentences, followed by the deep language parameters, namely , , , . These data have been calculated manually as described in [,,].
For Dickens’ novels, besides their different sizes—from the shortest (A Tale of Two Cities) to the longest (David Copperfield)—it is interesting to note the very similar average structure of sentences in terms of words per sentence , approximately These average values, however, give rise to significant differences when the sentences cannel is studied in Section 6 (fine tuning) by considering also the spreading of the data due to correlation coefficients.
The average values reported in Table 1 and Table 2 can be analyzed in two interesting ways: (a) by studying the relationship between and , and its very likely connection with Miller’s law []; (b) by showing a high-level overall view of the literary works in a Cartesian plane.
5.1. Relationship between and , Miller’s Law
An interesting observation of the averages reported in Table 3 is the range of the word interval from approximately 5.2 (Women in Love) to 7.8 (The Hound of The Baskervilles), a significant interval in Miller’s law . Because is very likely an estimate of the capacity of the short-term memory buffer, the short-term memory of the intended readers of David Copperfield () is less engaged than that of the readers of Bleak House ().
Figure 1 shows the scatter plot between the average values of and , for the works listed in Table 3, together with the non-linear regression line (least-square best-fit) that models, as in [], the average value of versus the average value of , given by
Figure 1.
Scatter plot between and of the literary works of Table 1 and Table 2, together with the non-linear regression line (best-fit line) that models, on average, versus for these works, Equation (1), and Miller’s bounds. Dickens’ novels are represented by the circles: The Adventures of Oliver Twist, blue; David Copperfield, red; Bleak House, magenta; A Tale of Two Cities, cyan; Our Mutual Friend, black. Matthew: yellow square. The blue crosses refer to the other works listed in Table 2. The mark on the far right refers to Robinson Crusoe.
In Equation (18), (words per interpunctions) is the horizontal asymptote, and (words per sentence) is the value of at which the exponential in Equation (18) falls at of its maximum value. Notice that the asymptotic value is very close to its center value 7, and all data fall within Miller’s range , the same as the literary works in the Italian literature [].
The trend modeled by Equation (18) can be justified as follows. As the number of words in a sentence, increases, the number of word intervals, , can increase but not linearly, because the short-term memory cannot hold, approximately, a number of words larger than that empirically predicted by Miller’s law; therefore, saturation must occur []. This is clearly shown by the right-most couple (57.747, 7.119) in Figure 1 due to Robinson Crusoe.
In other words, scatter plots, such as that shown in Figure 1, drawn also for other literature [], should give an insight into the short-term memory capacity engaged in reading the texts. The values found for each author set the average size of the short-term memory capacity that their readers should have in order to read the literary work more easily.
The average value of the deep language parameters can be used to provide a first assessment of how much the literary works are similar, or “close”, by reporting them in a Cartesian plane as vectors, a graphical representation discussed in detail in [,,] and here briefly recalled.
5.2. The Vector Plane
Let us consider the following six vectors of the indicated components ), ), ), ), ), ) and their resulting vector:
The choice of which parameter represents the component in the abscissa and ordinate axes is not important because, once the choice is made, the numerical results will depend on it, but not the relative comparisons and general conclusions.
Figure 2 shows the resulting vector (19). The Cartesian coordinates reported have been “normalized” so that Of Mice and Men is located at (0,0) (blue pentagon) and Moby Dick is located at (1,1) (green triangle with vertex pointing down). This normalized representation allows us to maintain the relative distances by assuming the same unit in the two coordinates.
Figure 2.
Normalized coordinates and of the resulting vector (18) of the literary works listed in Table 1 and Table 2, normalized so that Of Mice and Men is located at the origin (0,0) and Moby Dick is located at (1,1). Dickens’ novels are represented by the circles: The Adventures of Oliver Twist, a, blue; David Copperfield, b, red; Bleak House, d, magenta; A Tale of Two Cities, c, cyan; Our Mutual Friend, e, black. Matthew: 1, yellow square. The black square B is the barycenter of Dickens’ works. The other novels are numbered according to the order reported in Table 2.
It can be noted that, compared to the other English works, Dickens’s novels are all very near to each other, within the circle drawn from their barycenter (black square) with the radius reaching David Copperfield (red circle). It can also be noted that there is a clear distinction between XIX century (magenta and green marks) and XX century novels (blue marks), therefore introducing a time arrow in the development of the English literature, at least for the sampled works. The outlying vector (1.443, 2.211) of Robinson Crusoe (1719) is not reported due to space constraints.
Curiously, the Gospel of Matthew in the King James version of the Bible (yellow square, see Table 3) is very near to Dicken’s barycenter. This is an unexpected coincidence, which requires further investigation. Did the classical New Testament books available at that time—namely the King James translations—affect the mathematical structure of Dickens’ writings? In Section 8, we propose a likely answer to this question by considering also the other three Gospels (Mark, Luke, John).
As stated before, all these findings and observations refer to a high-level comparison because they involve only average values. In literary works, however, there are multiple communication channels [,], one of which is the so-called “sentences channel”, a channel that linearly links the sentences of two literary works for an equal number of words. The theory of these channels includes not only averages, such as regression lines, but also correlation coefficients, as recalled in the first part of this article; therefore, in the next section, we apply the theory to Dickens’ works.
6. Experimental Signal-to-Noise Ratio of Self- and Cross-Channels
As discussed in Section 3, we consider the values of concerning the sentences channel. Table 4 lists the slope and the correlation coefficient of the regression line between the number of sentences
(dependent variable) and the number of words (independent variable) per chapter in Dickens’ works and in Matthew. Four decimal digits are reported because some values differ only from the third digit. These data are the parameters of the input literary works required by the theory.
Table 4.
Slope and correlation coefficient of the regression line between the number of sentences (dependent variable) and the number of words (independent variable). Four decimal digits are reported because some values differ only from the third digit.
We can notice, for example, that the slopes of Matthew and A Tale of Two Cities are equal to the forth decimal digit, but not so the correlation coefficients—recall the sensitivity of the signal-to-noise ratio to this parameter, discussed in []. Therefore, in this case, is practically given by the regression noise; see Equation (8). In other words, a comparison based only on averages would conclude that the two texts are mathematically identical: only a “fine tuning” study of the sentences channel would show, as we do below, that similarity does exist but not to this extent.
Table 5 shows the average value (dB) and standard deviation (dB) of (dB) in the self- and cross-channels, the correlation coefficient , and the slope of the regression line between the number of sentences of Oliver Twist (output channel) versus the number of sentences in the other Dickens novels (input channels), for an equal number of words.
Table 5.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, average value and standard deviation of correlation coefficient , and slope of the regression line between the number of sentences of Oliver Twist (channel output) versus the number of sentences in the other Dickens novels (channel input), for equal number of words, in the sentences channel. Four decimal digits are reported because some values differ only from the third digit.
Table 6, Table 7, Table 8 and Table 9 show the results when the output channel is David Copperfield, Bleak House, A Tale of Two Cities, Our Mutual Friend. For example, according to Table 4, in David Copperfield, 100 words give sentences on average; therefore, from Table 5, this number of sentences is “translated” into sentences in Oliver Twist, with correlation coefficient . Of course, the largest statistical value cannot exceed 1.
Table 6.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of David Copperfield (channel output, self-channel) versus the number of sentences in the other Dickens novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Table 7.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Bleak House (channel output, self-channel) versus the number of sentences in the other Dickens novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Table 8.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of A Tale of Two Cities (channel output, self-channel) versus the number of sentences in the other Dickens novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Table 9.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Our Mutual Friend (channel output, self-channel) versus the number of sentences in the other Dickens novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Finally, notice the asymmetry typical of linguistic channels [,,,]. For example, from Table 5, in the cross-channel of David Copperfield, (dB), while, from Table 6, in the cross-channel of Oliver Twist, (dB).
Notice that the standard deviation of in self-channels is approximately dBs, independently of the average value, and when the average value of a cross-channel moves closer to that of the self-channel, also its standard deviation tends to assume the same value (e.g., Our Mutual Friend in Table 7), a typical feature of cross-channels being very similar to self-channels [,,].
Now, as discussed in Section 3, self-channels can describe all possible literary works that, by maintaining the same statistical properties of the original work, the author might have written at the same time as the original one. Therefore, the closer the parameters of the cross-channels to those of the self-channel, the more similar are the input and output works. In other words, the Gaussian probability density function of a cross-channel can largely overlap with that of the self-channel. This superposition is quantified by the likeness index, as shown in [] for . In the next section, we show this overlap for the channel capacity and calculate the likeness index.
7. Capacity of Self- and Cross-Channels and Likeness Index
In this section, we calculate the capacity of self- and cross-channels. We assume that the signal-to-noise ratio of these channels is Gaussian, with the average value and standard deviation given in Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10, and calculate from Equation (13), by running a Monte Carlo simulation (100,000 simulations).
Table 10.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Matthew (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Figure 3 shows the results for Bleak House (Table 7), together with the theoretical Gaussian probability density function calculated according to the approximation given by Equations (16) and (17). The simulated data show a probability density function that agrees extremely well with the theoretical model. In fact, the average value and standard deviation of the simulated data agree to the forth digit with those calculated with Equations (16) and (17), therefore confirming the validity of the hypotheses assumed for the Gaussian model, in Equation (17).
Figure 3.
Upper panel: Probability density histograms of self- and cross-channel capacity. Bleak House: magenta (output); The Adventures of Oliver Twist: blue; David Copperfield: red; A Tale of Two Cities; cyan; Our Mutual Friend: black. Matthew: yellow. The continuous black lines are the theoretical densities given by Equation (17). Lower panel: Probability distribution functions.
Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 show the Monte Carlo results for the channels listed in Table 5, Table 6, Table 7, Table 8 and Table 9. We see a further confirmation that the capacity of all self- and cross-channels can be very well modeled as Gaussian. This result applies also to the case in which Matthew is the output text (Figure 8). Moreover, the average value of the worst cross-channels tends to be approximately half of that of the self-channel.
Figure 4.
Upper panel: Probability density histograms of self- and cross-channel capacity. The Adventures of Oliver Twist: blue (output); David Copperfield: red; Bleak House: magenta; A Tale of Two Cities; cyan; Our Mutual Friend: black. Matthew: yellow. Lower panel: Probability distribution functions.
Figure 5.
Upper panel: Probability density histograms of self- and cross-channel capacity. David Copperfield: red (output); The Adventures of Oliver Twist: blue; Bleak House: magenta; A Tale of Two Cities; cyan; Our Mutual Friend: black. Matthew: yellow. Lower panel: Probability distribution functions.
Figure 6.
Upper panel: Probability density histograms of self- and cross-channel capacity. A Tale of Two Cities; cyan (output); David Copperfield: red; The Adventures of Oliver Twist: blue; Bleak House: magenta; Our Mutual Friend: black. Matthew: yellow. Lower panel: Probability distribution functions.
Figure 7.
Upper panel: Probability density histograms of self- and cross-channel capacity. Our Mutual Friend: black (output); A Tale of Two Cities; cyan; David Copperfield: red; The Adventures of Oliver Twist: blue; Bleak House: magenta; Matthew: yellow. Lower panel: Probability distribution functions.
Figure 8.
Upper panel: Probability density histograms of self- and cross-channel capacity. ; Matthew: yellow (output); Our Mutual Friend: black; A Tale of Two Cities; cyan; David Copperfield: red; The Adventures of Oliver Twist: blue; Bleak House: magenta. Lower panel: Probability distribution functions.
In conclusion, if the stochastic variable of a linguistic channel is Gaussian, then also its capacity (bits per symbol) is Gaussian.
Now, from the Gaussian probability density functions shown in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8, the likeness index can be calculated. As discussed in Section 4, this index coincides with the likeness index ; therefore, we do not distinguish between them in the following. Table 11 reports the results for Dickens’ works. The title in the first line indicates the output novel, and the title in the first column indicates the input novel (regression line given in Table 4). The output novel is the work that produces the regression lines in Step 2 of the Monte Carlo simulation, and the input work (with fixed regression line, Table 4) produces the cross-channel.
Table 11.
Likeness index between the indicated literary works, sentences channel. The work in the first line indicates the output text; the text in the first column indicates the input text (regression line given in Table 3). For example, in the channel Our Mutual Friend Bleak House, , while in the reverse channel, Bleak House Our Mutual Friend, .
For example, in the channel Our Mutual FriendBleak House, . In the reverse channel, Bleak HouseOur Mutual Friend, . The meaning of these results is the following: in the channel Bleak HouseOur Mutual Friend, the regression line between the sentences and words in every new Our Mutual Friend simulated in Step 2 of the Monte Carlo algorithm of Section 3 is very similar to that of the input text Bleak House (regression given in Table 4), so that the theory of Section 2 produces, in the end, this large . In other words, the regression of Bleak House belongs to the set of regression lines (self-channel) of Our Mutual Friend, a “belonging” described by the two Gaussian densities and measured by .
Now, similarly, in the channel Our Mutual FriendBleak House, the regression line between the sentences and words in every new Bleak House is quite similar to that of the input text Our Mutual Friend (Table 3), so that . However, in this case, this asymmetry may indicate a time arrow, because Bleak House (written earlier) seems to be more “contained” in A Mutual Friend (written later) than the reverse, because . The time arrow, however, is not evident in other cases.
It is very interesting and surprising to compare them with Matthew. It can be noticed, in fact, that Bleak House () and A Tale of Two Cities () are more similar to Matthew than the other novels. In other words, Matthew seems to have affected the statistics of the sentences in these two Dickens works more than those found in Our Mutual Friend, Oliver Twist, and David Copperfield. Appendix A reports the tables for the other Gospels.
Now, let us consider in more depth the likely influence of the Gospels of the King James translation on Dickens’ writing.
8. The Likely Influence of the Gospels on Dickens’ Novels
The very similar values of the deep language parameters in Dickens’ novels and in the Gospel according to Matthew (Figure 1 and Figure 2) may be a trace of the influence unconsciously left in Dickens’ style after researching the life of Jesus of Nazareth and writing The Life of Our Lord for his children, published only in 1934 []. Dickens felt the need to impart some religious instruction to his children by writing a simplified version of the Gospels. According to scholars [,,], in his novels, all the strongest illustrations are derived from the New Testament because he gave priceless value to its Books.
Figure 9 shows a detail of Figure 2, with the insertion of the other Gospels, whose deep language parameters are reported in Table 12. Notice that only the three synoptic Gospels (Matthew, Mark, Luke) fall within the circle of Dickens’ novels, while John is clearly further away. In other words, John does not seem to have notably influenced Dickens’ writing.
Figure 9.
Normalized coordinates and of the resulting vector (18) of the literary works listed in Table 1 and Table 2 (detail) and the canonical Gospels (Matthew, Mark, Luke, John, yellow marks), normalized so that Of Mice and Men is located at the origin (0,0) and Moby Dick is located at (1,1). Dickens’ novels are represented by the circles: The Adventures of Oliver Twist, a, blue; David Copperfield, b, red; Bleak House, d, magenta; A Tale of Two Cities, c, cyan; Our Mutual Friend, e, black. The other novels are numbered according to the order reported in Table 2.
Table 12.
Gospels statistics, King James version. Number of chapters (i.e., the number of samples considered in calculating the regression lines of the theory), total number of characters contained in the words, total number of words and sentences, deep language parameters , , , , with standard deviation reported in the second line.
Table 13 reports the slope and correlation coefficient of the regression line between the number of sentences (dependent variable) and the number of words (independent variable) per chapter, in the Gospels. Four decimal digits are reported because some values differ only from the third digit. Notice that Matthew and Luke almost coincide, as already observed in the general study reported in [] on the original Greek texts. Mark is not far, therefore distinguishing the synoptic Gospels from John.
Table 13.
Slope and correlation coefficient of the regression line between the number of sentences (dependent variable) and the number of words (independent variable) per chapter, in the canonical Gospels. Four decimal digits are reported because some values differ only from the third digit.
Table 14 summarizes the likeness index between the indicated Gospel (output) and Dickens’ novels (input). This table shows that Dickens’ novels have been likely influenced by the synoptic Gospels, especially the last three novels, which were written shortly after The Life of Our Lord.
Table 14.
Likeness index between the indicated Dickens novels and the four Canonical Gospels in the King James translation (output), sentences channel. The Gospel in the first line indicates the output text; the text in the first column indicates the input text (regression line given in Table 10). For example, in the channel Our Mutual Friend Bleak House, , while in the reverse channel, Bleak House Our Mutual Friend, .
In conclusion, we conjecture that the synoptic Gospels, read and studied by Dickens, affected and shaped, unconsciously, the deep language parameters of his writing style.
9. Final Remarks
We can now link the results shown in Figure 2 (vectors) with the likeness index . We have noticed that Dickens’ novels are concentrated within a circle, which includes very few novels by other authors.
In the Cartesian plane, we can calculate the Pythagorean distance between a literary work and a reference work, and correlate with the corresponding . Such an exercise is shown in Figure 10, as an example, where the reference (output) is Bleak House. It is clearly evident that decreases sharply as increases.
Figure 10.
Likeness index versus Pythagorean distance between a literary work and a reference (output) work (Bleak House).
For small distances (, and show a tight inverse relationship. The closest work to Bleak House is The Jungle Book (13), followed by Little Women (8), Treasure Island (9), and The Adventures of Huckleberry Finn (10). Although these novels fall within the circle, they could not have influenced Dickens’ style because they were published later than Bleak House (Table 2 and Table 3); therefore, a small distance is a necessary but not a sufficient condition for two literary works being mathematically similar.
10. Conclusions
In the first part of the article, we have recalled our general theory of linguistic channels and have studied the Shannon capacity and interdependence of these channels. In the second part, to show some features and usefulness of the theory, we have applied it to novels written by Charles Dickens and other authors of English literature, including the Gospels of the classical King James version of the Bible.
In literary works (or in any long texts), there are multiple communication channels, one of which is the channel that linearly links the sentences of two literary works for an equal number of words, as explicitly studied in the article. The theory of these channels considers not only averages, such as regression lines, but also correlation coefficients.
A Monte Carlo simulation addresses the inaccuracy in estimating the slope and correlation coefficient of regression lines due to the small sample size (i.e., given by the number of chapters of each literary work). However, besides the usefulness of the simulation as a “renormalization” tool shown in the article, there is another very likely more interesting property concerning the new generated literary works. In fact, because the mathematical theory does not consider meaning, the simulated texts might have been “written” by the author, as they maintain the main statistical properties of the deep language parameters of the original text. In other words, they are “literary works” that the author might have written at the time that he wrote the original work.
We have shown that the probability density function of the capacity of self- and cross-channels (defined and studied in the article) is a Gaussian stochastic variable. The closer the parameters of the cross-channels are to those of the self-channel, the more similar are the two literary works. The similarity is measured by the likeness index.
We have found that Dickens’ novels show striking and unexpected mathematical/statistical similarity to the synoptic Gospels. The similarity may be a trace of the influence unconsciously left in Dickens’ deep language style after researching the life of Jesus of Nazareth and writing The Life of Our Lord for his children.
We have shown that the Pythagorean distance (in a suitably defined Cartesian plane involving the deep language parameters) and the likeness index, for a reference literary work compared to all others, correlates with the corresponding likeness index with a tight inverse proportional relationship.
A similar approach can be applied, of course, to any literary corpus written in any alphabetical language, and this would allow us to compare different texts, even in translation.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The author declares no conflict of interest.
Appendix A. Statistics of Gospels of Mark, Luke, John in the King James Translation
Table A1, Table A2 and Table A3 report the statistics concerning the Gospels of Mark, Luke and John. The King James translation necessary to calculate the likeness index is reported in Table 12. All Gospels have been downloaded from https://www.biblegateway.com/versions/King-James-Version-KJV-Bible/#booklist (accessed on 4 October 2022).
Table A1.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Mark (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Table A1.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Mark (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input, cross-channels), for equal number of words, in the sentences channel.
| Novel | ||||||
|---|---|---|---|---|---|---|
| Ave | Dec | Ave | Dev | |||
| Mark (self-channel) | 24.57 | 6.74 | 0.9957 | 0.0089 | 0.9968 | 0.0292 |
| Oliver Twist | 17.09 | 3.51 | 0.9936 | 0.0075 | 1.0960 | 0.0322 |
| David Copperfield | 16.47 | 3.31 | 0.9931 | 0.0137 | 1.1130 | 0.0321 |
| Bleak House | 22.33 | 5.89 | 0.9947 | 0.0081 | 0.9816 | 0.0285 |
| A Tale of Two Cities | 22.70 | 6.60 | 0.9935 | 0.0135 | 1.0237 | 0.0298 |
| Our Mutual friend | 19.05 | 5.41 | 0.9902 | 0.0089 | 0.9888 | 0.0287 |
Table A2.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Luke (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input, cross-channels), for equal number of words, in the sentences channel.
Table A2.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of Luke (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input, cross-channels), for equal number of words, in the sentences channel.
| Novel | ||||||
|---|---|---|---|---|---|---|
| Ave | Dec | Ave | Dev | |||
| Luke (self-channel) | 26.39 | 6.21 | 0.9978 | 0.0030 | 0.9984 | 0.0245 |
| Oliver Twist | 21.07 | 3.54 | 0.9977 | 0.0033 | 1.0676 | 0.0268 |
| David Copperfield | 16.98 | 3.67 | 0.9920 | 0.0080 | 1.0838 | 0.0269 |
| Bleak House | 23.70 | 4.45 | 0.9979 | 0.0032 | 0.9562 | 0.0237 |
| A Tale of Two Cities | 20.87 | 5.75 | 0.9930 | 0.0072 | 0.9967 | 0.0246 |
| Our Mutual friend | 22.38 | 4.96 | 0.9960 | 0.0047 | 0.9615 | 0.0239 |
Table A3.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of John (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input cross-channels), for equal number of words, in the sentences channel.
Table A3.
Average value (dB) and standard deviation (dB) of (dB) in self- and cross-channels, correlation coefficient , and slope of the regression line between the number of sentences of John (channel output, self-channel) versus the number of sentences in Dickens’ novels (channel input cross-channels), for equal number of words, in the sentences channel.
| Novel | ||||||
|---|---|---|---|---|---|---|
| Ave | Dec | Ave | Dev | |||
| Luke (self-channel) | 26.88 | 6.67 | 0.9977 | 0.0040 | 0.9986 | 0.0206 |
| Oliver Twist | 12.40 | 1.08 | 0.9970 | 0.0040 | 1.2227 | 0.0257 |
| David Copperfield | 11.30 | 1.51 | 0.9936 | 0.0086 | 1.2412 | 0.0255 |
| Bleak House | 18.74 | 2.40 | 0.9975 | 0.0045 | 1.0951 | 0.0227 |
| A Tale of Two Cities | 15.12 | 2.44 | 0.9942 | 0.0083 | 1.1421 | 0.0236 |
| Our Mutual friend | 16.77 | 2.40 | 0.9946 | 0.0054 | 1.1023 | 0.0229 |
References
- Matricciani, E. Deep Language Statistics of Italian throughout Seven Centuries of Literature and Empirical Connections with Miller’s 7 ∓ 2 Law and Short-Term Memory. Open J. Stat. 2019, 9, 373–406. [Google Scholar] [CrossRef]
- Matricciani, E. A Statistical Theory of Language Translation Based on Communication Theory. Open J. Stat. 2020, 10, 936–997. [Google Scholar] [CrossRef]
- Matricciani, E. Linguistic Mathematical Relationships Saved or Lost in Translating Texts: Extension of the Statistical Theory of Translation and Its Application to the New Testament. Information 2022, 13, 20. [Google Scholar] [CrossRef]
- Matricciani, E. Multiple Communication Channels in Literary Texts. Open J. Stat. 2022, 12, 486–520. [Google Scholar] [CrossRef]
- Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
- Catford, J.C. A Linguistic Theory of Translation. An Essay in Applied Linguistics; Oxford University Press: Oxford, UK, 1965. [Google Scholar]
- Munday, J. Introducing Translation Studies. Theories and Applications, 2nd ed.; Routledge: London, UK, 2008. [Google Scholar]
- Proshina, Z. Theory of Translation, 3rd ed.; Far Eastern University Press: Manila, Philippines, 2008. [Google Scholar]
- Trosberg, A. Discourse analysis as part of translator training. Curr. Issues Lang. Soc. 2000, 7, 185–228. [Google Scholar] [CrossRef]
- Tymoczko, M. Translation in a Post—Colonial Context: Early Irish Literature in English Translation; St. Jerome Publishing: Manchester, UK, 1999. [Google Scholar]
- Warren, R. (Ed.) The Art of Translation: Voices from the Field; North-Eastern University Press: Boston, MA, USA, 1989. [Google Scholar]
- Williams, I. A corpus-based study of the verb observar in English-Spanish translations of biomedical research articles. Target 2007, 19, 85–103. [Google Scholar] [CrossRef]
- Wilss, W. Knowledge and Skills in Translator Behaviour; John Benjamins: Amsterdam, The Netherlands; Philadelphia, PA, USA, 1996. [Google Scholar]
- Wolf, M.; Fukari, A. (Eds.) Constructing a Sociology of Translation; John Benjamins: Amsterdam, The Netherlands; Philadelphia, PA, USA, 2007. [Google Scholar]
- Gamallo, P.; Pichel, J.R.; Alegria, I. Measuring Language Distance of Isolated European Languages. Information 2020, 11, 181. [Google Scholar] [CrossRef]
- Barbançon, F.; Evans, S.; Nakhleh, L.; Ringe, D.; Warnow, T. An experimental study comparing linguistic phylogenetic reconstruction methods. Diachronica 2013, 30, 143–170. [Google Scholar] [CrossRef]
- Bakker, D.; Muller, A.; Velupillai, V.; Wichmann, S.; Brown, C.H.; Brown, P.; Egorov, D.; Mailhammer, R.; Grant, A.; Holman, E.W. Adding typology to lexicostatistics: Acombined approach to language classification. Linguist. Typol. 2009, 13, 169–181. [Google Scholar] [CrossRef]
- Petroni, F.; Serva, M. Measures of lexical distance between languages. Phys. A Stat. Mech. Appl. 2010, 389, 2280–2283. [Google Scholar] [CrossRef]
- Carling, G.; Larsson, F.; Cathcart, C.; Johansson, N.; Holmer, A.; Round, E.; Verhoeven, R. Diachronic Atlas of Comparative Linguistics (DiACL)—A database for ancient language typology. PLoS ONE 2018, 13, e0205313. [Google Scholar] [CrossRef]
- Gao, Y.; Liang, W.; Shi, Y.; Huang, Q. Comparison of directed and weighted co-occurrence networks of six languages. Phys. A Stat. Mech. Appl. 2014, 393, 579–589. [Google Scholar] [CrossRef]
- Liu, H.; Cong, J. Language clustering with word co-occurrence networks based on parallel texts. Chin. Sci. Bull. 2013, 58, 1139–1144. [Google Scholar] [CrossRef]
- Gamallo, P.; Pichel, J.R.; Alegria, I. From Language Identification to Language Distance. Phys. A 2017, 484, 162–172. [Google Scholar] [CrossRef]
- Pichel, J.R.; Gamallo, P.; Alegria, I. Measuring diachronic language distance using perplexity: Application to English, Portuguese, and Spanish. Nat. Lang. Eng. 2019, 26, 433–454. [Google Scholar] [CrossRef]
- Eder, M. Visualization in stylometry: Cluster analysis using networks. Digit. Scholarsh. Humanit. 2015, 32, 50–64. [Google Scholar] [CrossRef]
- Brown, P.F.; Cocke, J.; Pietra, A.D.; Pietra, V.J.D.; Jelinek, F.; Lafferty, J.D.; Mercer, R.L.; Roossin, P.S. A Statistical Approach to Machine Translation. Comput. Linguist. 1990, 16, 79–85. [Google Scholar]
- Koehn, F.; Och, F.J.; Marcu, D. Statistical Phrase-Based Translation. In Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology (HLT-NAACL 2003), Edmonton, AB, Canada, 27 May–1 June 2003; pp. 48–54. [Google Scholar]
- Carl, M.M.; Schaeffer, M. Sketch of a Noisy Channel Model for the Translation Process. In Empirical Modelling of Translation and Interpreting; Hansen-Schirra, S., Czulo, O., Hofmann, S., Eds.; Language Science Press: Berlin, Germany, 2017; pp. 71–116. [Google Scholar] [CrossRef]
- Elmakias, I.; Vilenchik, D. An Oblivious Approach to Machine Translation Quality Estimation. Mathematics 2021, 9, 2090. [Google Scholar] [CrossRef]
- Lavie, A.; Agarwal, A. Meteor: An Automatic Metric for MT Evaluation with High Levels of Correlation with Human Judgments. In Proceedings of the Second Workshop on Statistical Machine Translation, Prague, Czech Republic, 23 June 2007; pp. 228–231. [Google Scholar]
- Banchs, R.; Li, H. AM–FM: A Semantic Framework for Translation Quality Assessment. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; Volume 2, pp. 153–158. [Google Scholar]
- Forcada, M.; Ginestí-Rosell, M.; Nordfalk, J.; O’Regan, J.; Ortiz-Rojas, S.; Pérez-Ortiz, J.; Sánchez-Martínez, F.; Ramírez-Sánchez, G.; Tyers, F. Apertium: A free/open-source platform for rule-based machine translation. Mach. Transl. 2011, 25, 127–144. [Google Scholar] [CrossRef]
- Buck, C. Black Box Features for the WMT 2012 Quality Estimation Shared Task. In Proceedings of the 7th Workshop on Statistical Machine Translation, Montreal, QC, Canada, 7–8 June 2012; pp. 91–95. [Google Scholar]
- Assaf, D.; Newman, Y.; Choen, Y.; Argamon, S.; Howard, N.; Last, M.; Frieder, O.; Koppel, M. Why “Dark Thoughts” aren’t really Dark: A Novel Algorithm for Metaphor Identification. In Proceedings of the 2013 IEEE Symposium on Computational Intelligence, Cognitive Algorithms, Mind, and Brain, Singapore, 16–19 April 2013; pp. 60–65. [Google Scholar]
- Graham, Y. Improving Evaluation of Machine Translation Quality Estimation. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; pp. 1804–1813. [Google Scholar]
- Espla-Gomis, M.; Sanchez-Martınez, F.; Forcada, M.L. UAlacant Word-Level Machine Translation Quality Estimation System at WMT 2015. In Proceedings of the Tenth Workshop on Statistical Machine Translation, Lisbon, Portugal, 17–18 September 2015; pp. 309–315. [Google Scholar]
- Costa-Jussà, M.R.; Fonollosa, J.A. Latest trends in hybrid machine translation and its applications. Comput. Speech Lang. 2015, 32, 3–10. [Google Scholar] [CrossRef]
- Kreutzer, J.; Schamoni, S.; Riezler, S. QUality Estimation from ScraTCH (QUETCH): Deep Learning for Word-Level Translation Quality Estimation. In Proceedings of the Tenth Workshop on Statistical Machine Translation, Lisbon, Portugal, 17–18 September 2015; pp. 316–322. [Google Scholar]
- Specia, L.; Paetzold, G.; Scarton, C. Multi-Level Translation Quality Prediction with QuEst++. In Proceedings of the ACL–IJCNLP 2015 System Demonstrations, Beijing, China, 26–31 July 2015; pp. 115–120. [Google Scholar]
- Banchs, R.E.; D’Haro, L.F.; Li, H. Adequacy-Fluency Metrics: Evaluating MT in the Continuous Space Model Framework. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 472–482. [Google Scholar] [CrossRef]
- Martins, A.F.T.; Junczys-Dowmunt, M.; Kepler, F.N.; Astudillo, R.; Hokamp, C.; Grundkiewicz, R. Pushing the Limits of Quality Estimation. Trans. Assoc. Comput. Linguist. 2017, 5, 205–218. [Google Scholar] [CrossRef]
- Kim, H.; Jung, H.Y.; Kwon, H.; Lee, J.H.; Na, S.H. Predictor-Estimator: Neural Quality Estimation Based on Target Word Prediction for Machine Translation. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 2018, 17, 1–22. [Google Scholar] [CrossRef]
- Kepler, F.; Trénous, J.; Treviso, M.; Vera, M.; Martins, A.F.T. OpenKiwi: An Open Source Framework for Quality Estimation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Florence, Italy, 28 July–2 August 2019; pp. 117–122. [Google Scholar]
- D’Haro, L.; Banchs, R.; Hori, C.; Li, H. Automatic Evaluation of End-to-End Dialog Systems with Adequacy-Fluency Metrics. Comput. Speech Lang. 2018, 55, 200–215. [Google Scholar] [CrossRef]
- Yankovskaya, E.; Tättar, A.; Fishel, M. Quality Estimation with Force-Decoded Attention and Cross-Lingual Embeddings. In Proceedings of the Third Conference on Machine Translation: Shared Task Papers, Belgium, Brussels, 31 October–1 November 2018; pp. 816–821. [Google Scholar]
- Yankovskaya, E.; Tättar, A.; Fishel, M. Quality Estimation and Translation Metrics via Pre-Trained Word and Sentence Embeddings. In Proceedings of the Fourth Conference on Machine Translation, Florence, Italy, 1–2 August 2019; Volume 3, pp. 101–105. [Google Scholar]
- Papoulis, A. Probability & Statistics; Prentice Hall: Hoboken, NJ, USA, 1990. [Google Scholar]
- Miller, G.A. The Magical Number Seven, Plus or Minus Two. Some Limits on Our Capacity for Processing Information. Psychol. Rev. 1955, 2, 343–352. [Google Scholar]
- Dickens, C. The Life of Our Lord; Simon & Schuster: New York, NY, USA, 1934. [Google Scholar]
- Walder, D. Dickens and Religion; George Allen & Unwin: London, UK, 1981. [Google Scholar]
- Hanna, R.C. The Dickens Family Gospel: A Family Devotional Guide Based on the Christian Teachings of Charles Dickens; Rainbow Publishers: Kochi, India, 1999. [Google Scholar]
- Hanna, R.C. The Dickens Christian Reader: A Collection of New Testament Teachings and Biblical References from the Works of Charles Dickens; AMS Press: New York, NY, USA, 2000. [Google Scholar]
- Matricciani, E.; Caro, L.D. A Deep-Language Mathematical Analysis of Gospels, Acts and Revelation. Religions 2019, 10, 257. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).