Next Article in Journal
Behavioral Propagation Based on Passionate Psychology on Single Networks with Limited Contact
Previous Article in Journal
Complexity and Entropy in Physiological Signals (CEPS): Resonance Breathing Rate Assessed Using Measures of Fractal Dimension, Heart Rate Asymmetry and Permutation Entropy
Previous Article in Special Issue
The Interpretation of Graphical Information in Word Processing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Entropy of Digital Texts—The Mathematical Background of Correctness

Faculty of Informatics, University of Debrecen, Kassai út 26., 4028 Debrecen, Hungary
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(2), 302; https://doi.org/10.3390/e25020302
Submission received: 31 December 2022 / Revised: 29 January 2023 / Accepted: 31 January 2023 / Published: 6 February 2023

Abstract

:
Based on Shannon’s communication theory, in the present paper, we provide the theoretical background to finding an objective measurement—the text-entropy—that can describe the quality of digital natural language documents handled with word processors. The text-entropy can be calculated from the formatting, correction, and modification entropy, and based on these values, we are able to tell how correct or how erroneous digital text-based documents are. To present how the theory can be applied to real-world texts, for the present study, three erroneous MS Word documents were selected. With these examples, we can demonstrate how to build their correcting, formatting, and modification algorithms, to calculate the time spent on modification and the entropy of the completed tasks, in both the original erroneous and the corrected documents. In general, it was found that using and modifying properly edited and formatted digital texts requires less or an equal number of knowledge-items. In information theory, it means that less data must be put on the communication channel than in the case of erroneous documents. The analysis also revealed that in the corrected documents not only the quantity of the data is less, but the quality of the data (knowledge pieces) is higher. As the consequence of these two findings, it is proven that the modification time of erroneous documents is severalfold of the correct ones, even in the case of minimal first level actions. It is also proven that to avoid the repetition of the time- and resource-consuming actions, we must correct the documents before their modification.

1. Introduction

Millions of digital texts have been created since word processors appeared on the market, and handling these documents became one of the most popular digital production-activity. However, Johnson [1] called attention to his findings as early as late the 90s that:
“…it is believed that a very little knowledge, a skimpy overview, is sufficient. The value of limited study of word processing applications is rather doubtful. In fact, a little bit of knowledge about word processing may be almost useless—and a quick overview is certainly not sufficient to realize most of the overwhelming benefits of using computers for writing.”
He went even further, claiming that:
“In order to take advantage of the powers of word processing requires considerable skill in its use. Those who understand only a little about word processors will probably employ them in a linear fashion like an expensive typewriter and compose in exactly the same way they would on a typewriter because they simply do not know how to use the sophisticated editing features of a word processor.”
What surprising is that the quality of digital texts has not improved considerably during the past years, and the mentioned misconceptions regenerate and intensify the problems. Johnson [1] only mentions the complexity of word processors and users’ lack of knowledge; however, we also must pay attention to the content and the structure of the texts, which is more complex and demanding than the technical details of using word processors. It is said that building football stadiums does not make football, having a telescope does not make astronomy. In a similar way, owning hardware and software does not make informatics, or computer sciences, which is clearly expressed both in science [2] and industry [3,4].
“When major new machinery comes along—as computers have—it’s rather disorientating.”
[2]
“Relying on technical innovation alone often provides only temporary competitive advantage.”
[3]
“Society has reached the point where one can push a button and be immediately deluged with technical and managerial information. This all very convenient, of course, but in one is not careful there is a danger of losing the ability to think. We must remember that in the end it’s the individual human being who must solve the problems.”
[4]
Consequently, the quality of word-processed documents is questionable due to the circulating misconceptions and the ill-use of these rather complex applications [5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20], which can be proven by thorough analyses of the available documents and rather inefficient modification processes. More and more word-processed documents are publicly available on the internet and circulate in closed networks. On one hand, the erroneous documents encourage untrained end-users to copy and follow these examples, and they use them as references. On the other hand, it allows researchers to build their corpora for various error detection studies on the subject. Considering the available word processors and their popularity at the time of writing the present paper, we selected the most widely used program and its documents for further analysis, which was Microsoft Word and documents with DOC and DOCX extensions [21]. One further reason for choosing Microsoft Word was that it allows users to seamlessly present non-printing characters on the interface which—despite its credible popularity—is crucial in the editing process of digital texts [22,23,24,25].
The question is whether self-taught [26], overconfident [27,28,29], tool-centered end-users (non-professional authors and/or editors of documents) [30,31,32,33,34,35,36,37,38,39,40,41,42,43], or those who raise doubts about the quality of word-processed texts have merit [5,6,7,8,9,10,11,12,13,14,15,16,17,18,19]. To find proof, instead of relying on legends and folktales [44,45], we must find facts by using the results of educational research connected to the subject [44,45,46,47,48,49,50,51,52,53]. Being aware of the low efficiency of word processing activities, we launched a research project to find an objective measuring system and define the entropy of erroneous and correct digital texts, what we call text-entropy.
Our previous research in the field of didactics of informatics has already revealed that in general, education performs a crucial role in developing students’ computational thinking skills [54], especially in handling digital texts [5,6,7,8,9,10,11,12,13,14,15,16,17,18,19]. It is also found that the cognitive load of handling digital text-based documents in a “user-friendly” GUI is so high [55] that novel methods and approaches are required to be able to cope with this complex problem.
Considering the role of education, the complexity of digital texts, the algorithm-driven, deceptive “user-friendly” word processing programs, and the millions of erroneous documents circulating either in closed communities (corporate, school, training group, etc.) or on the internet available for everyone, it is found that we are in great need of approaches and methods that can handle errors. Surprisingly, educational research has paid little attention to learning from errors [56], with a few exceptions [57,58,59]. Primarily sports, where ignoring errors would lead to serious injuries and underperformance in competitions [60,61,62,63,64]. Not less surprising is that the profit-oriented “real-world” and their production systems also pay minimal attention to errors. However, the Toyota Production System claims that they must bring problems to the surface, make them visible, and go to work immediately on countermeasures [4]. One of the principles of the Toyota Way is that:
“We view errors as opportunities for learning. Rather than blaming individuals, the organization takes corrective actions and distributes knowledge about each experience broadly. Learning is a continuous company-wide process as superiors motivate and train subordinates; as predecessors do the same for successors; and as team members at all levels share knowledge with one another.”
However, most of these principles, and educational and/or training research consider students’, employees’, athletes’, etc., errors committed by themselves which must be corrected as soon as possible to avoid further damages.
Based on this concept, but using already existing erroneous documents and teaching-learning materials as examples, the Error Recognition Model (ERM) was introduced [7,11,12]. We use the very same model to define the entropy of digital texts (text entropy) and decide whether a text is correct or not.

1.1. Error Recognition Model

The Error Recognition Model (ERM) [7,11,12] would be a solution for increasing the effectiveness of digital text management, since it has been proven more efficient and effective than the widely accepted tool-centered approaches, and the circulating misconceptions. Both are primarily rooted in the published teaching-learning materials, or rather tutorials, the disadvantages of tool-centered computer courses [30,31,32,33,34,35,36,37,38,39,40,41,42,43], and the believes of overconfident but ignorant end-users [27,28,29].
“… the significant advantages of word processing are available exclusively to those who are proficient in the use of the hardware and software; they will be inaccessible to those who have only a little understanding of word processing. A stand-alone computer skills course (taught by a school or by a computer dealer) may not be the best means to teach substantial knowledge of word processing; examples and practice will inevitably be simulated and artificial, and there will be little motivation to fully understand the applications.”
[1]
It is proven that applying ERM [12] in the teaching-learning process has advantages which the widely accepted tool-centered, low-mathability approaches miss [2,65,66,67,68,69,70]. These advantages are due to “real-world” texts used as examples and the consideration of cognitive load at each step of method. The erroneous “real-world” texts perform a crucial role in motivation, where students claim proudly that they can do better. By keeping the cognitive load at bay [44,71,72,73,74], we can make countermeasures immediately and develop students’ computational thinking skills step-by-step. Consequently, we can avoid unrepairable damage when students and end-users create, edit, and modify their own or somebody else’s documents.

1.1.1. Properly Formatted Text

The ERM accepts the definition of the properly formatted text [7,11], and, based on it, builds up a teaching-learning approach that focuses on the text, instead of the word processing tools and interfaces.
The definition of the properly formatted text has two constrains which are the following:
  • The text fulfils the requirements of printed documents (quantitative requirements, and errors detailed in Section 1.1.2).
  • The content is indifferent to modification (qualitative requirements, and errors detailed in Section 1.1.2)—the document is editable, but applied changes must be limited to the correct form of those actions that the user originally intended.

1.1.2. Error Categories

The focuses of ERM are the recognition of errors in digital texts (e.g., finding, being aware of, avoiding, and correcting them), learning how to avoid them, and transferring this knowledge and skills to other types of digital texts (e.g., from word processors to presentations and web pages, and the other way around). ERM teaches that we must be aware of errors, we must learn how to handle and avoid them to improve the quality of our works [3,4,56,57,58,59]. One of the primary principles of ERM is that errors are user-friendly tools in handling digital contents, as they are in the real world [3,4,56,57,58,59].
In accordance with the definition of the properly formatted text and due to the extremely high number of errors in digital texts, they must be categorized. Considering the terminology of both natural and artificial languages in connection with computers, errors are listed in two hypernym, then three and three hyponym categories [11].
  • Quantitative error categories (recognizable in any printed or electronically displayed form, can be seen by the reader or viewer of the document):
    a.
    syntactic (the grammar of the text);
    b.
    semantic (the content of the text);
    c.
    typographic (the appearance of the text) [75,76,77,78].
  • Qualitative error categories (recognizable in editable digital form only; therefore, invisible to the target audience, and seen exclusively by the author(s) and other participants handling the original or the modified file):
    a.
    layout (the arrangement of characters and objects of the text);
    b.
    formatting (applying formatting commands);
    c.
    style (handling—applying, defining, reformatting—styles).
It is not rare that one error category triggers another or a third. In these cases, primary and secondary categories are applied to the errors, depending on which triggers the other. For example, in Figure 1 Line 19 has a manual hyphenation which is a layout error. However, the manual hyphenation (with the Enter and Space characters in the middle of the word ‘parental’) creates syntactical and/or semantical errors (with the separate words ‘pa’ and ‘rental’).

1.1.3. The Application of ERM

The application of ERM in an educational environment consists of three major steps depending on the tools used by students/teachers can be unplugged (UP), semi-unplugged (SUP), and digital (D) [12].
  • The first step is error recognition in two clearly distinguishable phases:
    a.
    The marking, explaining (writing), and later categorizing the quantitative errors of the text in its printed form (paper, picture, or PDF). In this phase, the only source is the printed text (UP). This is the form of the document which readers and viewers see.
    b.
    The marking, explaining (writing), and later categorizing the qualitative errors of the text in its editable (digital) form. In this phase, the primary source is the editable, opened document (if available with the turned-on Show/Hide button to clearly present the non-printing characters [22,23,24,25]). The secondary source is the printed document where notetaking takes place. To distinguish the two types of error hypernyms, different colors are used to mark the errors. The coloring system would also serve students in practicing, rehearsing, and catching up (SUP).
  • The second step is correcting the errors of the document. In this phase, a correction algorithm is set up whose primary concept is to start with the errors of the widest range (domain), and finish with the least significant. The algorithm also considers correction with the Replace… command or with macros when repeated errors must be corrected (e.g., multiple empty paragraph marks, tabulator and/or space characters, repeated formatting) (D).
  • The third phase is the formatting of the text which, similar to correction, is based on an algorithm. The steps of the algorithm are primarily decided focusing on the range of the planned formatting commands, starting from those of the widest range to the least significant (D).
We must note at this point that there are cases where some steps of the third phase (formatting) must be conducted in the second phase (e.g., modifying page setup, setting Normal style).
The steps of ERM reveal that the focus is on the text [12]. On one hand, the model supports the readers by making the text as legible and understandable as possible. On the other hand, it helps in avoiding and correcting qualitative errors, and making the text ready for effective modification and formatting restricted only to the intention of the author/editor. In this scenario, the text-editing application, the use of the GUI performs only a secondary role, meanwhile building proficiency in the ERM text-handling process.
Unlike traditional course books, teaching-learning materials, and tutorials [30,31,32,33,34,35,36,37,38,39,40,42,43], this concept does not start teaching text management through:
  • operating system instructions (this knowledge should have been brought in through knowledge-transfer, and if not, then it should be introduced and practiced somewhere else) [30,32];
  • toying with font-face appearances [33,34,35,36];
  • new features of the application [30];
  • the introduction of the GUI objects [30];
  • typing text.
Instead of these, in ERM real world texts with real contents are provided during the teaching-learning process, which idea is in complete accordance with the main concept of:
  • Informatics Reference Framework [79,80];
  • TPCK (Technological Pedagogical Content Knowledge) [81,82,83];
  • Polya’s concept-based problem solving approach [65];
  • Kahneman’s thinking fast and slow theory [71] along with the cognitive load theory [72];
  • Hattie’s results on teaching effectively [73];
  • Chen’s Meaning System Model [66].
Considering the combination of these theories, the teaching-learning materials can be adjusted to the students’ (in general, and end-users) age, background knowledge, interests, school subjects, work requirements, etc.

1.2. Detecting End-User Activities—ANLITA

To record what actions end-users perform in digital texts, our research group developed a dedicated logging application, named ANLITA (Atomic Natural Language Input Tracking Application). In the present research, ANLITA is primarily used to log all the keyboard actions and to screen-record the complete editing process conducted in Word documents. The program outputs two files:
  • a text file containing all the keyboard events;
  • a video file, in which the entire text handling process was recorded.
Using these two files, the complete text handling process can be retraced, the techniques and algorithms (if there was any) applied to the text can be revealed. Based on this data, an objective measurement system can be set up which might help us to provide information on the quality of the text, more precisely how efficiently the message of the natural language digital text is communicated.

2. Sample

For the present study, three documents of different content and length were selected, each carrying both quantitative and qualitative errors:
  • A single page proposition of a Grade 7 student voicing requests of their class for better conditions in the school (for short: medicine) (Figure 1). This document includes text-content only. (Line numbering shown in Figure 1 was added for referencing purposes in the present paper.)
  • A teacher’s three-page long test paper (Figure 2). This document includes a combination of text-content, pictures, other graphical objects, and two table-imitations (for short: frenchfood).
  • An equation cut from a longer document created by a senior pre-service teacher of mathematics and library information (for short: equation) (Figure 3).
These documents are from our private collection, gathered from students and colleagues rendered unidentifiable.
In general, our study checked for the flexibility of these documents, e.g., how efficiently the intended modifications can be applied to them. Consequently, our primarily concerns were the qualitative errors [7,11]. To identify those, the visibility of non-printing characters perform a crucial role in the analysis, so all recordings and figures were created with the Show/Hide button turned-on [22,23,24,25] and some with the text boundaries of the document sections made visible.

2.1. Document: medicine

The medicine document contains several layout errors in the form of multiple Space and Enter characters and Enter characters at the end of the lines (Lines 4, 5, 6, 12, 18, and 19). In general, these characters were meant to create left indentation and vertical spacing between paragraphs, and right alignment in Line 30. We could not explain the multiple Space characters in Line 34. There is a formatting and style error in Line 12 which is unreasonable (style Heading 1 with the Keep with next paragraph formatting, reformatted to match the appearance of the main body of the text). Line numbers were not part of the original document, we turned those on to enable us referring to the location of the errors in a more convenient and reliable way.

2.2. Document: frenchfood

The original frenchfood document is a three-page long text created by a teacher with testing purposes. The length of the document depends on the settings of the Normal style, but it was meant to be three pages long. The first page (Section 1) has Portrait, while the second and third pages Landscape orientation (Section 2) set up.
The document is burdened:
  • with layout errors in the form of multiple Tabulator, Space, and Enter characters,
  • a mixture of layout and formatting errors on the graphical objects (lines and textboxes both in the body and the fake headers),
  • style and formatting errors on fonts and paragraphs.

2.3. Document: equation

The content of the equation document is a portion of a senior pre-service teacher’s lesson plan. Figure 3 clearly presents the document creator’s attempted manipulation of the text (Ben-Ari [5] calls this bricolage). Instead of inserting proper math formulae into the document using the readily available built-in equation editor, the author fabricated various text-positioning tricks to construct these formula-lookalikes. The errors in this piece of text are the following.
  • The first formula starts with a manual numbering. (Later in the analysis, the manual numbering was ignored since it does not perform any role in the editing of the equations.)
  • Numerators and denominators are typed in separate paragraphs.
  • Vincula are imitated by applying underline character formatting on the numerators.
  • Minus signs are substituted with underscore characters.
  • Plus signs and whole number are received a subscript character formatting.
  • Equal signs are imitated by blanks (Space characters) with double-underline formatting applied.
  • Space characters are used to horizontally position numerators and denominators.
This document has one further characteristic we must mention; the overconfident author blamed MS Word not allowing equation sign at end the of the lines (marked by a footnote in Formula 1, and by the imitation of footnotes in Formulas 2 and 3) (Figure 3) [27,28,29].

3. Methods

On the selected documents, a rigorous analysis was conducted to reveal their structures and to detect their errors. This is primarily an unplugged and/or semi-unplugged process according to ERM [12].
Following the steps of ERM, in the second phase, the errors of the original documents were corrected, then the proper formatting took place. Finally, both to the original and the corrected-formatted documents, modifications were applied to measure and compare the entropy of the documents.
We must note that all these steps can be repeated by anyone interested. However, the steps detailed in the present paper were conducted by two of the authors. They both are professionals in text management, which helped them to set up a normalized algorithm to each text and to minimize the time assigned to the algorithm steps. Consequently, less experienced end-users repeating all these steps might provide different time and probability. However, these differences not necessarily deteriorate the entropy as the limit to how effectively one can handle a digital text, to how effectively one can communicate the outcome of a text handling process (Section 6).

3.1. Correction of the Documents

Considering the content, each piece of text is unique. However, the analyses revealed that the primary concern our analysis is the improper use of non-printing characters. Since we were not present at the creation of the documents, we do not know whether the authors had these turned on or off [22,23,24,25]. Regardless of the authors’ awareness of the non-printing characters and our lack of this information, all the processes applied to the texts during the study are conducted and recorded with the turned-on Show/Hide button. We can use this method to disambiguate whether characters or formatting are used to set up an appearance, which perform crucial role both in the analyses and the documentation of the analyses, and later in the dissemination of the findings.
A correction-algorithm was set up for to the medicine and frenchfood documents, along which their correction was executed. Beyond correcting the errors of a document, the algorithm serves our research to set up a lower bound on the average size (number of instructions and time) of each modification that end-users can apply to a text. With this method, we can define correction-entropy and the number of bits a text requires to be corrected.
These steps of the algorithms are presented in in Section 5, along with the time spent on each step.

3.2. Formatting of the Documents

Following the correction of the documents, all the three texts were properly formatted which includes the modification of the Normal style. Furthermore, where it was necessary, minor adjustments were applied. This happened in the case of the frenchfood document where some of the pictures on the second page must be resized to fit the content of the three rows on one page.
In a similar way to correction, a lower bound on the average size of each formatting is set up. The steps of the algorithms are presented in in Section 5, along with the time spent on each step.
The aims of the correction and formatting processes were to set up the entropy of the documents and find an objective measuring system which can distinguish erroneous texts from their properly formatted counterpart.

3.3. Correction, Formatting, and Statistics

In the following, the methods of the correction and formatting of the medicine and the frenchfood documents are detailed. Beyond the algorithms of these processes, the statistics and other graphical tools provided by the software (GUI) are added for giving more explanatory details.

3.3.1. Document: medicine

The corrected medicine document consists of 8 paragraphs (Figure 4, right) compared to the 34 paragraphs in the original (Figure 4, left).
The text boundaries (Figure 4) reveal that MS Word is able to recognize each individual paragraph of the document, including the empty ones. However, the document statistics calculated and shown by the application are manipulated (Figure 5) by leaving out the empty paragraphs, despite that they perform a crucial role in the text-handling processes. According to the statistics (Figure 5), the original medicine document has 14 paragraphs, because MS Word counts those paragraphs only where at least one alphabetical/numerical/special printable character content is present.
In the case of the medicine document, the aim of the correction was to clear all the unnecessary and incorrect characters and formats of the document. While in the formatting phase, we wanted to apply proper formatting steps to set up a correct document as similar in appearance to the original one as possible.

3.3.2. Document: frenchfood

The comparison of the boundaries of the erroneous (Figure 6) and the corrected document (Figure 7), and the statistics of the original and the correct frenchfood documents (Figure 8) reveals that the empty paragraphs are not counted in the statistics. Furthermore, it can be concluded that only those paragraphs are counted which have printable characters, and, in that respect, neither figures nor drawn shapes count as paragraph-content.
According to the MS Word statistics (Figure 8), in the original document there are 15 paragraphs: 10 + 1 on Page 1 (P1), and 3 + 1 on Page 2 (P2). On P1, the paragraph of the orange does not count (there is no printable character in this paragraph), while both of the text boxes do (one on P1 and the other on P2). In the corrected document, there are 10 paragraphs on P1 and 3 on P2. However, the text boundaries reveal that the original frenchfood document is loaded with uncounted empty paragraphs, and there are two unnecessary text boxes.
Figure 9 presents the corrected and formatted frenchfood document. In this document, the first paragraph of P1 is moved to the header, a table was created on each page, and the pictures and food names were put side by side on P1, and names were arranged below the pictures on P2. Furthermore, on P2 a line (bottom border with paragraph domain) was added to each food name to mimic the appearance seen on the original document.
In the correction phase of the analysis, the aim was to create two tables holding all the pictures and food names. On P1, we kept the original order of the pictures and names, while on P2, the pictures and names got rearranged into matching pairs.

3.4. Modification of the Documents

Beyond calculating correction- and formatting-entropy of the documents, we also wanted to calculate the modification-entropy. In this context, modification means that either content is added to the document, or a formatting is performed.
The modification of the documents was planned and executed in accordance with the content and the characteristics of the texts. Considering these criteria, the following modifications were applied to both the original and the corrected documents (Table 1).

3.4.1. Document: medicine

In the medicine documents two modifications were applied (both to the original and the corrected). In the first analysis, two words were added to the second paragraph (as counted in the correct document) (Figure 10), while in the second analysis, the font size of the paragraph was changed (Figure 11).

3.4.2. Document: frenchfood

In the original frenchfood document on P2, ten figures were arranged into three rows. The first two rows consist of four and four pictures, while the third row contained only two pictures. To fill in this gap, two pictures and two food names were added to the content of P2 (Figure 12).
The second modification to the frenchfood document was the changing of the font size. Figure 13 presents the results of this modification both in the original (left) and in the corrected (right) documents. While in the original document further amendments were required to match the pictures and the food names, in the corrected document no additional steps were required.

3.4.3. Document: equation

Introducing the equation editor in word processors is considered difficult, thus it is seldom taught in schools and trainings. In general, only advanced or special classes deal with the subject. However, to measure the difficulty and the complexity of creating equations, we launched several subjective and an objective test. In the subjective tests, the equation editor was introduced in a Grade 7 math class. In the subjective tests, the equation editor was introduced in a Grade 7 math class. Before we introduced the equation editor, these students had already used touch screen displays and their fingers to enter equations. During the testing period where the equation editor was introduced and used, both the activities and the results of the students, along with the time spent on the computer were observed. Furthermore, the eligibility of the formulae were considered.
To find objective measures, we first applied our planned modifications to the formula-lookalikes in the erroneous document following the jiggery-pokery method of the original author, then to the properly constructed equations. (Figure 3).
The modifications performed in this objective measuring process were:
  • calculating the common denominator of the fractions (Figure 14),
  • adding a new fraction to the equation (Figure 15),
  • solving the substitution problem by simplifying the equation (Figure 16).
The final step of the test was to measure the entropy of both solutions, to see how much data is required to put on the communication channel to handle equations (formulas).

4. Setting Up the Theory

4.1. Connection to Communication Entropy

In Shannon’s original paper [84,85], the probabilities of events were calculated primarily from their relative frequency (Equation (3)). In our case, we were obliged to find other sources to calculate the probabilities. We found that the duration of events can substitute the frequency and serve as the basis for calculating text-entropy.
The correction, formatting, and modification phases of the documents are digital processes, and, as such, can be logged with ANLITA. As it was mentioned earlier (Section 1.2) ANLITA produces two files for each monitored session (what we call a modification task):
  • a text file logging the keyboard activities and their timestamp,
  • a video file with the entire modification process recorded.
The recorded files can be used to calculate the time spent on the events (atomic steps). The sum of these event-times provides the time spent on the modification task (Equation (1)). Using these durations and the total time, we are able to calculate the probability of a modification task (Equations (1) and (2)) [84,85,86,87,88,89,90].

4.2. Calculating the Entropy of Digital Texts—Text-Entropy

The text-entropy can originate from three different type of tasks:
  • formatting,
  • correction (in erroneous documents),
  • modification.
To all types, the same method is applied to calculate their text-entropy. Based on the algorithm of the task, events (atomic steps) are set up, and to each event, a time is assigned (tk, recorded in the logging process and measured in seconds with two decimals). The sum of these time values provides the time of the task (t) (Equation (1)).
t = t 1 + + t n = k = 1 n t k
Based on the duration of the events (tk) and duration of the task (t), the probability of each event is calculated (pk) (Equation (2)). The sum of these probabilities is 1 (Equation (3)). Calculating the probability of the events from the time assigned to them, allows us to distinguish text-entropy from information-entropy, while still holding to a common theoretical background.
p k = t k t
p 1 + + p n = k = 1 n p k = 1
From these probabilities (pk) the information content of the events (Ik) were calculated (Equation (4)). (The choice of a logarithmic base is explained in Section 4.4)
I k p = - l o g 2 p k
The final step of calculating text-entropy is presented in Equation (5). The sum of the products of probabilities and information contents allows us to quantify the entropy of a task (E). In other words, we are able to tell how many bits of data are required to be put on the communication channel to conduct the planned task (Equation (5)).
E Χ = k = 1 n p k · I k = - k = 1 n p k · log 2 p k = k = 1 n E k
To provide text-entropy as an integer, one further step is applied, where E is rounded up (Eb) (Equation (6)).
E b = E Χ + 1

4.3. The Characteristics of Data in Text-Entropy

One further question is what kind of data can be considered when text management is discussed. Shannon, in his original schematic diagram of a general communication system, called the data being put on the communication channel “information source”. In our case, the situation is the same. To handle digital texts, end-users must have sources, which enable them to conduct the planned actions. Calculating the entropy of an action does not specify these sources. However, our knowledge connected to text management can identify two sources with certainty. These are education, and the user interface of the word processor which is recently a graphical interface (GUI). Further sources can be considered, but they are beyond the scope of the present paper.
Considering the possible sources of information (the data being put on the communication channel), we must keep in mind that the analyses presented in the paper were conducted by the professionals of our research group. In the case of general end-users, measuring the entropy of the selected modification tasks might require further steps, items, and concerns, which might increase the number of bits to be put on the channel (further discussed in Section 6).

4.4. The Choice of a Logarithmic Base

According to Shannon [84,85], any base can be selected to describe the entropy. Since in computer sciences base 2 is the most widely accepted, we decided to use the same for calculating the amount of data required to complete a modification task.
“If the number of messages in the set is finite then this number or any monotonic function of this number can be regarded as a measure of the information produced when one message is chosen from the set, all choices being equally likely. As was pointed out by Hartley the most natural choice is the logarithmic function. Although this definition must be generalized considerably when we consider the influence of the statistics of the message and when we have a continuous range of messages, we will in all cases use an essentially logarithmic measure.”
“The choice of a logarithmic base corresponds to the choice of a unit for measuring information. If the base 2 is used the resulting units may be called binary digits, or more briefly bits, a word suggested by J.W. Tukey. A device with two stable positions, such as a relay or a flip-flop circuit, can store one bit of information. N such devices can store N bits, since the total number of possible states is 2N and log22N = N. If the base 10 is used the units may be called decimal digits.”

5. Results

To calculate the entropy of a document we decided to take into consideration:
  • first level modifications, which are fundamental actions (typing, changing font size, and inserting pictures),
  • the correction of the document to avoid carrying the inefficient modification processes,
  • the formatting of a document.
If the document is properly edited and formatted, the correction and the formatting of the document is left out from both the text-management process and the analysis. Furthermore, if the original document carries errors, modifications were applied in both the original and the corrected documents to be able to compare the messages which the two forms of the document carry.
One further technique in the process of measuring the entropy of digital texts is the grouping of repeating steps. With this method, for example handling multiple Space, Tab, and Enter characters are not logged character by character. This method allows space for various solutions handling these repetitions (e.g., deleting with replacement, selecting by blocks or double clicks, etc.), reduces the steps of the algorithm to a reasonable size, and the time assigned to these steps can be measured in seconds. The results of the analyses are shown in the following tables (Tables in Section 5) where the steps are presented as modification events (Column Algorithm), to which the duration, the information content and the entropy values are assigned(Time (tk), Ik, and Ek, respectively).
Considering the formatting of the documents, in the testing process these steps were performed only in the correct version of the documents to avoid the multiplication of errors.
Following these concerns, the next sections provide the results of how the algorithms were set up and the entropy assigned to these algorithms.

5.1. Document: medicine–Correction

The algorithm of the correction of the medicine document is presented in Table 2. In general, it is found that two bits of data (Eb) are required to put on the communication channel to correct this document. The crucial steps of the algorithm are the following:
  • the recognition of the extra Space and Enter characters,
  • how to delete these characters,
  • how to remove all the font and paragraph formats to clear the typographic and formatting errors,
  • recognizing and correcting syntax errors.
Clearing all the errors of the medicine document took around 230 seconds. To remove multiple Space and Enter characters two methods were recorded:
  • selecting blocks and deleting them,
  • using replacement (Replace… command), where two Space or Enter characters were substituted by one and handling the line opening and closing Space characters.
In this document, there was no significant difference between the times spent on the two methods.
After correcting the medicine document, it has eight paragraphs, and there is no unnecessary character left (Figure 4 right, and Figure 5 right). In the next step, the formatting of the corrected document took place.

5.2. Document: medicine–Formatting

The algorithm of the formatting, along with the time spent on the atomic steps (events) and the information content of the steps of the medicine document is presented in Table 3. To format this document three bits of data (Eb) must be put on the channel.
At first sight (and originally meant to be by the author), this document is an easy to handle little thing. However, the power of this document is underestimated. The entropy of the formatting reveals that the knowledge pieces required by this document are extremely demanding. The three bits of data needed to format this document indicate that firm background knowledge is necessary to complete the task. Calculating the entropy of this “easy” task explains the errors of the original document. Furthermore, it proves that minimal guidance [74] is not enough to teach fundamental word processing. A lot more data must be put on the channel than course books and other teaching-learning materials suggest [33,34,35,36,37,38,39,40].

5.3. Document: medicine–Modification

The modification of the medicine document has four stages:
In the erroneous text, typing two words into a paragraph requires additional steps to restore the appearance similar to the original, imitating the left indent. These additional steps are typing and deleting Space and Enter characters. With this method, three bits of data must be put on the channel to reach the expected arrangement (Table 4). However, we must mention a further side effect of this solution: whenever the content of the text is changed, adjusting the text to the fake ident must be repeated, which is an extremely time- and resource-consuming process.
Completing the “insert two words” task in the corrected document requires only one bit of data. In this case, purely the intention of the end-user (the typing) needs to be completed, no collateral activities are imposed.
The comparison of the entropy of the two corrections reveals that handling an erroneous document is much more demanding than a correct one. In this case, typing two words required almost three times more time, and around three times more data put on the channel in the erroneous document.
Further typing and content modification would have the same requirements with a linear increase in time. However, we calculated the time and the entropy of formatting of the medicine document (Table 3) and found that these tasks required three bits of data. The comparison of the entropy of typing in the erroneous document is around as much as the entropy of the formatting. We can conclude that if end-users want to work effectively in a digital text, the content first need to be correctly edited and formatted. In any other cases, the word processing activities are disadvantageous, we lose time, money, and resources.
Increasing the font size of a paragraph in the original (erroneous) document requires four bits of data (Eb). This can be explained by the collateral actions imposed on the user to realign the fabricated left indent. These actions are beyond the modification intention of the end-user (which was simply to just change the font size) and based on the entropy this is quite demanding.
In the corrected document, changing the font size of the selected paragraph took less than 5 s, and one bit of data must be put on the channel.

5.4. Document: frenchfood–Correction

The frenchfood document is supposed to have two tables, one on Page 1 (P1) and the other on Page 2 (P2). Instead of using two actual tables, the original document is loaded with fiddly attempts to imitate them. Due to the high number of layout errors of the frenchfood document, the correction took more than 380 s, and four bits of data (Eb) were needed (Table 8). This shows that recognizing the errors of a document is quite demanding and should be handled more consciously during the teaching-learning process; primarily, to avoid these errors, secondarily to be able to correct them before any modification is applied to the document.
The steps of the correction algorithm reveal that after deleting the multiple Space, Tab, and Enter characters, the most important knowledge pieces are the inserting and formatting tables.

5.5. Document: frenchfood–Formatting

The formatting of the corrected document took around 100 s, and three bits of data must be put on the channel (Table 9). End-users intended to modify this document must apply fundamental font and paragraph formatting in tables, handle borders in table with paragraph domain, format header with positioned tabulators, and set up the location of the header.

5.6. Document: frenchfood—Modification: Inserting Picture and Text

In the modification phase of the research, two pictures of food and their names were inserted into the empty cells of T2 on P2 (Figure 12). The modification was conducted both in the original (erroneous) and the corrected and formatted documents. The steps of these processes are presented in Table 10 and Table 11.
It is found that the modification of the original document is more demanding than that of the correct one. It required more than twice as much time and knowledge pieces in the erroneous document than in the corrected one. In both cases, the pictures needed to be inserted and the words needed to be typed, which took around the same amount of time. Adjusting the size of the pictures was conducted in both documents without any serious alteration in time. However, in the original document additional steps were required to adjust the pictures and the words (one below the other and horizontally centered). These steps took up time.

5.7. Document: frenchfood–Modification: Changing Font Size

Changing the Normal style with the font (Arial) and font size (32 pt) of the original and the corrected documents show remarkable differences. While in the original document it took around 300 s (Table 12), in the corrected document, only around 20 s were needed (Table 13). Furthermore, it is also remarkable that changing the font size and then adjusting the pictures and text require four bits of data (Table 12), while in the corrected document one bit is needed (Table 13). In general, we can conclude that handling erroneous documents is more difficult, complicated, and demanding than handling correct ones.

5.8. Document: equation–Modification

The equation document is different both from medicine and frenchfood in the sense that it cannot be corrected. The formulae of the equation document are so poorly designed that our research group gave up on its correction but created the correct formulae using the built-in equation editor of the word processor. Figure 17 presents the paragraphs of the erroneous and corrected equation documents, with the text boundaries made visible.
In this phase of our research, the original equation document was modified by adding new fractions to the original formula in accordance with the following concept:
Calculating the common denominator of the fractions (Figure 14) requires five bits of data in the erroneous (Table 14), and four in the correct (Table 15) document. Furthermore, the time spent on the actions is almost four times more in the erroneous document than in the correct one (Table 14 and Table 15). In the correct document, creating the common denominator requires nothing else but positioning the cursor, deleting the old value and typing the new. In the case of the erroneous formulae, additional time is required to adjust the position of denominators and numerators with Space characters, and formatting the equal sign and the vincula with further adjustments.
The following step of the analysis was to imitate the solving of the presented problem in the original lesson plan (Figure 16 and Figure 17). In this phase, several atomic actions were performed, for example inserting new fractions, typing numerators and denominators, simplifications, and calculations. Similar to calculating the common denominator, beyond the required elements in the formulae, collateral adjustments and formatting were imposed. Considering the number of knowledge pieces required to complete this task, in both cases five bits of data must be put on the channel (Table 16 and Table 17). However, the time spent on the task is 2.6 times more in the erroneous document (Table 16 and Table 17).
In the third test of the equation documents, a new fraction was added to the existing one (Figure 15). In this case, the number of bits which must be put on the channel in the erroneous (Table 18) and the correct (Table 19) documents was 3 and 2, respectively. The time spent on this modification is 3.3 times more in the erroneous (Table 18) than in the correct (Table 19) document.

6. Discussion

6.1. Changing Duration and Adding a New Atomic Step

In the previous section, the entropy of correction, formatting, and modification tasks in the three sample documents were detailed, and the results presented using the newly introduced concept of “text-entropy”. In the discussion section, we focus on the question, how the entropy of a task changes when:
  • more time is spent on an already existing atomic step (event), or
  • a new atomic step is inserted.
It is found that changes in the duration of atomic steps is a subcategory of the insertion of a new atomic step. Consequently, this latter, more general modification is analyzed. Figure 18 presents the modification of the correct frenchfood document. An additional fictitious atomic step is added to the steps listed in Table 11 (with a t = 100 s duration mapped in Figure 18). The time assigned to the action is in the range of [1, 1000] with a 1 s difference.
It is found that in both cases the entropy increases until we reach a certain amount of time (tmax), which is understood as further knowledge pieces are added as input data. However, when tmax is reached the entropy starts to decrease, this also means that without adding further atomic steps to the process, the time conceals further events. We call this concept hidden or time-concealed data.

6.2. Where We Stand

Millions of ill-treated Word documents circulate both in closed communities and on the internet providing bad samples. In both cases, the available documents include educational materials meant to teach word processing and offering exercises on the subject. In general, it is found that these matters and problems are invisible to IT professionals, corporate managers, and researchers in both the fields of information systems and computer education.
In Figure 19, Figure 20 and Figure 21 erroneous Word tutorials and exam papers are presented. The samples show that even educational documents do not fulfill the requirements of the properly structured and formatted text and are set up as bad examples.
These exam papers repeat exactly the kind of errors our research analyzed. The examples prove the negligent text-handling habits and demonstrate errors that exist in the teaching–learning process of text-editing in general.
The selected errors are the following:

6.3. Building Corpora

To our knowledge, open-source datasets of Word documents do not exist. Another problem, considering corpora, is that the copyright issues are not defined. Consequently no one knows under what conditions such corpora can be created and used. This explains our choice of samples in the paper, where the Word documents are from our private collection.
  • Author-data is rarely available, consequently there is no place to ask for permission.
  • There are no designated data banks of corpora through which copyright issues might be handled.
  • This entire process is a black box.
  • Documents found on the public internet are not guaranteed to stay there.

7. Conclusions

This paper presents how text-entropy can be calculated from text handling processes, which are primarily formatting, correction, and modification (including the creation of a text). The word processing tasks can be algorithmized, broken into events (atomic steps) to which durations of time can be assigned. These lengths of time serve as sources of probabilities, bases of information entropy, and, from now on, of text-entropy.
In the present study, three erroneous MS Word documents were selected to demonstrate how text-entropy can be applied to real-world digital texts. In this process, we have to build the algorithm of the tasks, measure the time spent on the events, and finally calculate the entropy of the completed tasks.
In the comparison of the erroneous and the correct documents, the text-entropy can be calculated from the entropy of modification tasks. In the present analysis, first level modifications [19]—typing, changing font size, inserting picture—are conducted in the medicine and the frenchfood documents, and inserting formulae in the equation documents. Furthermore, the entropy of correction and formatting of the original (erroneous) documents must be taken into consideration when calculating text-entropy. These pieces provide data about the complexity of documents, and on how correct or how erroneous they are.
We found that modifying a properly structured and formatted digital text is less demanding than its erroneous counterpart. It means that less data must be put on the communication channel than in the case of erroneous documents. The analysis also revealed that not only the quantity of the data is lower, but the quality of the data (the required knowledge pieces) is higher. As the consequence of these two findings, it is proven that the modification time of erroneous documents is severalfold that of the correct ones, even in the case of minimal first level tasks. It is also proven that to avoid the multiplication of time- and resource-consuming tasks, we need to correct the documents before their modification.
With the concept and the methods provided in the present paper, we were able to find an objective measurement—the entropy of digital texts—that can describe the quality of documents. Based on this value, it can be told how correct or how erroneous a digital text-based document is.
The analyses can be further extended to several directions. First, we must emphasize that the algorithms were set up and the steps were conducted by the members of our research group, who are professionals in text management and handling word processors. Less experienced end-users might need more time and different algorithms to conduct the modifications, which can generate results different from those presented here. Furthermore, it is clear from the results that end-users, primarily students must be taught how to design and create properly structured and formatted digital texts, and how to work effectively with digital texts by using complex word processors. As the results of our analyses revealed, the focus of education should be realigned from the interface of an application to the content and structure of the digital texts.
It is also found that the presented approach for calculating text-entropy provides a method that can be followed both in educational and industrial (business/office/firm) environments to test the correctness of documents and reveal their discrepancies, if there is any. The aim of introducing this concept and testing method is to reduce the use and waste of both human and machine resources in the process of handling digital text-based documents.

Author Contributions

Conceptualization, M.C. and T.N.; methodology, M.C., K.N. and T.N.; formal analysis, M.C. and T.N.; investigation, M.C. and T.N.; resources, M.C. and T.N.; data curation, M.C. and T.N.; writing—original draft preparation, M.C.; writing—review and editing, M.C., K.N. and T.N.; visualization, M.C. and T.N.; supervision, M.C.; project administration, M.C. and T.N.; funding acquisition, M.C. and T.N. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by the KDP-2021 Program of the Ministry for Innovation and Technology from the Source of the National Research, Development and Innovation Fund.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Johnson, E. A Little Learning about Word Processing. Available online: https://www.uv.es/~fores/programa/johnson_wordprocessing2.html (accessed on 21 January 2023).
  2. Wolfram, C. The Math(s) FIX: An Education Blueprint for the AI Age; Wolfram Media, Inc.: Champaign, IL, USA, 2020. [Google Scholar]
  3. Rother, M. Toyota Kata: Managing People for Improvement, Adaptiveness, and Superior Results; McGraw Hill: New York, NY, USA, 2010. [Google Scholar]
  4. Liker, J.K. The Toyota Way: 14 Management Principles from the World’s Greatest Manufacturer; McGraw-Hill: New York, NY, USA, 2021. [Google Scholar]
  5. Ben-Ari, M. Bricolage Forever! In Psychology of Programming Interest Group; University of Leeds: Leed, UK, 1999. [Google Scholar]
  6. Ben-Ari, M.; Yeshno, T. Conceptual Models of Software Artifacts. Interact. Comput. 2006, 18, 1336–1350. [Google Scholar] [CrossRef]
  7. Csernoch, M. Teaching Word Processing—The Theory Behind. Teach. Math. Comput. Sci. 2009, 7, 119–137. [Google Scholar] [CrossRef]
  8. Csernoch, M. Teaching Word Processing—The Practice. Teach. Math. Comput. Sci. 2010, 8, 247–262. [Google Scholar] [CrossRef]
  9. Csernoch, M. Methodological Questions of Teaching Word Processing. In 3rd International Conference on Applied Informatics; Eszterházy Károly Catholic University: Eger-Noszvaj, Hungary, 1997; pp. 375–382. [Google Scholar]
  10. Csernoch, M.; Biró, P. Wasting Human and Computer Resources. Int. J. Soc. Educ. Econ. Manag. Eng. 2015, 9, 573–581. [Google Scholar]
  11. Csernoch, M. Thinking Fast and Slow in Computer Problem Solving. J. Softw. Eng. Appl. 2017, 10, 11–40. [Google Scholar] [CrossRef]
  12. Sebestyén, K.; Csapó, G.; Csernoch, M.; Aradi, B. Error Recognition Model: High-Mathability End-User Text Management. Acta Polytech. Hung. 2022, 19, 151–170. [Google Scholar]
  13. Csapó, G. The Maturation Exam of Informatics and the Text-Management Culture of First Year Students of Informatics. In Hungarian: Az informatika érettségi és az informatikushallgatók szövegkezelési kultúrája. In Válogatott Tanulmányok a Pedagógiai Elmélet és Szakmódszertanok Köréből; Tibor, K.J., Ed.; International Research Institute: Budapest, Hungary, 2017; pp. 387–394. [Google Scholar]
  14. Csernoch, M.; Dani, E. Do You Speak, Write and Think in Informatics? Acta Polytech. Hung. 2022, 19, 113–131. [Google Scholar] [CrossRef]
  15. Biró, P.; Csernoch, M. Teaching Word Processing with and without Algorithms. In Proceedings of the Hungarian Conference on Educational Research, Szeged, Hungary, 29–30 May 2014. [Google Scholar]
  16. Csernoch, M. Clearing Up Misconceptions About Teaching Text Editing. In 4th International Conference of Education, Research and Innovation; IATED: Valencia, Spain, 2011; pp. 407–415. [Google Scholar]
  17. Csernoch, M. Do You Speak and Write in Informatics? In Proceedings of the 10th International Multi-Conference on Complexity, Informatics and Cybernetics (IMCIC 2019), Orlando, FL, USA, 12–15 March 2019; International Institute of Informatics and Cybernetics: Orlando, FL, USA, 2019; Volume I, pp. 147–152. [Google Scholar]
  18. Biró, P.; Csernoch, M. User-Friendly Errors of Text-Management. In Hungarian: Felhasználóbarát Szövegkezelési Hibák. In A Pedagógusképzés XXI. Századi Perspektívái; Babeș–Bolyai Tudományegyetem Pedagógia és Alakalmazott Didatika Intézet és a DOCEO Egyeszület: Kolozsvár, Romania, 2016; pp. 177–186. [Google Scholar]
  19. Hankiewicz, K.; Butlewski, M. Efficiency in Performing Basic Tasks Using Word Processing Programs by the Elderly as a Measure of the Ergonomic Quality of Software. In Human-Computer Interaction. Theories, Methods, and Tools. HCI 2014; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2014; pp. 481–488. [Google Scholar]
  20. Kirschenbaum, M.G. Track Changes: A Literary History of Word Processing; The Belknap Press of Harvard University Press: Cambridge, MA, USA, 2016. [Google Scholar]
  21. Walker, A. 23 Best Word Processing Software. 2023. Available online: https://www.guru99.com/best-free-word-processor.html (accessed on 16 January 2023).
  22. Curts, E. How to Show Non-Printing Characters in Google Docs. Available online: https://laptrinhx.com/how-to-show-non-printing-characters-in-google-docs-989631243/ (accessed on 21 January 2023).
  23. Kenyon, C.K. Show/Hide Non-Printing Formatting Characters. Available online: https://www.addbalance.com/word/nonprinting.htm (accessed on 21 January 2023).
  24. Barnhill, S. Word’s Nonprinting Formatting Marks. Available online: http://wordfaqs.ssbarnhill.com/NonprintChars.htm (accessed on 24 January 2023).
  25. Dodlapati, S.R.; Lakkaraju, P.; Tulluru, N.; Zeng, Z. Non Printable & Special Characters: Problems and How to Overcome Them. Available online: https://www.lexjansen.com/nesug/nesug10/ff/ff04.pdf (accessed on 22 January 2023).
  26. Prensky, M. Digital Natives, Digital Immigrants. Horizon 2001, 9, 1–6. [Google Scholar]
  27. Kruger, J.; Dunning, D. Unskilled and Unaware of It: How Difficulties in Recognizing One’s Own Incompetence Lead to Inflated Self-Assessments. J. Personal. Soc. Psychol. 1999, 77, 1121–1134. [Google Scholar] [CrossRef]
  28. Gibbs, S.; Moore, K.; Steel, G.; McKinnon, A. The Dunning-Kruger Effect in a Workplace Computing Setting. Comput. Hum. Behav. 2017, 72, 589–595. [Google Scholar] [CrossRef]
  29. Staub, S.; Kaynak, R. Is an Unskilled Really Unaware of It? Procedia Soc. Behav. Sci. 2014, 150, 899–907. [Google Scholar] [CrossRef]
  30. Bucki, L.A. Microsoft Word 2013 Bible; Wiley: Hoboken, NJ, USA, 2013. [Google Scholar]
  31. Philo, A.; Angstadt, M. Microsoft Word 2016 Step By Step; MC-NPL Computer Lab: Norristown, PA, USA, 2016. [Google Scholar]
  32. Gookin, D. Word 2019 for Dummies; John Wiley & Sons Inc.: Hoboken, NJ, USA, 2019. [Google Scholar]
  33. Lénárd, A.; Sarbó, G.; Tarné Éder, M.; Turzó-Sovák, N. Digital Culture 3, in Hungarian: Digitális Kultúra 3. Oktatási Hivatal. 2022. Available online: https://www.tankonyvkatalogus.hu/pdf/OH-DIG03TA__teljes.pdf (accessed on 6 July 2022).
  34. Abonyi-Tóth, A.; Farkas, C.; Turzó-Sovák, N.; Varga, P. Digital Culture 6, in Hungarian: Digitális kultúra 6. Oktatási Hivatal. 2021. Available online: https://www.tankonyvkatalogus.hu/pdf/OH-DIG06TA__teljes.pdf (accessed on 11 April 2022).
  35. Abonyi-Tóth, A.; Farkas, C.; Turzó-Sovák, N.; Varga, P. Digital Culture 7, in Hungarian: Digitális Kultúra 7. Oktatási Hivatal. 2022. Available online: https://www.tankonyvkatalogus.hu/pdf/OH-DIG07TA__teljes.pdf (accessed on 11 April 2022).
  36. Varga, P.; Jeneiné Horváth, K.; Reményi, Z.; Farkas, C.; Takács, I.; Siegler, G.; Abonyi-Tóh, A. Digital Culture 9, in Hungarian: Digitális Kultúra 9. Oktatási Hivatal. 2020. Available online: https://www.tankonyvkatalogus.hu/pdf/OH-DIG09TA__teljes.pdf (accessed on 11 April 2022).
  37. Joos, I.; Wolf, D.M.; Nelson, R. Introduction to Word Processing. In Introduction to Computers for Healthcare Professionals; Jones & Bartlett Learning: Burlington, MA, USA, 2021; pp. 215–272. [Google Scholar]
  38. Role, J. Domestic Data Entry Operator. Textbook for Class IX. YOUR SCHOOL POINT. 1 January 2020. pp. 45–97. Available online: https://stgregoriosschooldelhi.ac.in/pdf/Information-Technology-Class-IX.pdf (accessed on 11 April 2022).
  39. Ciocaru, L.; Penea, S.; Rusu, O.; Mitrache, C.E. Informatics and ICT 7. In Hungarian: Informatika és IKT. Tankönyv a VII. Osztály Számára; Ministerului Educației și Cercetării: Bucharest, Romania, 2017. [Google Scholar]
  40. Florea, A.; Săcuiu, S.E. Informatics and ICT 7. In Hungarian: Informatika és IKT VII. Osztály; Ministerul Educației: Bucharest, Romania, 2017. [Google Scholar]
  41. Gove, M. Digital Literacy Campaign—Michael Gove’s Speech in Full. Available online: http://www.theguardian.com/education/2012/jan/11/digital-literacy-michael-gove-speech (accessed on 21 January 2023).
  42. ECDL Base: Word Processing Training Course & Certification Information. Available online: http://www.ecdlonline.uk/ecdl-courses/certification/ecdl-basic-word-processing.html (accessed on 24 January 2023).
  43. ECDL, Foundation. Best Practice Awards; Digital Literacy AG: Basel, Switzerland, 2015. [Google Scholar]
  44. Kirschner, P.A.; De Bruyckere, P. The Myths of the Digital Native and the Multitasker. Teach. Teach. Educ. 2017, 67, 135–142. [Google Scholar] [CrossRef]
  45. Lister, R. After the Gold Rush: Toward Sustainable Scholarship in Computing. In Proceedings of the Tenth Conference on Australasian Computing Education; Australian Computer Society, Inc.: Darlinghurst, Australia, 2008; Volume 78, pp. 3–17. [Google Scholar]
  46. Gibbs, S.F.; McLennan, T.J. Has the Computing Competence of First Year University Students Increased during the Last Decade? In Ascilite 2008 Melbourne: Hello! Where Are You in the Landscape of Educational Technology?: Proceedings: 25th Annual ASCILITE Conference: 30 November–3 December 2008; ASCILITE: Burwood, VIC, Australia, 2008; pp. 632–640. [Google Scholar]
  47. Gibbs, S.; McKinnon, A.; Steel, G. Are Workplace End-User Computing Skills at a Desirable Level? A New Zealand Perspective. Available online: https://aisel.aisnet.org/amcis2014/EndUserIS/GeneralPresentations/5/ (accessed on 22 January 2023).
  48. Gibbs, S. Computer Self-Efficacy—Is There a Gender Gap in Tertiary Level Introductory Computing Classes? J. Appl. Comput. Inf. Technol. 2013, 17, 2013. [Google Scholar]
  49. Gibbs, S.F. The Above-Average Effect in an End-User Computing Context. Ph.D. Thesis, Lincoln University, Christchurch, New Zealand, 2016. [Google Scholar]
  50. Gibbs, S.F.; Steel, G.; Kuiper, A. Expectations of Competency: The Mismatch between Employers’ and Graduates’ Views of End-User Computing Skills Requirements in the Workplace. J. Inf. Technol. Educ. Res. 2011, 10, 371–382. [Google Scholar] [CrossRef] [PubMed]
  51. Malmi, L.; Sheard, J.; Kinnunen, P.; Simon; Sinclair, J. Computing Education Theories. What Are They and How Are They Used? In Proceedings of the 2019 ACM Conference on International Computing Education Research, Toronto, ON, Canada, 12–14 August 2019; pp. 187–197. [Google Scholar]
  52. Malmi, L.F.; Sheard, J.; Kinnunen, P.; Simon; Sinclair, J. Development and Use of Domain-Specific Learning Theories, Models, and Instruments in Computing Education. ACM Trans. Comput. Educ. 2022, 23, 1–48. [Google Scholar] [CrossRef]
  53. Freiermuth, K.; Hromkovič, J.; Steffen, B. Creating and Testing Textbooks for Secondary Schools. In Proceedings of the International Conference on Informatics in Secondary Schools—Evolution and Perspectives ISSEP 2008: Informatics Education—Supporting Computational Thinking, Torun, Poland, 1–4 July 2008; pp. 216–228. [Google Scholar]
  54. Wing, J.M. Computational Thinking. Commun. ACM 2006, 49, 33–35. [Google Scholar] [CrossRef]
  55. Csernoch, M.; Máth, J.; Nagy, T. The Interpretation of Graphical Information in Word Processing. Entropy 2022, 24, 1492. [Google Scholar] [CrossRef]
  56. Tulis, M.; Steuer, G.; Dresel, M. Learning from Errors: A Model of Individual Processes. Frontline Learn. Res. 2016, 4, 12–26. [Google Scholar] [CrossRef]
  57. McLaren, B.M.; Adams, D.; Durkin, K.; Goguadze, G.; Mayer, R.E.; Rittle-Johnson, B.; Sosnovsky, S.; Isotani, S.; van Velsen, M. To Err Is Human, to Explain and Correct Is Divine: A Study of Interactive Erroneous Examples with Middle School Math Students. In 21st Century Learning for 21st Century Skills: EC-TEL 2012; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; pp. 222–235. [Google Scholar]
  58. Metcalfe, J. Learning from Errors. Annu. Rev. Psychol. 2017, 68, 465–489. [Google Scholar] [CrossRef]
  59. Webb, D.C. Troubleshooting Assessment: An Authentic Problem Solving Activity for It Education. Procedia Soc. Behav. Sci. 2010, 9, 903–907. [Google Scholar] [CrossRef]
  60. Verkhoshansky, Y.V. Main Features of a Modern Scientific Sports Training Theory. Available online: https://www.verkhoshansky.com/Portals/0/Articles/English/Main%20Features%20of%20a%20Modern%20Scientific%20Sports%20Training%20Theory.pdf (accessed on 22 January 2023).
  61. Buns, M.T. Skill Acquisition—Evidence-Based Practices for Detecting and Correcting Errors. Available online: https://coachesinsider.com/track-x-country/skill-acquisition-evidence-based-practices-for-detecting-and-correcting-errors-article/ (accessed on 22 January 2023).
  62. Cushion, E.; Howe, L.; Read, P.; Spence, A. A Process for Error Correction for Strength and Conditioning Coaches. Strength Cond. J. 2017, 39, 84–92. [Google Scholar] [CrossRef]
  63. Heinen, T. The Error-Correction Process—A Heuristic Concept. In Advances in Visual Perception Research; Nova Biomedical: New York, NY, USA, 2015; pp. 193–202. [Google Scholar]
  64. Zhao, X. Analysis and Correction of Wrong Technical Actions in Juvenile Sports Training Based on Deep Learning. Comput. Intell. Neurosci. 2022, 2022, 6492410. [Google Scholar] [CrossRef]
  65. Polya, G. How to Solve It; Princeton University Press: New York, NY, USA, 1945. [Google Scholar]
  66. Chen, J.; Morris, D.; Mansour, N. Sience Teachers’ Beliefs: Perceptions of Efficacy and the Nature of Scientific Knowledge and Knowing. In International Handbook of Research on Teachers’ Beliefs; Routledge: New York, NY, USA, 2014; pp. 382–398. [Google Scholar]
  67. Baranyi, P.; Gilányi, A. Mathability: Emulating and Enhancing Human Mathematical Capabilities. In Proceedings of the 2013 IEEE 4th International Conference on Cognitive Infocommunications (CogInfoCom), Budapest, Hungary, 2–5 December 2013. [Google Scholar]
  68. Baranyi, P.; Csapó, Á.; Sallai, G. Cognitive Infocommunications (COGINFOCOM); Springer International Publishing: Cham, Switzerland, 2015; p. 191. [Google Scholar]
  69. Biró, P.; Csernoch, M. The Mathability of Computer Problem Solving Approaches. In Proceedings of the 2015 6th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), Győr, Hungary, 19–21 October 2015; pp. 111–114. [Google Scholar]
  70. Wolfram, C. Evidence: Let’s Promote Not Stifle Innovation in Education. Available online: https://www.conradwolfram.com/writings/2015/5/21/role-of-evidence-in-education-innovation (accessed on 24 January 2023).
  71. Kahneman, D. Thinking, Fast and Slow; Farrar, Straus and Giroux: New York, NY, USA, 2011. [Google Scholar]
  72. Sweller, J.; Ayres, P.; Kalyuga, S. Cognitive Load Theory in Perspective. In Cognitive Load Theory; Springer: New York, NY, USA, 2011; pp. 237–242. [Google Scholar]
  73. Hattie, J. Visible Learning for Teachers: Maximizing Impact on Learning; Routledge: New York, NY, USA, 2012. [Google Scholar]
  74. Kirschner, P.A.; Sweller, J.; Clark, R.E. Why Minimal Guidance during Instruction Does Not Work: An Analysis of the Failure of Constructivist, Discovery, Problem-Based, Experiential, and Inquiry-Based Teaching. Educ. Psychol. 2006, 41, 75–86. [Google Scholar] [CrossRef]
  75. Virágvölgyi, P. The Mastery of Typography with Computers. In Hungarian: A Tipográfia Mestersége: Számítógéppel; Osiris: Budapest, Hungary, 2004. [Google Scholar]
  76. Jury, D. About Face: Reviving the Rules of Typography; Rotovision: Brighton, UK, 2004. [Google Scholar]
  77. Jury, D. What Is Typography? Ivy Press: Brighton, UK, 2017. [Google Scholar]
  78. Reynolds, G. Presentation Zen: Simple Ideas on Presentation Design and Delivery; New Riders: Berkeley, CA, USA, 2020. [Google Scholar]
  79. Vahrenhold, J.; Caspersen, M.; Berry, G.; Gal-Ezer, J.; Kölling, M.; McGettrick, A.; Westermeier, M. Informatics Education in Europe: Are We All in the Same Boat? The Committee on European Computing Education (CECE): New York, NY, USA, 2017. [Google Scholar]
  80. Caspersen, M.; Diethelm, I.; Gal-Ezer, J.; McGettrick, A.; Nardelli, E.; Passey, D.; Rovan, B.; Webb, M. Informatics Reference Framework for School; Informatics for All. 2022. Available online: https://www.informaticsforall.org/wp-content/uploads/2022/03/Informatics-Reference-Framework-for-School-release-February-2022.pdf (accessed on 22 December 2022).
  81. Mishra, P.; Koehler, M.J. Technological Pedagogical Content Knowledge: A Framework for Teacher Knowledge. Teach. Coll. Rec. Voice Scholarsh. Educ. 2006, 108, 1017–1054. [Google Scholar] [CrossRef]
  82. Angeli, C. Teaching Spreadsheets: A TPCK Perspective. In Improving Computer Science Education; Routledge: New York, NY, USA; London, UK, 2013; pp. 140–153. [Google Scholar]
  83. Angeli, C.; Valanides, N. Technological Pedagogical Content Knowledge Exploring, Developing, and Assessing TPCK; Springer: Boston, MA, USA, 2015. [Google Scholar]
  84. Shannon, C. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  85. Shannon, C.E. Prediction and Entropy of Printed English. Bell Syst. Tech. J. 1951, 30, 50–64. [Google Scholar] [CrossRef]
  86. Gray, R.M. Entropy and Information Theory; Springer: Stanford, CA, USA, 2011. [Google Scholar]
  87. O’Neil, C.; Schutt, R. Entropy. In Doing Data Science; O’Reilly Media: Sebastopol, CA, USA, 2014; pp. 186–187. [Google Scholar]
  88. Kinsner, W. Is Entropy Suitable to Characterize Data and Signals for Cognitive Informatics. In Novel Approaches in Cognitive Informatics and Natural Intelligence? IGI Global: Hershey, PA, USA, 2008; pp. 28–51. [Google Scholar]
  89. Rényi, A. On Measures of Information and Entropy. In Berkeley Symposium on Mathematical Statistics and Probability; University of California Press: Berkeley, CA, USA, 1961; pp. 547–561. [Google Scholar]
  90. Tóthné Laufer, E. The Fundamantals of Information and Code Theory. In Hungarian: Az Információ- és Kódelmélet Alapjai; Óbudai Egyetem: Budapest, Hungary. Available online: http://siva.bgk.uni-obuda.hu/~laufer/bevinfo_tankonyv/Informatika%20alapjai%20jegyzet.pdf (accessed on 22 December 2022).
Figure 1. The translated version of the original Hungarian single page document, named medicine.
Figure 1. The translated version of the original Hungarian single page document, named medicine.
Entropy 25 00302 g001
Figure 2. The original three-page long frenchfood document.
Figure 2. The original three-page long frenchfood document.
Entropy 25 00302 g002
Figure 3. The original formula-lookalikes from a senior pre-service teacher’s lesson plan, the document named equations.
Figure 3. The original formula-lookalikes from a senior pre-service teacher’s lesson plan, the document named equations.
Entropy 25 00302 g003
Figure 4. The paragraphs of the original and the corrected medicine documents, with the text boundaries made visible.
Figure 4. The paragraphs of the original and the corrected medicine documents, with the text boundaries made visible.
Entropy 25 00302 g004
Figure 5. MS Word statistics of the original (left) and the corrected (right) medicine documents.
Figure 5. MS Word statistics of the original (left) and the corrected (right) medicine documents.
Entropy 25 00302 g005
Figure 6. The original three-page long frenchfood document with visible text boundaries.
Figure 6. The original three-page long frenchfood document with visible text boundaries.
Entropy 25 00302 g006
Figure 7. The corrected two-page long frenchfood document with visible text boundaries.
Figure 7. The corrected two-page long frenchfood document with visible text boundaries.
Entropy 25 00302 g007
Figure 8. MS Word statistics of the original (left) and the corrected (right) frenchfood documents.
Figure 8. MS Word statistics of the original (left) and the corrected (right) frenchfood documents.
Entropy 25 00302 g008
Figure 9. The corrected and formatted two-page long frenchfood document with visible text boundaries.
Figure 9. The corrected and formatted two-page long frenchfood document with visible text boundaries.
Entropy 25 00302 g009
Figure 10. The expression “and vitamins” (red box) is inserted into the corrected (upper) and the erroneous (lower) medicine documents.
Figure 10. The expression “and vitamins” (red box) is inserted into the corrected (upper) and the erroneous (lower) medicine documents.
Entropy 25 00302 g010
Figure 11. Font size is changed in the corrected (upper) and the original, erroneous (lower) the medicine documents.
Figure 11. Font size is changed in the corrected (upper) and the original, erroneous (lower) the medicine documents.
Entropy 25 00302 g011
Figure 12. The second page of the original (upper) and the corrected (lower) frenchfood documents.
Figure 12. The second page of the original (upper) and the corrected (lower) frenchfood documents.
Entropy 25 00302 g012
Figure 13. The results of changing the font size in the original (left) and the corrected and formatted (right) frenchfood documents.
Figure 13. The results of changing the font size in the original (left) and the corrected and formatted (right) frenchfood documents.
Entropy 25 00302 g013
Figure 14. Calculating the common denominator of the fractions in the original (left) and in the corrected (right) equation documents.
Figure 14. Calculating the common denominator of the fractions in the original (left) and in the corrected (right) equation documents.
Entropy 25 00302 g014
Figure 15. Adding a new fraction (red box) to the original (left) and the corrected (right) equation documents.
Figure 15. Adding a new fraction (red box) to the original (left) and the corrected (right) equation documents.
Entropy 25 00302 g015
Figure 16. Solving the substitution problem in the original (left) and in the corrected (right) equation documents.
Figure 16. Solving the substitution problem in the original (left) and in the corrected (right) equation documents.
Entropy 25 00302 g016
Figure 17. The paragraphs of the original (left) and correct (right) equation documents with the text boundaries made visible.
Figure 17. The paragraphs of the original (left) and correct (right) equation documents with the text boundaries made visible.
Entropy 25 00302 g017
Figure 18. The changes in the entropy of a newly introduced atomic step (Ek) and the entropy (E) of a modification task in the function of time in the frenchfood document (E_original = 1.83 (Table 11) indicates the entropy before the modification).
Figure 18. The changes in the entropy of a newly introduced atomic step (Ek) and the entropy (E) of a modification task in the function of time in the frenchfood document (E_original = 1.83 (Table 11) indicates the entropy before the modification).
Entropy 25 00302 g018
Figure 19. Errors in a Word exam paper.
Figure 19. Errors in a Word exam paper.
Entropy 25 00302 g019
Figure 20. Errors in a Word exam paper.
Figure 20. Errors in a Word exam paper.
Entropy 25 00302 g020
Figure 21. Errors in a Word exam paper.
Figure 21. Errors in a Word exam paper.
Entropy 25 00302 g021
Figure 22. Errors in a Word exam paper.
Figure 22. Errors in a Word exam paper.
Entropy 25 00302 g022
Table 1. Modifications applied to the three sample documents of the study.
Table 1. Modifications applied to the three sample documents of the study.
DocumentModification1Modification2
medicinetyping two wordschanging font size
frenchfoodadding two pictures and nameschanging font size
equationadding new fractionssimplifying formula
Table 2. The steps of the correction of the medicine document.
Table 2. The steps of the correction of the medicine document.
AlgorithmTime (tk)IkEk
clearing all font and paragraph formatting5.005.52620.1199
deleting Space blocks22.933.32630.3316
deleting leading Spaces13.034.14310.2345
deleting empty paragraphs25.073.19760.3485
moving date above signature2.966.27650.0810
deleting end of line Enters, hyphenation, and paragraph-closing Space27.133.08440.3637
correcting syntax errors133.950.78040.4543
Total230.07 1.9336
Table 3. The formatting steps of the medicine document.
Table 3. The formatting steps of the medicine document.
AlgorithmTime (tk)IkEk
modifying Normal style18.102.52170.4391
left indent and copying15.892.70960.4142
titles: Font size, vertical Spacing, and copying32.121.69420.5236
date: vertical Spacing14.002.89230.3896
signature: right alignment10.883.25600.3408
hyphenation12.953.00470.3744
Total103.94 2.4817
Table 4. The steps of inserting two words into the original (erroneous) medicine document.
Table 4. The steps of inserting two words into the original (erroneous) medicine document.
AlgorithmTime (tk)Ik
positioning the cursor5.092.69410.4163
typing6.862.26360.4714
typing Enter5.122.68560.4174
typing Space7.012.23240.4751
deleting Enter2.863.52580.3061
deleting Space6.002.45680.4475
Total32.94 2.5338
Table 5. The steps of inserting two words into the corrected medicine document.
Table 5. The steps of inserting two words into the corrected medicine document.
AlgorithmTime (tk)IkEk
positioning the cursor5.091.23130.5245
typing6.860.80070.4597
Total11.95 0.9841
Table 6. The steps of changing font size in the original medicine document.
Table 6. The steps of changing font size in the original medicine document.
AlgorithmTime (tk)IkEk
selection1.896.09470.0892
changing font size3.034.05820.2436
typing Enter1.174.74590.1769
typing Space4.804.29970.2183
deleting Enter2.982.30860.4660
deleting Space4.064.37630.2107
typing Enter16.143.72670.2815
typing Space3.853.98780.2514
deleting Enter6.043.14810.3551
deleting Space5.043.00510.3743
typing Enter9.023.32840.3313
typing Space9.964.31400.2169
deleting Enter7.965.40280.1277
deleting Space4.024.72190.1789
Total79.96 3.5219
Table 7. The steps of changing font size in the corrected medicine document.
Table 7. The steps of changing font size in the corrected medicine document.
AlgorithmTime (tk)IkEk
selection1.891.38030.5302
changing font size3.030.69930.4307
Total4.92 0.9609
Table 8. The correction steps of the erroneous frenchfood document.
Table 8. The correction steps of the erroneous frenchfood document.
AlgorithmTime (tk)IkEk
deleting all font and paragraph formatting6.145.96270.0956
changing proofing language to French16.914.50110.1988
deleting textbox on P1 and moving the text36.093.40740.3211
deleting the textbox on P219.024.33150.2151
deleting lines on P232.853.54310.3040
checking the paper size51.032.90760.3875
setting the margins to default14.984.67590.1829
deleting multiple Tabs30.953.62900.2933
replacing Space characters with a Tab on P21.038.53830.0230
moving the first paragraph of P1 to the Header7.155.74300.1072
deleting underscore characters 22.944.06110.2433
deleting paragraph closing Tabs42.023.18790.3498
creating table on P1 (T1)18.814.34750.2136
formatting T135.093.44790.3160
creating table on P2 (T2)12.924.88940.1650
formatting T234.993.45200.3154
Total382.92 3.7316
Table 9. The formatting steps of the frenchfood document.
Table 9. The formatting steps of the frenchfood document.
AlgorithmTime (tk)IkEk
setting font size in T18.933.48560.3112
setting alignment in T13.005.05930.1517
adding lines (borders) in T211.113.17050.3521
adding vertical space in T222.842.13080.4865
setting alignment in T213.162.92620.3850
borders in T2 (empty lines)40.991.28710.5274
formatting Header with positioned tabs, setting header only on P138.931.36150.5299
Total100.03 2.7438
Table 10. The steps of inserting two pictures and names into the original (erroneous) frenchfood document.
Table 10. The steps of inserting two pictures and names into the original (erroneous) frenchfood document.
AlgorithmTime (tk)IkEk
inserting new paragraphs9.055.47660.1230
moving words99.842.01300.4987
typing Tabs/Spaces65.182.62820.4251
adjusting non-printing characters25.054.00780.2491
deleting characters from the end of the document3.846.71340.0640
adding pictures31.953.65680.2899
adjusting picture size for using only two pages60.072.74600.4093
adding food names33.843.57390.3001
adjusting with Spaces and Tabs74.152.44220.4494
Total402.97 2.8087
Table 11. The steps of inserting two pictures and names into the corrected frenchfood document.
Table 11. The steps of inserting two pictures and names into the corrected frenchfood document.
AlgorithmTime (tk)IkEk
moving words to the empty text cells in T296.960.93190.4885
deleting characters from the end of the document4.045.51690.1205
adding pictures to the empty picture cells in T234.832.40900.4536
adding food names in the empty text cells in T220.133.20000.3482
adjusting picture size for using only two pages29.022.67220.4192
Total184.98 1.8300
Table 12. The steps of changing the font size in the original (erroneous) frenchfood document.
Table 12. The steps of changing the font size in the original (erroneous) frenchfood document.
AlgorithmTime (tk)IkEk
changing Normal style16.004.22870.2255
clear formatting6.965.42960.1260
deleting Tabs on P15.055.89240.0992
modifying textbox122.003.76930.2764
modifying textbox216.054.22420.2260
deleting Tabs on P213.024.52610.1964
arranging lines51.822.53330.4376
deleting Tabs and Spaces30.143.31510.3331
arranging lines69.962.10030.4898
deleting Tabs31.003.27450.3384
arranging lines37.982.98160.3775
Total299.98 3.1260
Table 13. The steps of changing the font size in the corrected frenchfood document.
Table 13. The steps of changing the font size in the corrected frenchfood document.
AlgorithmTime (tk)IkEk
changing Normal style12.060.66150.4182
clear formatting7.021.44310.5307
Total19.08 0.9489
Table 14. The steps of calculating the common denominator in the erroneous equation document.
Table 14. The steps of calculating the common denominator in the erroneous equation document.
AlgorithmTime (tk)IkEk
copying equation9.054.24660.2237
adjusting with Spaces21.023.03080.3708
formatting equal sign8.974.25940.2224
typing numerator7.024.61300.1885
typing denominator2.955.86380.1007
adjusting with Spaces4.115.38540.1288
typing numerator12.943.73070.2810
typing denominator4.965.11420.1477
formatting vincula13.073.71630.2827
typing numerator8.074.41190.2073
typing denominator2.945.86870.1004
adjusting with Spaces10.913.97690.2526
typing numerator11.093.95330.2552
typing denominator2.935.87360.1002
formatting vincula6.974.62330.1876
typing numerator3.095.79690.1043
typing denominator3.905.46100.1240
formatting vincula4.095.39240.1284
trial and error14.753.54190.3041
adjusting with Spaces11.913.85040.2669
formatting vincula7.054.60690.1891
Total171.79 4.1664
Table 15. The steps of calculating the common denominator in the correct equation document.
Table 15. The steps of calculating the common denominator in the correct equation document.
AlgorithmTime (tk)IkEk
copying equation7.052.63980.4236
typing equal sign5.013.13270.3572
typing numerator5.882.90160.3883
typing denominator2.054.42180.2063
typing numerator3.053.84870.2671
typing denominator2.943.90160.2611
typing numerator4.073.43240.3179
typing denominator2.933.90660.2605
typing numerator4.003.45750.3147
typing denominator2.034.43600.2049
typing numerator2.044.42890.2056
Total43.94 3.4655
Table 16. The steps of solving the problem in the erroneous equation document.
Table 16. The steps of solving the problem in the erroneous equation document.
AlgorithmTime (tk)IkEk
copying equation12.585.44590.1249
adjusting with Spaces12.585.44590.1249
formatting equal sign9.735.81630.1032
typing numerator10.685.68150.1107
typing denominator4.067.07610.0524
typing numerator5.756.57650.0689
typing denominator1.958.13240.0290
adjusting with Spaces5.816.55950.0695
typing numerator10.715.67760.1109
typing denominator3.937.12530.0510
adjusting with Spaces15.465.14870.1451
typing numerator2.977.53080.0407
typing denominator1.908.17620.0283
typing numerator3.917.13240.0508
typing denominator0.959.16880.0159
formatting vincula19.484.81480.1711
copying equation6.866.32000.0791
changing order43.773.64710.2911
adjusting with Spaces15.455.14960.1451
formatting vincula40.753.75040.2787
copying equation32.124.09350.2398
numerator (calculation)22.504.60730.1890
denominator (calculation)7.656.16330.0860
adjusting with Spaces16.645.04200.1530
numerator1 (substitution)23.304.55650.1936
numerator2 (substitution)1.818.24430.0272
adjusting with Spaces9.745.81480.1033
formatting equal sign10.855.65940.1120
formatting vincula17.484.97130.1585
adjusting with Spaces7.726.15050.0866
simplifying fraction128.104.28660.2196
simplifying fraction21.998.10400.0295
adjusting with Spaces18.564.88490.1653
formatting equal sign18.527.59860.0392
formatting vincula18.414.88790.1651
formatting equal sign15.554.89630.1644
result2.835.14050.1457
changing size, position24.394.49060.1998
adjusting with Spaces40.903.74490.2793
Total548.36 4.8485
Table 17. The steps of solving the problem in the correct equation document.
Table 17. The steps of solving the problem in the correct equation document.
AlgorithmTime (tk)IkEk
copying equation7.904.73900.1775
typing equal sign3.116.08390.0897
typing numerator5.995.13830.1459
typing denominator2.916.17980.0852
typing numerator4.175.66080.1119
typing denominator2.896.18980.0848
typing numerator2.956.16010.0861
typing denominator2.056.68520.0650
typing numerator5.005.39890.1280
typing denominator2.126.63680.0667
typing numerator4.935.41920.1266
typing denominator2.056.68520.0650
typing numerator3.006.13590.0873
typing denominator3.955.73900.1075
copying equation14.893.82460.2699
changing order19.003.47290.3128
copying equation8.064.71000.1800
calculating32.032.71950.4129
copying equation9.084.53810.1953
numerator1 (substitution)12.854.03710.2459
numerator2 (substitution)15.983.72260.2820
copying equation16.183.70470.2841
simplifying fraction19.834.42360.2061
simplifying fraction214.063.90730.2604
result5.985.14070.1457
Total210.96 4.2223
Table 18. The steps of adding a new fraction to the formula in the erroneous equation document.
Table 18. The steps of adding a new fraction to the formula in the erroneous equation document.
AlgorithmTime (tk)IkEk
typing numerator11.992.24530.4736
adjusting with Spaces11.942.25140.4728
typing denominator1.974.85090.1681
adjusting with Spaces2.964.26350.2220
formatting vincula9.122.64010.4235
trial and error9.882.52460.4387
formatting operator8.992.66080.4208
Total56.85 2.6195
Table 19. The steps of adding a new fraction to the formula in the correct equation document.
Table 19. The steps of adding a new fraction to the formula in the correct equation document.
AlgorithmTime (tk)Ik
inserting fraction10.980.63740.4098
typing numerator2.912.55320.4350
typing denominator3.192.42070.4521
Total17.08 1.2969
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Csernoch, M.; Nagy, K.; Nagy, T. The Entropy of Digital Texts—The Mathematical Background of Correctness. Entropy 2023, 25, 302. https://doi.org/10.3390/e25020302

AMA Style

Csernoch M, Nagy K, Nagy T. The Entropy of Digital Texts—The Mathematical Background of Correctness. Entropy. 2023; 25(2):302. https://doi.org/10.3390/e25020302

Chicago/Turabian Style

Csernoch, Mária, Keve Nagy, and Tímea Nagy. 2023. "The Entropy of Digital Texts—The Mathematical Background of Correctness" Entropy 25, no. 2: 302. https://doi.org/10.3390/e25020302

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop