Next Article in Journal
Self-Organized Complexity and Coherent Infomax from the Viewpoint of Jaynes’s Probability Theory
Next Article in Special Issue
Emergence and Evolution of Meaning: The General Definition of Information (GDI) Revisiting Program—Part I: The Progressive Perspective: Top-Down
Previous Article in Journal
Sentence Comprehension as Mental Simulation: An Information-Theoretic Perspective
Previous Article in Special Issue
From Genomics to Scientomics: Expanding the Bioinformation Paradigm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Epistemic Information in Stratified M-Spaces

Department of Mathematics, University of California, Los Angeles, 405 Hilgard Ave. Los Angeles, CA 90095, USA
Information 2011, 2(4), 697-726; https://doi.org/10.3390/info2040697
Submission received: 15 September 2011 / Revised: 24 November 2011 / Accepted: 1 December 2011 / Published: 16 December 2011
(This article belongs to the Special Issue Selected Papers from "FIS 2010 Beijing")

Abstract

: Information is usually related to knowledge. However, the recent development of information theory demonstrated that information is a much broader concept, being actually present in and virtually related to everything. As a result, many unknown types and kinds of information have been discovered. Nevertheless, information that acts on knowledge, bringing new and updating existing knowledge, is of primary importance to people. It is called epistemic information, which is studied in this paper based on the general theory of information and further developing its mathematical stratum. As a synthetic approach, which reveals the essence of information, organizing and encompassing all main directions in information theory, the general theory of information provides efficient means for such a study. Different types of information dynamics representation use tools of mathematical disciplines such as the theory of categories, functional analysis, mathematical logic and algebra. Here we employ algebraic structures for exploration of information and knowledge dynamics. In Introduction (Section 1), we discuss previous studies of epistemic information. Section 2 gives a compressed description of the parametric phenomenological definition of information in the general theory of information. In Section 3, anthropic information, which is received, exchanged, processed and used by people is singled out and studied based on the Componential Triune Brain model. One of the basic forms of anthropic information called epistemic information, which is related to knowledge, is analyzed in Section 4. Mathematical models of epistemic information are studied in Section 5. In Conclusion, some open problems related to epistemic information are given.

1. Introduction

Explication and clarification of the concept of information requires not only elaboration and exploration of a general unified definition of information but also needs separation and examination of basic special types of information. The general theory of information gives the most general definition of information, organizing and encompassing all main directions in information theory [1]. According to the main principle of general theory of information, which impeccably correlates with people's practice and observations, when information comes to a system and is accepted, it causes changes in the system. This feature of information makes it reasonable to model information by operators and information dynamics by actions of these operators. This approach is adopted in the general theory of information.

In addition, the general theory of information provides efficient means for information classification and study of the basic information types. One of such types is information that is received, exchanged, processed and used by people. It is called anthropic information. We explain the methodological principles of the anthropic information explication and classification, which bring us to cognitive information and its important subclass called epistemic information.

Portions of epistemic information are modeled/represented by epistemic information operators acting in spaces of knowledge, which are represented by a formal construction called a Mizzaro space. These spaces consist of knowledge items often unified by structural relations.

In a general setting, epistemic information has been studied by different authors. Bar-Hillel and Carnap [2], Hintikka [3-5] and Israel and Perry [6] explored information in knowledge represented by means of mathematical logic. Shreider [7], Mackay [8], Brookes [9], Mizzaro [10,11] and Gackowski [12] base their theories on the following assumption:

Information is a change in a knowledge system

Later this principle has been made more exact [13] and formulated as

Epistemic information is a change in a knowledge system

The general theory of information [1] makes the next step to the better understanding of epistemic information. Namely, it is explained that

Epistemic information is a capacity to cause changes

in a knowledge system and it is possible to measure

this capacity by changes in the knowledge system

impacted by information

Such changes in the knowledge system of a system R reflect accepted information. Note that information can be accepted not only as true information but also as false information. In this case, changes in the knowledge system can result in exclusion of some knowledge or in labeling this knowledge as false, e.g., treating it as a misconception or blunder.

The approach of Mizzaro [10,11,13] and Burgin [1] to epistemic information does not consider knowledge in general, but makes use of the term a knowledge state (KS) of a dynamic knowledge system or a cognitive agent. It is postulated that knowledge of a knowledge system/agent consists of atomic components called knowledge items (KI) and the number of these items in a portion of knowledge gives an adequate measure of this knowledge. Such knowledge states are set-theoretical or unstructured Mizzaro spaces defined in [1].

This is the first approximation to modeling information processes related to knowledge transformations. In it, the substantial dimensions of knowledge are reflected, while the relational dimensions, which describe relations between knowledge items, are studied in higher-level models of operational information theory, which goes beyond the scope of this paper. Taking into account such relations allows one to achieve higher precision in measuring information and knowledge. However, there are many situations where precision provided by quantities of knowledge items is sufficient. For instance, exactly this level of precision is successfully used in software engineering and technology where program instructions, which belong to the procedural type of knowledge, are used as knowledge items that determine measures of information and knowledge transformations [14].

Taking knowledge states as the base, Mizzaro [13] considers and utilizes natural set-theoretical operations, such as union, difference, complement, and intersection, as well as set-theoretical relations, such as inclusion, emptiness, and membership, on the set of all knowledge states. These and more sophisticated operations and relations are advanced to the level of local information operators and studied in [1]. Namely, in the context of the general theory of information, transformations in knowledge states caused by receiving data (messages) are represented by local information operators, which are projections of global information operators also described in [1]. Consequently, the theory of Mizzaro spaces and local information operators in these spaces is a localization of the mathematical (formalized) stratum of the general theory of information.

In this paper, we continue exploration of epistemic information, extending Mizzaro spaces to M-spaces, which represent and model information dynamics by algebraic structures. In addition to uniform Mizzaro spaces studied in [1,11,13], here we consider Mizzaro spaces and M-spaces with stratification and extend the scope of epistemic information operators. In [1,11,13], only two operators, addition and deletion are studied. Here we consider five basic epistemic information operators: addition, deletion, moving, replication, and transformation, as well as their compositions. Section 2 gives a compressed description of the parametric phenomenological definition of information in the general theory of information. In Section 3, anthropic information that is received, exchanged, processed and used by people is separated and studied based on the Componential Triune Brain model introduced in [15] and further developed in [1]. This model allows overcoming the pitfalls of the simple linear hierarchy suggested by MacLean [16,17] by considering not anatomically localized but distributed in the brain basic systems and including such an important psychological construct as the Will [18,19] in the brain structure. Based on this model, one of the basic forms of anthropic information called epistemic information, which is related to knowledge, is analyzed in Section 4. Mathematical models of epistemic information are studied in Section 5. In Conclusion, we give some open problems related to epistemic information.

2. The Concept of Information

Our study of epistemic information is based on the system of ontological principles from the general theory of information. All these principles serve as an extended definition of information. A concise definition is given in the second ontological principle, which has several forms.

Ontological Principle O2 (the General Transformation Principle)

In a broad sense, information for a system R is a capacity to cause changes in the system R.

Thus, we may understand information in a broad sense as a capacity (ability or potency) of things, both material and abstract, to change other things. Information exists in the form of portions of information. Informally, a portion of information is such information that can be separated from other information. Information is, as a rule, about something. What information is about is called an object of this information.

The Ontological Principle O2 has several consequences. First, it demonstrates that information is closely connected to transformation. Namely, it means that information and transformation are functionally similar because they both point to changes in a system [20]. At the same time, they are different because information is potency for (or in some sense, cause of) change, while transformation is the change itself, or in other words, transformation is an operation, while information is what induces this operation.

Second, the Ontological Principle O2 explains why information influences society and individuals all the time, as well as why this influence grows with the development of society. Namely, reception of information by individuals and social groups induces transformation. In this sense, information is similar to energy. Moreover, according to the Ontological Principle O2, energy is a kind of information in a broad sense. This well correlates with the Carl Friedrich von Weizsacker's idea [21,22] that energy might in the end turn out to be information.

Third, the Ontological Principle O2 makes it possible to separate different kinds of information. For instance, people, as well as any computer, have many kinds of memory. It is even supposed that each part of the brain has several types of memory agencies that work in somewhat different ways, to suit particular purposes [23]. It is possible to consider each of these memory agencies as a separate system and to study differences between information that changes each type of memory. This might help to understand the interplay between stability and flexibility of mind, in general, and memory, in particular.

In essence, we can see that all kinds and types of information are encompassed by the Ontological Principle O2.

However, the common usage of the word information does not imply such wide generalizations as the Ontological Principle O2 implies. Thus, we need a more restricted theoretical meaning because an adequate theory, whether of the information or of anything else, must be in significant accord with our common ways of thinking and talking about what the theory is about, else there is the danger that theory is not about what it purports to be about. To achieve this goal, we use the concept of an infological system IF(R) of the system R for the information definition. It is done in two steps. At first, we make the concept of information relative and then we choose a specific class of infological systems to specify information in the strict sense. That is why it is impossible and, as well as, counterproductive to give an exact and thus, too rigid and restricted definition of an infological system. Indeed, information is a very rich and widespread phenomenon to be reflected by a restricted rigid definition (cf., for example, [24,25]).

The concept of infological system plays the role of a free parameter in the general theory of information, providing for representation of different kinds and types of information in this theory. That is why the concept of infological system, in general, should not be limited by boundaries of exact definitions. A free parameter must really be free. Identifying an infological system IF(R) of a system R, we can define information relative to this system. This definition is expressed in the following principle.

Ontological Principle O2g (the Relativized Transformation Principle)

Information for a system R relative to the infological system IF(R) is a capacity to cause changes in the system IF(R).

As a model example of an infological system IF(R) of an intelligent system R, we take the system of knowledge of R. In cybernetics, it is called the thesaurus Th(R) of the system R. Another example of an infological system is the memory of a computer. Such a memory is a place in which data and programs are stored and is a complex system of diverse components and processes.

Elements from IF(R) are called infological elements.

There is no exact definition of infological elements although there are various entities that are naturally considered as infological elements as they allow one to build theories of information that inherit conventional meanings of the word information. For instance, knowledge, data, images, algorithms, procedures, scenarios, ideas, values, goals, ideals, fantasies, abstractions, beliefs, and similar objects are standard examples of infological elements.

When we take a physical system D as the infological system and allow only for physical changes, we see that information with respect to D coincides with (physical) energy.

Taking a mental system B as the infological system and considering only mental changes, information with respect to B coincides with mental energy.

These ideas are crystallized in the following principle.

Ontological Principle O2a (the Special Transformation Principle)

Information in the strict sense or proper information or, simply, information for a system R, is a capacity to change structural infological elements from an infological system IF(R) of the system R.

As the concept of mental energy is much less understood than the concept of physical energy, we give some explanations based on the origin, development and contemporary understanding of mental energy.

Considering mental energy on the level of individual mentality, it is possible, as the first approximation, to equate it with psychic energy. The concept of psychic energy first entered physiology and to some extent psychology through the discussions of Ernst Brücke, Herman Helmholtz, and Emil Du Bois-Reymond, who during the years 1838-1842, worked in the laboratory of the German physiologist Johannes Muller. At the same time, according to Jung [26], Nicolas von Grot was the first to explicitly define the concept of psychic. He wrote [27]:

“The concept of psychic energy is as much justified in science as that of physical energy, and psychic energy has just as many quantitative measurements and different forms as has physical energy.”

Later this concept was further developed by Sigmund Freud [28,29], Brücke's student, and then by Carl Jung, Freud's student. Jung [26] regarded psychic energy as a basic life-force, manifesting itself through actions, such as eating, moving, thinking, sex, remembering, etc.

Contemporary understanding determines mental energy as the ability to perform mental tasks, the intensity of feelings of energy/fatigue, and the motivation to accomplish mental and physical tasks [30,31]. This shows that basic aspects of mental energy manifestation include: (a) cognition (knowledge that is gained through perception, reasoning or intuition), (b) changing of moods or feelings (states of mind), and (c) motivation (an incentive for action). Factors that can influence mental energy include, among others, psychological issues such as interest, passion, desire, concern, and biological issues such as genetics, nutrition, pain, and sleep [31].

Often psychic/mental energy is confused with psychic/mental force. For instance, some psychoanalysts do not distinguish energy and force, particularly, when they follow Freud's occasional practice of calling libido itself a force. Actually, as in physics, energy and force are different phenomena. Energy is a potential to make changes, while force is what is making changes. This allows one to take the Will as a representative of force in the individual mentality and above. This is expressed in the remark attributed to Loewenstein that “force is energy in action” (cf. [32]).

3. The phenomenological stratum of the general theory of information

In this paper, we are primarily interested in information received, exchanged, processed and used by people. We call it anthropic information. A relevant infological system IF(R) for anthropic information is the human brain. Thus, to further classify and study anthropic information, we need to utilize our knowledge about the structure and functions of the brain.

In our case, the most relevant is the Triune Brain model introduced and studied by Paul MacLean [33,34]. The main conception of his approach is existence of three levels of perception and action that are controlled by three corresponding centers of perception in the human brain. These three centers together form the Triune Brain and have the structure of a triad. It is natural to call the initial MacLean's structure by the name Anatomic Triune Brain model because it is based on the anatomy of the brain where three indispensable parts are distinguished: the neocortex, limbic system and R-complex.

According to the theory of MacLean, the neural basis, or framework, of the brain consists of three parts: the spinal cord, hindbrain, and midbrain. In addition to it, centuries of evolution have endowed people with three distinct cerebral systems. The oldest of these is called the reptilian brain or R-complex. It programs behavior that is primarily related to instinctual actions based on ancestral learning and memories, satisfying basic needs such as self-defense, reproduction and digestion. The reptilian brain is fundamental in acts such as primary motor functions, primitive sensations, dominance, establishing territory, hunting, breeding, and mating.

Through evolution, people have developed a second cerebral system, the limbic system, which MacLean refers to as the paleomammalian brain and which contains hippocampus, amygdala, hypothalamus, pituitary gland, and thalamus. This system is situated around the R-complex, is shared by humans with other mammals, and plays an important role in human emotional behavior.

The most recent addition to the cerebral hierarchy is called the neomammalian brain, or the neocortex. It constitutes 85% of the whole human brain mass and receives its information from the external environment through the eyes, ears, and other organs of senses. This brain component (neocortex) contains cerebrum, corpus callosum, and cerebral cortex. The cerebrum and cerebral cortex are divided into two hemispheres, while the corpus callosum connects these hemispheres. The neocortex deals with information in a logical and algorithmic way. It governs people creative and intellectual functions like social interaction and advance planning. The left hemisphere works with symbolic information, applying step-by-step reasoning, while the right hemisphere handles images processed by massively parallel (gestalt) algorithms.

Even psychologists who have objections to the Anatomic Triune Brain model admit that it is a useful, although oversimplified, metaphor, as the structure presented as the triune brain is based on a sound idea of three functional subsystems of the brain [16,35]. In the development of neurophysiology and neuropsychology, MacLean's theory was used as a base for the Whole Brain model, developed by Herrmann [36]. The main idea of this development is a synthesis of the Anatomic Triune Brain model with the two-hemisphere approach to the brain functioning.

The theory of the triune brain (reptilian, old mammalian and new mammalian) is used as a metaphor and a model of the interplay between instinct, emotion, and rationality in humans. Cory [37] applied this model to economic and political structures. In Cory's schema, the reptilian brain mediates the claims of self-interest, whereas the old mammalian brain mediates the claims of empathy. If selfish interests of an individual are denied for too long, there is discontent due to a feeling of being unjustly treated. If empathic interests are denied for too long, there is discontent due to guilt. In either case, the center of intelligence at the prefrontal cortex plays the role of a mediator. Its executive function is required to restore balance, generating the reciprocity required for effective social and economic structures.

The Triune Brain model is used to explain hyperactivity of youngsters studied by Zametkin [38] and other researchers. Peter Levine bases his approach to trauma treatment on the Triune Brain model [39]. According to Levine, there are tree types of the uniform stress and relaxation responses to a threatening situation that are active in all animal species through the autonomic nervous system. In the everyday language, these responses are metaphorically called fight, flight or freeze. The first two of them are well-known, while the third one was introduced by Levine. The freezing or immobility response has evolved over millions of years and it has served an adaptive purpose well for all species—except humans. In an individual, it can lead to trauma. Many physical ailments are actually residues of thwarted trauma reactions incurred during stressful events. What usually happens to nonhuman species is that after the threatening situation resolves itself, the animal forgets the stress and goes on its way without being traumatized.

In contrast to this, people can get stuck in the freezing response, while the reasoning mind resists or blocks the natural bodily sensations and fine motor movements needed to come out of the freeze response. The contemporary rationalistic culture is not helpful in supporting people in such a traumatizing situation. The feelings that people go through after experiencing a traumatic event are outside of their voluntary control, often being frightening and even potentially re-traumatizing.

Levine [39] postulates that trauma exists not in the event or in the story of the event, but is stored within the nervous system. The main principle of the Levine's treatment approach is that the body has a natural, innate, and miraculous capacity to heal once these reactions are understood and guided.

Although the Triune Brain has become a well-known model in contemporary psychology, it caused several objections on the ground of the development and structure of the triune brain system [16,40]. First, there is evidence that the, so-called, paleomammalian and neomammalian brains appeared, although in an undeveloped form, on much earlier stages of evolution than it is assumed by MacLean. Second, there are experimental data that in the neocortex, regions that are homological to the, so-called, paleomammalian and reptilian brains exist. For instance, neuropsyhological data give evidence that amygdala, which is a part of a limbic system, performs the low-level emotion processing, while the ventromedial cortex performs the high-level emotion processing. This shows that emotions exist, at least, on three levels: on the subconscious level of limbic system, on the conscious intuitive level, and on the conscious rational level in the cortex. The first level utilizes direct affective information, while the second and, to some extent, the third levels make use of cognitive emotional information [1].

At the same time, the development of the system of Will demands inclusion of some regions that are not included into the R-complex (the reptilian brain in MacLean's theory) into this system. It means that the centers of rational intelligence, emotion and will are not concentrated in three separate regions of the brain but are highly distributed among several components of the brain. Thus, it is better to call them not centers but systems of intelligence, emotion and will. This extension of functional characteristics results in the Componential Triune Brain model described in [1], making this model more adequate to experimental data and overcoming the pitfalls of the simple linear hierarchy suggested by MacLean [16,40].

Thus, the Componential Triune Brain model consists of three basic systems of the brain:

-

the System of Rational Intelligence (also called System of Reasoning) (SRI);

-

the System of Emotions (or more generally, of Affective States) (SAS);

-

the System of Will and Instinct (SWI).

All three systems of the brain are schemes in the sense of the schema theory, which is developed as a specific direction of the brain theory [41-44]. According to this theory, brain schemes are anatomically distributed in the brain and interact in a way of concurrent competition and coordination. All these interactions are based on physical processes but have an inherent informational essence related to a specific type of information. Information processes in the brain are more exactly reflected by the theory of the triadic mental information than by the conventional information theory that deals only with cognitive information.

In standard structuring of the brain, we also find these three systems. In the conventional setting (cf., for example, [40,45-47], the brain includes three components: the forebrain, midbrain, and hindbrain.

The forebrain is the largest division of the brain involved in a wide range of activities that make people human. The forebrain has a developed inner structure. It includes the cerebrum, which consists of two cerebral hemispheres. The cerebrum is the nucleus of the System (Center) of Rational Intelligence.

Under the cerebrum, is the diencephalon, which contains the thalamus and hypothalamus. The thalamus is the main relay center between the medulla and the cerebrum. The hypothalamus is an important control center for sex drive, pleasure, pain, hunger, thirst, blood pressure, body temperature, and other visceral functions. The forebrain also contains the limbic system, which is directly linked to the experience of emotion. The limbic system is the nucleus of the System (Center) of Emotions (or more generally, of Affective States).

The midbrain is the smallest division and it makes connections with the other two divisions—forebrain and hindbrain—and alerts the forebrain to incoming sensations.

The hindbrain is involved in sleeping, waking, body movements and the control of vital reflexes such as heart rate, blood pressure. The structures of the hindbrain include the pons, medulla and cerebellum. The hindbrain is the nucleus of the System (Center) of Will and Instinct.

The System of Rational Intelligence realizes rational thinking. It includes both symbol and image processing, which go on in different hemispheres of the brain [48]. The System of Emotions governs sensibility and the emotional sphere of personality. The System of Will and Instincts directs behavior and thinking. Two other systems influence behavior only through the will. For instance, a person can know that it is necessary to help others, especially, those who are in need and deserve helping. However, in many cases, this person does nothing without a will to help. In a similar way, we know situations when an individual loves somebody but neither tells this nor explicitly shows this due to an absence of a sufficient will.

It is necessary to remark that discussing the will of an individual we distinguish conscious will, unconscious will, and instinct. All of them are controlled by the Center of Will and Instincts (SWI). In addition, it is necessary to make distinctions between thoughts about intentions to do something and the actual will to do this. Thoughts are generated in the Center of Rational Intelligence, while the will dwells in the Center of Will and Instincts. In other words, thoughts and words about wills, wishes, and intentions may be deceptive if they are not based on a Will.

Will is a direct internal injunction, as well as any kind of motivation [19]. That is, the forces that act on or within an organism to initiate and direct behavior, has to be transformed into a will in order to cause the corresponding action. Will is one of the basic components of different models of personality. For instance, in psychosynthesis, which is a holistic transpersonal psychology and philosophy of life developed by Roberto Assagioli, a contemporary of both Freud and Jung, the Will is the basic component of the self [18].

Assagioli explicated essential properties of the Will—it can be assertive, aggressive, and controlling. In addition, there are three categories of Will: the accepting Will, yielding Will, and dedicated Will.

Thus, we can see that the Componential Triune Brain model corresponds to the three basic functions of the brain:

-

Reasoning as symbolic information processing and information processing of images.

-

Emotions and feelings as affective states of the brain.

-

Will and instinct as forces of the psyche.

The Anatomic Triune Brain model of MacLean also corresponds to the basic functions of the brain but misses one of them, namely, the Will.

Many other psychological theories and psychiatric tecniques consider the Will as a primary factor of human behavior and dispositions [19,49].

Often the Will is considered as a process that deliberates on what is to be done [50].

The Componential Triune Brain model is not only a necessary extension of the Triune Brain model but it also continues a long standing approach to the brain stratification. As Smith [35] demonstrates, triadic models of the brain and psyche have featured through two and half millennia of Western thought, starting with works of Pythagoras, Plato and Aristotle and receiving a modern airing in Paul MacLean's the Triune Brain model. A generation later after Pythagoras, Plato and Aristotle, Herophilus and Erasistratus from Alexandrian put together a more anatomically informed triadic theory, which was modified by Galen in the 2nd century and remained the prevailing paradigm for nearly fifteen hundred years until it was overturned by the great thinkers of the Renaissance. Nonetheless, the notion that the human neuropsychological system is somehow best thought of as having a triadic (tripartite) structure has remained remarkably resilient and has reappeared time and again in modern and early modern times. For instance, the Triune Brain model well correlates with the Freud's model of personality, which has the structure of the triad (Id, Ego, Super-ego). In the correspondence between the Triune Brain model and the Freud's model, the reptilian complex corresponds to Id, the limbic system corresponds to Ego, and the neocortical complex corresponds to Super-ego. In the context of triadic models, it is also possible to consider the triarchic theory of intelligence developed by Robert Sternberg [51].

Taking each of these the centers, SRI, SAS, and SWI, as a specific infological system, we find three types of information. One is the conventional information that acts on the center of reasoning and of other higher brain functions (SRI), which is situated in the neocortex. This information gives knowledge, changes beliefs and generates ideas. Thus, it is natural to call it cognitive information. Information of the second type acts on the system of emotions (SAS), which includes the paleomammalian brain. It is natural to call this information by the name direct emotional information, or direct affective information or emotive information. Information of the third type acts on the System of Will and Instinct (SWI), which contains the reptilian brain. It is natural to call this information by the name direct regulative or direct effective information.

Thus, anthropic information has three dimensions:

Cognitive information changes the content of the SRI, which includes the knowledge system (thesaurus) and neocortex (neomammalian brain) as its carrier.

Direct emotional/affective information changes the content of the SAS, which includes the paleomammalian brain (limbic system).

Direct regulative/effective information changes the content of the SWI, which includes the reptilian brain or R-complex.

However, in general, emotions constitute only one part of affective states, which also include moods, feelings, etc. That is why in general, direct affective information is more general than direct emotional information. However, as there is no consensus on the differences between emotions and affective states, these two types of information are used without differentiation.

Interactions between the basic brain systems imply dependencies between thinking, emotions, and actions of people. Emphasizing some of these relations, psychologists build their theories and psychotherapists develop their therapeutic approaches. Giving priority to the System of Rational Intelligence (SRI), the so-called “cognitive revolution” has taken hold around the world. It influenced both psychology, resulting in the emergence of cognitive psychology [52], and psychotherapy, inspiring creation of cognitive therapy [53]. In psychology, the word cognitive often means thinking in many contexts of contemporary life [54]. The cognitive therapeutic approach begins by using the extremely powerful reasoning abilities of the human brain. This is important because our emotions and our actions are not separate from our thoughts. They are all interrelated. Thinking (SRI) is the gateway to our emotions (SAS)—and our emotions are the gateway to our actions through motivation and will (SWI). This is only another way of saying that information from the System of Rational Intelligence (SRI) goes to the System of Emotions (SAS)—and from it to the System of Will and Instincts (SWI) that controls our actions. Consequently, the cognitive psychotherapeutic approach, which has been successfully utilized for treating many mental disorders, gives additional supportive evidence for the theory of the triune brain and behavior, as well as for the theory of the triadic mental information. The latter explains that while going from the System of Rational Intelligence to the System of Emotions and to the System of Will and Instincts, information is transformed from cognitive information, to direct emotional/affective information to direct effective/regulative information. As a result, we have Interaction of the personality components presented in Figure 1.

We can see that cognitive information is one of the three basic types of anthropic information with the corresponding infological system CIF(R), which contains (stores and processes) cognitive elements or constituents, such as knowledge, data, ideas, beliefs, images, algorithms, procedures, scenarios, ideas, values, goals, ideals, fantasies, abstractions, etc. Cognitive infological systems are very important, especially, for intelligent systems as the majority of researchers believe that information is intrinsically connected to cognition. This peculiarity is reflected in the following Cognitive Transformation Principle.

Ontological Principle O2c (the Cognitive Transformation Principle)

Cognitive information for a system R, is a capacity to cause changes in the cognitive infological system CIF(R) of the system R.

In the case of a cognitive infological system CIF(R), it looks like it is possible to give an exact definition of cognitive information. However, now cognitive sciences do not know all structural elements involved in cognition. A straightforward definition specifies cognition as activity (process) that gives knowledge. At the same time, we know that knowledge, as a rule, comes through data and with data. So, data are also involved in cognition and thus, have to be included in cognitive infological systems. In addition, cognitive processes employ such structures as ideas, images, texts, beliefs, values, measures, problems, schemas, procedures, tasks, goals, etc. Thus, to comprehensively represent cognitive information, it is imperative to include all such objects in cognitive infological systems.

As the cognitive infological system contains knowledge of the system it belongs, cognitive information is the source of knowledge changes.

There are also different types of cognitive information. One approach to classification is based on the structures in the brain. Researchers found that a longitudinal fissure separates the human brain into two distinct cerebral hemispheres, connected by the corpus callosum. The sides resemble each other and each hemisphere's structure is generally mirrored by the other side. Yet despite the strong similarities, the functions of each cortical hemisphere are different. As it is known (cf., for example, [55]), the left hemisphere operates with symbolic information representation, performing logical and analytical functions, such as linear reasoning, numeric manipulation, language processing, and mental arithmetic. It is also supposed that the left hemisphere works in the sequential mode on the level of consciousness. The right hemisphere processes images and realizes intuitive, creative and synthesizing functions of the brain, such as processing of visual and audiological stimuli, spatial manipulation, facial perception, and artistic performance. It is also supposed that the right hemisphere works in the parallel (concurrent) mode on the level of consciousness. In a similar way, Herrmann [36] differentiates functioning of the left parts and right parts of both the cerebral and limbic regions of the brain.

Taking each of these hemispheres as an infological system, we come to two types of cognitive information.

Symbolic (discrete) cognitive information transforms (or can transform) the symbolic content of the left hemisphere.

Holistic (continuous) cognitive information transforms (or can transform) the integral (gestalt) content of the right hemisphere.

This classification of cognitive information is complementary to another classification, in which one of the basic types of cognitive information is epistemic information, which can be both symbolic and holistic, belonging to both, the right hemisphere and left hemisphere. It is studied in Section 5.

4. Stratification of Knowledge Systems

There are different kinds of stratification.

In physical stratification, each stratum is a separate physical system. Any distributed database is physically stratified.

In analytical stratification, each stratum is determined by a specific name (label) and all elements from this stratum have this label (name). Knowledge base stratification used for handling inconsistent knowledge bases [56,57], for constructing models of a knowledge base [58] and for merging multiple knowledge bases [59,60] is analytical. The same is logic stratification used for formalization of commonsense reasoning [61]. In this section, we are mostly interested in analytical stratification based on knowledge classification. Different classes of knowledge form corresponding strata of knowledge systems.

There are different principles of knowledge classification, which allow us to build several types of knowledge system stratifications.

Time is an important characteristic of knowledge, giving different stratifications.

The temporal stratification.

  • The past stratum of knowledge consists of knowledge obtained/accepted in the past.

  • The current stratum of knowledge consists of the actual (used now) knowledge.

  • The future stratum of knowledge.

For instance, the knowledge “the Earth is flat” is past knowledge, while the knowledge “the Earth is round” is current knowledge.

The past stratum of knowledge consists of three substrata: the forgotten past knowledge, outdated but preserved past knowledge and still actual past knowledge.

The current stratum of knowledge consists of three substrata: the disappearing current knowledge, consolidated current knowledge, and emergent current knowledge.

The future stratum of knowledge consists of three substrata: the tentative/potential future knowledge, realizable future knowledge and emergent future knowledge.

More precise temporal stratifications are used in temporal the knowledge and databases. A temporal knowledge/database is a database with built-in time aspects. In particular, it supports a temporal knowledge/data model and has a temporal version of the query language [62,63]. Temporal knowledge/data stored in a temporal knowledge/database are different from the knowledge/data stored in non-temporal knowledge/database in that a time coordinate is attached to the knowledge/data. This is different from the conventional knowledge/data, which are usually considered to be valid now. Past and future knowledge/data are not stored. Usually past knowledge/data are modified, overwritten with new (updated) knowledge/data or deleted to achieve their temporal relevancy. Future knowledge/data are not considered because it is assumed that we do not receive information about the future.

There are many complexity measures of algorithms, methods and procedures. Taking a complexity measure C, it is possible to partition all algorithms (methods or/and procedures) into separate classes that have different complexity measures. Each such a partition induces a corresponding stratification of knowledge with respect to such knowledge characteristics as accessibility, inference, and generation, which are specific forms of knowledge acquisition. Here are some examples of such stratifications.

The accessibility stratification.

  • Directly accessible knowledge.

  • n-step accessible knowledge.

Another stratification is based on complexity of knowledge inference.

The inference stratification.

  • Directly implied knowledge.

  • n-step inferable knowledge.

One more stratification is based on complexity of knowledge generation.

The generation stratification.

  • Directly generable/computable knowledge.

  • n-step generable/computable knowledge.

Steps in generation, inference and access may be determined by:

-

Time slicing when each step is assigned some period of time for realization.

-

Elementary operations.

For instance, it is possible to assume that knowledge acquisition is direct if it demands less than 3 seconds. The first step of knowledge acquisition can be estimated as an interval from 3 seconds to 30 seconds. The second step of knowledge acquisition can be estimated as an interval from 30 seconds to 1 minute. The third step of knowledge acquisition can be estimated as an interval from 1 minute to 3 minutes and so on.

It is also possible to measure complexity, e.g., effort in generation, by the power of algorithms [64]. In this case, we have an algorithmic ladder, which consists of classes of algorithms with increasing computing power.

The traditional algorithmic ladders have one of the following forms:

  • Finite automata, deterministic pushdown automata, pushdown automata, and Turing machines.

  • Regular, or linear grammars, context-free grammars, context-sensitive grammars, and unrestricted, or phrase-structure grammars.

New achievements of the theory of algorithms and computation extend these ladders:

(1’)

Finite automata, deterministic pushdown automata, pushdown automata, Turing machines, inductive Turing machines [65], and infinite-time Turing machines [66].

(2’)

Regular, or linear grammars, context-free grammars, context-sensitive grammars, unrestricted, or phrase-structure grammars, grammars with prohibition [67], and Boolean grammars [68].

Inductive Turing machines give an example of an algorithmic ladder. Namely, the n-th strata of the inductive algorithmic ladder consists of inductive Turing machines with the structured memory that have order n but do not have order n + 1 [65]. It is also possible to build an algorithmic ladder using inductive or limit) Turing machine have structured program (rules for computation) or structured (heads) operating devices [65].

5. Stratified M-spaces and Information Operators

Let us consider a universal set or multiset W of knowledge items (units). It is possible to take the set WC of elementary knowledge units mathematically modeled in [1,69] as a universal set (multiset) W, obtaining a reasonable formalization of the concept of a knowledge state. Another possibility for W is realized by the set (multiset) WL of propositions and/or predicates from some logical language L. This logical approach was adopted in works of Bar-Hillel and Carnap [2], Hintikka [3-5] and some other authors. Shreider [7] interpreted knowledge items as texts in a thesaurus. Many researchers employ schemas as knowledge items in the brain (cf., for example, [41-44]. One more possibility for W is the set, or more exactly, a multiset, WS of situations possible in a world U, which are taken as knowledge items or knowledge units (cf., for example, [70,71].

The set W is called universal because we assume that the following axiom is true.

MA1 (the Internal Representation Axiom)

For any cognitive system (agent) A, knowledge states KAi of A are subsets (submultisets) of the set (multiset) W.

It is possible to interpret W as the base of all knowledge that agents are able to have about their environment.

Another aspect of universality of the set (multiset) W may be in the possibility to describe all possible (existing) worlds utilizing knowledge only from W. For instance, when W is the set (multiset) WL of propositions and/or predicates from some logical language L, then it is possible to build all descriptions of all possible worlds by combining elements from WL . This possibility is reflected in the following axiom.

MA2 (the External Representation Axiom)

For any environment (situation) D, there is a subset (submultiset) WD of the set (multiset) W that contains all accessible knowledge about D.

Taking these two axioms as the foundation, we develop a theory of cognitive systems/agents called the theory of M-spaces.

Definition 5.1 [1]

  • Subsets of W are called Mizzaro spaces.

  • Submultisets of W are called Mizzaro multispaces.

In some cases, only specific subsets (submultisets) of W are used in the theory. For instance, if elements of W are propositions and the model satisfies conditions of classical logical calculi, then only consistent subsets of propositions are acceptable as Mizzaro spaces. When, in addition, all deducible propositions are also included in such a logical Mizzaro space, then Mizzaro spaces are components of a logical variety [72].

Taking a knowledge system K, we model the states of K by Mizzaro spaces (Mizzaro multispaces), i.e., each knowledge state is represented by a Mizzaro space (Mizzaro multispace) KKi. The whole knowledge system K is modeled by an M-space.

In modeling knowledge systems and information processes, we consider two structures – sets and multisets – because using the classical background it is possible to consider only sets, which make the model simpler. However, many real cognitive systems contain several copies of the same element. For instance, the same element of knowledge can be stored in different parts of a computer memory or of the brain. This makes utilization of multisets necessary.

Definition 5.2

A type of structures is a system of conditions (axioms) that all these structures, i.e., sets with relations, satisfy.

To define an M-space, we consider a type θ of structures in Mizzaro spaces (Mizzaro multispaces).

Definition 5.3

An M-space (M-multispace) M is a structure of the form

M = { KS M ; OS M }
where KS consists of Mizzaro spaces (Mizzaro multispaces) K with the structure of the type θ, and KSM is a system of information operators in OS acting on Mizzaro spaces (Mizzaro multispaces).

Thus, KSM = {Ki|iI} and OSM = {At|tT}.

Example 5.1

It is possible to represent a logical variety or a prevariety V [72] as an M-space where KSM consists of the components of V and operators from OSM apply mappings fi: AiL and gi: TiL (iI), which form connections between components of the variety (prevariety).

Definition 5.4

  • The set KSM is called the state space of the M-space M.

  • The set U = ∪iI Ki is called the universe of the M-space M.

    K i KS M

  • The system OS is called the operating system of the M-space M.

The simplest structure of Mizzaro spaces Ki of the type θ is the structure of a set and the simplest structure of Mizzaro multispaces Ki of the type θ is the structure of a multiset. However, Mizzaro spaces Ki can be logical calculi, linear spaces or groups.

In the algebraic context, each M-space M is a universal algebra [73] with the support KS and system of operations OS. In this paper, we consider only unary Mizzaro spaces (unary Mizzaro multispaces) in which each information operator maps one Mizzaro space (Mizzaro multispace) Ki into another (may be the same) Mizzaro space (Mizzaro multispace) Kj.

Information operators from OS are global epistemic information operators in KS. When an operator acts only on one Mizzaro space (Mizzaro multispace), then it is a local epistemic information operator. Local epistemic information operators in non-structured Mizzaro spaces are studied in [1,10,11,13].

Note that it is possible to consider any system that contains a knowledge system as a knowledge system itself. Thus, any intelligent agent is a knowledge system.

There are two basic types of epistemic information operators: content, bond and mixed operators.

Definition 5.5

A content epistemic information operator acts on knowledge items in a knowledge state.

For instance, all information operators studied in [1,10,11,13] are content epistemic information operators.

Definition 5.6

A bond epistemic information operator acts on connections (bonds or relations) between knowledge items in a knowledge state.

Such operators as interpretation and reinterpretation of information/knowledge items are bond epistemic information operators.

Definition 5.7

A mixed epistemic information operator acts both on knowledge items and on connections (bonds or relations) between knowledge items in a knowledge state.

Operators of logical inference, such as rules of deduction, are mixed epistemic information operators act because they add new knowledge items in the form of propositions or/and predicates and establish relations of inferrability/deducibility between propositions or/and predicates.

Here we are mostly interested in content epistemic information operators, which we simply call epistemic information operators.

To correctly model stratified knowledge system, the modeling M-space also has to be stratified.

Definition 5.8

  • An M-space M = {KS; OS} is stratified if there is a set J and each Mizzaro space (Mizzaro multispace) Ki from KS has the form

    K i = j J K ij

  • A stratification of an M-space M = {KS; OS} is strict if for each Mizzaro space (Mizzaro multispace) Ki from KS, KijKik= ∅ when jkJ.

  • An M-space M = {KS; OS} is linearly stratified if each Mizzaro space (Mizzaro multispace) Ki from MK has the form

    K i = n = 1 K in

    in the case when the stratification is infinite and the form

    K i = n = 1 m K in

    in the case when the stratification is finite (cf. Figure 2).

Linear stratification means that the set of stratum indices J is finite or countable and linearly ordered.

There are M-space stratifications with a non-linear topology. For instance, the stratification in Figure 3 has a cyclic topology. Important cases of M-space stratifications have structures of a tree or a forest.

Example 5.2

People, as well as computers, have many kinds of memory. It is even supposed that each part of the brain has several types of memory agencies that work in somewhat different ways, to suit particular purposes [23]. It is possible to consider each of these memory agencies as a separate system and to study differences between information that changes each type of memory. This might help to understand the interplay between stability and flexibility of the mind, in general, and memory, in particular.

Psychologists differentiate three types of human memory: sensory memory, short-term memory, and long-term memory. It is the most important and best documented by scientific research memory stratification. However, memory researchers do not employ uniform terminology. Sensory memory is also known as sensory register, sensory store, sensory information storage, eidetic memory and echoic memory. Short- and long-term memories are also referred to as primary memory and secondary memory, correspondingly. Each component of memory differs with respect to its function, the form of information held, the length of time information is retained, and the amount of information-handling capacity.

Thus, human memory has three basic strata. As a result, all knowledge in the memory of an individual is also stratified into three components: knowledge/data in the sensory memory, knowledge in the short-term memory, and knowledge in the long-term memory of this individual (cf. Figure 4). Additional stratification of human memory as a knowledge space induces additional stratification of knowledge.

Sensory memory acts as a buffer for stimuli received through the senses, which are then filtered and passed from sensory memory into short-term memory by attention.

In tern, sensory memory is also stratified by different sensory channels. There are iconic memory for visual stimuli, echoic memory for aural stimuli and haptic memory for touch.

Long-term memory is naturally stratified. The most popular stratification divides it into two parts: episodic memory and semantic memory. Episodic memory stores knowledge of events and experiences in a serial form. In contrast to this, semantic memory is a structured record of facts, concepts and skills that people have acquired. The information in semantic memory is derived from that in the episodic memory of the same individual.

Neuroscientists distinguish three main activities related to long term memory: storage, deletion and retrieval. These operations are modeled by epistemic information operators. Storage is modeled by the epistemic information operator REPL. In information storage, information from sensory memory, at first goes to short-term memory and then is stored in long-term memory, usually by the process called rehearsal. Rehearsal is the repeated exposure to a stimulus of a knowledge/data portion, which transfers it into long-term memory. Deletion of a knowledge/data portion is modeled by the epistemic information operator DEL and is mainly caused by decay and interference. Emotional factors essentially affect the long-term memory functioning. According to contemporary there are two types of information retrieval: recall and recognition. In knowledge/data recall, the information is reproduced from memory. Recall is modeled by the epistemic information operator COPY. Knowledge/data recognition is based on information that this knowledge/data portion has been seen before and is assisted by the provision of retrieval cues to enable better access in the long-term memory. Recognition is modeled by the epistemic information operator GEN.

Scientists also use another stratification of the human memory: personal memory, semantic memory, perceptual memory, and skill memory, which includes, motor skill memory, cognitive skill memory, and rote linguistic skill memory.

Example 5.3

The computer memory is also a complex system of diverse components and processes. Memory of a computer includes such three basic components as the random access memory (RAM), read-only memory (ROM), and secondary storage. While RAM forgets everything whenever the computer is turned off and ROM cannot learn anything new, secondary storage devices allow the computer to record information for as long period of time as we want and change it whenever we want. Now the following devices are utilized for log-term computer memory: magnetic tapes and corresponding drives, magnetic disks and corresponding drives, flash memory storage devices and corresponding drives, and optical disks and corresponding drives.

Computer memory is intrinsically stratified by the hierarchy in which levels are distinguished by the response time, complexity, and capacity. The overall goal of using a memory stratification is to obtain the higher possible average access performance, while minimizing the cost of the entire memory system.

Often four major memory levels are separated:

  • Internal memory – Processor registers and cache.

  • Main memory – the system RAM and controller cards.

  • On-line mass storage – Secondary storage.

  • Off-line bulk storage – Tertiary and Off-line storage.

At the same time, other experts use another stratification of computer memory:

  • Processor registers have the fastest possible access (in just a few cycles) and usually stores tens of kilobytes

  • Level 2 (L2) cache usually stores 512 KiB or more

  • Level 3 (L3) cache usually stores 2048 KiB or more

  • Main memory, access to which may take hundreds of cycles, but which usually stores multiple gigabytes.

  • Disk storage, access to which may take millions of cycles latency if not cached, but which usually is very large

  • Tertiary storage, access to which may take several seconds latency and which can be huge

Such memory stratifications are linear, reflecting the access time with the fast CPU registers at the top and the slow hard drive at the bottom.

Another, although related, stratification is induced by different electronic devices, which include CPU registers, on-die SRAM caches, external caches, DRAM, paging systems, and virtual memory or swap space on a hard drive. These devices are called RAM (random access memory) by many developers, even though the various subsystems can have very different access times, violating the original concept behind the random access term in RAM. RAM consists of two strata: dynamic random access memory, which requires the stored information to be periodically re-read and rewritten, or refreshed in order not to lose it, and static memory, never needs to be refreshed as long as power is applied, although it can lose its content if power is removed.

Usually each stratum in RAM is also stratified. For instance, in DRAM, strata are defined by the row, column, bank, rank, and channel.

In addition, computer memory is also stratified by data storage technologies, such as semiconductor, magnetic, and optical technologies. In modern computers, primary storage almost exclusively consists of dynamic volatile semiconductor memory, which uses semiconductor-based integrated circuits to store information. Since the turn of the century, a type of non-volatile semiconductor memory known as flash memory has steadily gained share as off-line storage in various advanced electronic devices and computers.

Magnetic storage, which is non-volatile, uses different types of magnetization on a magnetically coated surface to store information. The information is accessed using one or more read/write heads which may contain one or more recording transducers. A read/write head only covers a part of the surface so that the head or medium or both must be moved relative to another in order to access data. In modern computers, there are following kinds of magnetic storage devices:

  • Magnetic disk, such as sloppy disks, used for off-line storage, and the hard disk drive, used for secondary storage

  • Magnetic tape data storage, used for tertiary and off-line storage

At the beginning of computer era, magnetic storage was also used for primary storage in a form of a magnetic drum, or core memory, core rope memory, thin-film memory, twistor memory or bubble memory, while magnetic tapes were often used for secondary storage.

Another popular type of storage is optical discs, which stores information in deformities on the surface of a circular disc, reading this information by illuminating the surface with a laser diode and observing the reflection. In modern computers, there are following kinds of optical storage devices:

  • CD, CD-ROM, DVD, BD-ROM: Read only storage, used for mass distribution of digital information (music, video, computer programs)

  • CD-R, DVD-R, DVD+R, BD-R: Write once storage, used for tertiary and off-line storage

  • CD-RW, DVD-RW, DVD+RW, DVD-RAM, BD-RE: Slow write, fast read storage, used for tertiary and off-line storage

  • Ultra Density Optical or UDO is similar in capacity to BD-R or BD-RE and is slow write, fast read storage used for tertiary and off-line storage.

In magneto-optical disc storage, the information is read optically and written into the magnetic state on a ferromagnetic surface by combining magnetic and optical methods. It is usually used for tertiary and off-line storage.

Paper data storage, typically in the form of paper tape or punched cards, has long been used to store information for automatic processing, particularly before electronic computers were invented.

There are also such memory devices as the vacuum tube memory, electro-acoustic memory, optical tapes, phase-change memory, holographic data storage, and molecular memory, which stores information optically inside crystals or photopolymers.

All these memory devices and components determine the corresponding stratification of knowledge stored in computers, e.g., of knowledge bases.

Example 5.4

Stratification is a popular technique in knowledge base theory and practice. For instance, Hunter and Liu [59] introduce knowledge base stratification to solve the problem of merging multiple knowledge bases. Benferhat and Baida [56] use stratified first order logic for access control in knowledge bases. Benferhat and Garcia [57] employ stratification for handling inconsistent knowledge bases. Lassez, et al, [58] show how stratification can be used as a tool in the interactive model-building process. Namely it is possible to reduce the computational complexity of the process by the use of stratification, which limits consistency checking to minimal strata.

Definition 5.9

  • An M-space M is a subspace of an M-space H if the state space KS is a substructure of the state space KSH and the operational system OS is a substructure of the operational system OSH.

  • If an M-space M is a subspace of an M-space H, then H is called a superspace of M.

In particular, the stratification of the knowledge space KS is induced by the stratification of the knowledge space KSH.

Example 5.5

If a structured M-space H models the group memory of a group G of several people, then a structured M-space M that models the memory of one individual from this group G is a subspace of H.

Subspaces of M-spaces and M-multispaces represent subsystems of knowledge systems. For instance, in large knowledge systems, such as a scientific theory, it is possible to separate the subsystem of denotational knowledge and the subsystem of operational knowledge.

Definition 5.10

If X is a structure, i.e., a set/multiset with relations, then X is the set of all elements from X and X is the multiset of all elements from X, while RelX is the set of all relations from X.

In such a way, ignoring the M-space stratification, it is possible to represent structured knowledge systems by uniform M-spaces, in which all knowledge states are sets or multisets. In this setting, content epistemic information operators act on elements from sets K or multisets K, while bond epistemic information operators act on elements from RelK.

Definition 5.11

  • A knowledge system (agent) A is called locally finite if any knowledge state of A is finite.

  • A knowledge system (agent) A is called finite if it has only a finite number of knowledge states and any knowledge state of A is finite.

  • An M-space M is called locally finite if each K from KS contains only a finite number of knowledge items.

  • An M-space M is called finite if it has only a finite number of Mizzaro spaces (Mizzaro multispaces) K each of which is also finite.

It looks like it might be sufficient to consider only finite or at least, locally finite agents. However, if knowledge is represented by logical statements and it is assumed (as, for example, in the theory of Bar-Hillel and Carnap [2]) that any knowledge system contains all logical consequences of all its elements, then an agent with such knowledge system is infinite. In information algebras, portions of information are represented by close subsets of sentences from a logical language L [74].

However, in conventional logics, closed with respect to such information operators as deduction sets are infinite because any sentence p implies pq for any sentence q from L, which is, as a rule, infinite (cf., for example, [75]). Thus, in the context of classical logic and information algebras any portion of information has infinitely many representations. Consequently, such a portion generates a system with the infinite number of knowledge items.

Lemma 5.1

An M-space M is finite if and only if it has a finite universe.

Stratification of the knowledge system and the corresponding M-space allows defining specific classes of epistemic information operators.

Definition 5.12

An epistemic information operator A is called stratified if for any jJ, there is kJ such that for any Ki from KS, we have A(Kij) ⊆ Kik.

Stratified information operators preserve the structure, i.e., stratification, of knowledge states. Note that adding and deletion operators are intrinsically stratified.

Definition 5.13

  • An epistemic information operator A is called closed if for any jJ and for any Ki from KS, we have A(Kij) ⊆ Kij.

  • An epistemic information operator A is called closed in a Mizzaro space Ki from KS if A(Ki) ⊆ Ki.

Lemma 5.2

Any closed epistemic information operator A is stratified.

Definition 5.14

  • A stratified epistemic information operator A in a linearly stratified M-space M is called monotone (antitone) if for any nN, there is kN such that kn (kn) and for any Ki from KS, we have A(Kij) ⊆ Kik.

  • A stratified epistemic information operator A in a linearly stratified M-space M is called strictly monotone (strictly antitone) if for any nN, there is kN such that k > n (k < n) and for any Ki from KS, we have A(KijKik.

Definitions imply the following property of epistemic information operators.

Lemma 5.3

Any strictly monotone (strictly antitone) epistemic information operator A is monotone (antitone).

Lemma 5.4

In a finite linearly stratified M-space M, there are no strictly monotone and strictly antitone information operators.

Definition 5.15

An epistemic information operator A is called contracting if there is kJ such that for any Ki from KS, we have A(Kij) ⊆ Kik.

Definitions imply the following result.

Lemma 5.5

Any contracting epistemic information operator A is stratified.

There are five types of basic epistemic operations: adding, deleting, moving, replicating, and ttransforming knowledge, and five types of corresponding basic epistemic information operators: adding AD, deletion DL, moving MV, replication REPL, generation GEN and transformation TR epistemic information operators.

Definition 5.16

A transformation epistemic information operator TR takes a group of knowledge items (may be, one item) from the current knowledge state and transforms it into another group of knowledge items (may be, into one item).

Definition 5.17

A generation epistemic information operator TR takes a group of knowledge items (may be, one item) from the current knowledge state and generates another group of knowledge items (may be, one item).

The difference between transformation and generation is that in generation the initial group of knowledge items is preserved, while in transformation it is not preserved.

Lemma 5.4

AD is equal to GEN with the empty set of the initial knowledge items.

Definition 5.18

A moving epistemic information operator MV moves a knowledge item from one stratum into another one.

Definition 5.19

  • A replica of a knowledge item is another knowledge item equivalent to the initial one.

  • A replication epistemic information operator REPL makes a replica of a knowledge item and adds it to the current knowledge state.

Example 5.5

Let us consider logical knowledge representation in which knowledge items are propositions. Then according to laws of logic there are equivalent propositions. For instance, taking the proposition (1) “B implies A”, we have equivalent propositions (2) “A follows from B”, (3) “If B, then A”, and (4) “A is a consequence of B”. All of them are replicas of one another although they are not copies.

If the proposition (1) belongs to the stratum K1, then its replication to the stratum K2 can introduce either proposition (1) or proposition (2) or proposition (3) to the stratum K2, while its copying to the stratum K2 can introduce only proposition (1) to the stratum K2.

An important special case of a replication epistemic information operator is a copying epistemic information operator COPY, which makes a copy of a knowledge item and adds it to the current knowledge state.

Another important special case of a replication epistemic information operator is a restricted replication epistemic information operator REPL0, which replicates a knowledge item and adds it only to a stratum of the current knowledge state that does not have the same replica. One of its special cases is a restricted copying epistemic information operator COPY0, which makes a copy of a knowledge item and adds it only to a stratum of the current knowledge state that does not have the same replica. Operators REPL0 and COPY0 are used in stratified M-spaces not to make these spaces stratified M-multispaces.

Lemma 5.5

The operator COPY0 can copy a knowledge item only to a different stratum, i.e., if aKi and COPY0aKj, then ij.

Indeed, if this condition is violated, then the initial M-space is converted to an M-multispace.

Complex information operations and operators are studied in [76].

Definition 5.20

An epistemic information operator C is called the sequential composition of an epistemic information operator A with an epistemic information operator B if C(x) is defined and equal to B(A(x)) when: 1) A(x) is defined and belongs to the domain of B; 2) B(A(x)) is defined. Otherwise, C gives no result being applied to x, i.e., C(x) = *.

Taking sequential composition of an epistemic information operator A with itself, we obtain sequential powers An of the operator A.

In the general case, the sequential composition of epistemic information operators is not commutative in M-spaces as the following example demonstrates.

Example 5.1

Let us consider a structured M-space M = {KS; OS} where KS = = ∪iI KS. In this space, the operator MVija moves an element a from the stratum KSi to the stratum KSj and does not change other elements from KS. Taking the sequential composition of such operators, we have

MV ij a MV ik a = MV ij a MV ik a MV ij a = MV ik a
if ij, kj, and ik. Thus, operators MVija do not commute with one another.

At the same time, all these operators are idempotents, i.e., MVijaMVija = MVija

It is necessary to remark that in a structured M-multispace M with an infinite number of elements a in each stratum KSi, operators MVija and MVika commute with one another. This demonstrates difference between M-spaces and M-multispaces.

Proposition 5.1

If A and B are closed (closed in a Mizzaro space Ki) operators, then their sequential composition AB is also a closed (closed in a Mizzaro space Ki) operator.

Indeed, if A and B are closed epistemic information operators in a structured M-multispace M, then for any jJ and for any Ki from KS, we have A(Kij) ⊆ Kij and B(Kij) ⊆ Kij. Thus, (A ∘ B)(Kij) = B(A(Kij) ⊆ B(Kij) ⊆ Kij.

For closed in a Mizzaro space Ki operators, the proof is similar.

Proposition 5.2

If A and B are contracting operators, then their sequential composition AB is also a contracting operator.

Proof is similar to the proof of Proposition 5.1.

Proposition 5.3

If A and B are stratified operators, then their sequential composition AB is also a stratified operator.

Proof is similar to the proof of Proposition 5.1.

Proposition 5.4

If A and B are (strictly) monotone [antitone] operators, then their sequential composition AB is also a (strictly) monotone [antitone] operator.

Indeed, if A and B are monotone epistemic information operators in a structured M-space M, then for any Ki from KS, we have A(Kij) ⊆ Kik with kj and B(Kik) ⊆ Kih with hk. Thus, (A ∘ B)(Kij) = B(A(Kij) ⊆ B(Kik) ⊆ Kih with hj.

Considerations for strictly monotone, antitone and strictly antitone epistemic information operators are similar.

Let us consider an M-space M with a finite linear stratification.

Proposition 5.5

For any monotone and any antitone epistemic information operator A, there is a number n such that the sequential power An is also a closed epistemic information operator.

Indeed, if A is a monotone epistemic information operator in a structured M-space M, then making each step, it either increases the number of the stratum or the image of a stratum remains in the same stratum. If the second case is true for all strata of M, then A itself is a closed epistemic information operator. Otherwise, A can increase the number of a stratum only for a finite number of steps because there are only a finite number of strata in M. Thus, after some number of repetitions, the image of a stratum remains in the same stratum. Taking the largest number of such steps, we obtain the necessary number n.

Note that n cannot be larger than the number of strata in M.

Let us explore relations between basic epistemic information operators.

Definition 5.21 [77]

Two operators A and B are functionally equivalent if they have the same definability domain D and A(x) = B(x) for any element x from D.

Proposition 5.6

A transformation epistemic information operator TR is functionally equivalent to the sequential composition of a deletion epistemic information operator DEL and adding epistemic information operator AD that act in the same stratum of the M-space.

Indeed, if TR takes items a1, a2, …, anfrom KS and transforms them into b1, b2, …, bm, then it is possible to achieve the same result by deleting a1, a2, …, an and adding b1, b2, …, bm to the corresponding stratum of KS.

Proposition 5.7

A moving epistemic information operator MV is functionally equivalent to deletion of a knowledge item in one stratum and adding the same knowledge item to another stratum.

Proposition 5.8

A replication epistemic information operator REPL is functionally equivalent to adding an equivalent knowledge item to the corresponding stratum.

Proposition 5.9

For any M-space M, there is a superspace H, in which all deletion and adding epistemic information operators DEL and AD in M are functionally equivalent to moving epistemic information operators MV in H.

Proof

To build a superspace H with the necessary properties, we add one more stratum E called the external stratum to the initial M-space M. In addition, we assume that E contains all elements from the universal set (multiset) W and each element has infinitely many copies in E. In this case, any deletion of an element a from a state K from M is equivalent to moving the same element a to the stratum E. In a similar way, any addition of an element a to a state K from M is equivalent to moving the same element a from the stratum E to the state K.

Proposition is proved.

Proposition 5.10

A moving epistemic information operator MV can be (functionally) simulated by copy COPY and deletion DEL epistemic information operators.

Indeed, instead of moving a knowledge item a from a stratum Ki of a state K to a stratum Kj of a state K′, it is possible to copy a from Ki to Ki and then to delete this element from Ki.

Proposition 5.11

A generation epistemic information operator GEN is functionally equivalent to the sequential composition of a transformation epistemic information operator TR and adding epistemic information operator AD that act in the same stratum of the M-space.

Proof is similar to the proof of Proposition 5.10.

Proposition 5.12

A transformation epistemic information operator TR is functionally equivalent to the sequential composition of a generation epistemic information operator GEN and deletion epistemic information operator DEL that act in the same stratum of the M-space.

Proof is similar to the proof of Proposition 5.10.

Definition 5.22

A system B of epistemic information operators is an operator basis of an M-space M if any A from OS is a composition of elements from B.

Operator bases can be useful in many situations. For instance, knowing properties of operators from such a base and properties of compositions, we can find properties of other operators.

Assuming that all operators in an M-space M are compositions of basic epistemic information operators, we have the following results.

Proposition 5.13

  • {AD, DEL} is an operator basis of an arbitrary (stratified) M-space M.

  • {TR, MV} is an operator basis of an arbitrary (stratified) M-space M.

  • {TR} is an operator basis of an arbitrary (i.e., non-stratified) M-space M.

Proof is based on Propositions 5.7 - 5.12.

Proposition 5.14

  • In a stratified M-space M with the external stratum, {REPL0, DEL} is an operator basis.

  • In a stratified M-space M with the external stratum, {MV} is an operator basis.

  • In a stratified M-multispace M with the external stratum, {REPL, DEL} is an operator basis.

Proof is based on Propositions 5.7 - 5.12.

6. Conclusions

Based on the principles of the general theory of information, epistemic information is singled out as a kind of antropic information and modeled by the algebraic construction of M-spaces. M-spaces represent information dynamics by information operators acting in knowledge spaces. The main emphasis of this study is made on stratified knowledge spaces and algebras of epistemic information operators in such spaces.

Obtained results bring us to the following problems.

It is possible to consider not only knowledge but also beliefs as basic components of cognitive infological systems and call information that acts on such systems by the name plausible epistemic information.

Problem 1

Mathematically describe and study plausible epistemic information.

Problem 2

Study other types of M-space stratifications and operators in these spaces.

Problem 3

Study M-spaces in which knowledge items are elements of logics, e.g., propositions or predicates and stratification of which includes the structure of the corresponding logic, e.g., the propositional logic or the first-order predicate logic.

Problem 4

Study categories of M-spaces and functors between these categories.

Problem 5

Study operations with M-spaces.

In this paper, we studied only content epistemic information operators, while bond epistemic information operators, which act on connections and relations between knowledge items, are also very important.

Problem 6

Study bond epistemic information operators.

In this paper, we studied only sequential composition of epistemic information operators, while other types of composition are also very important.

Problem 7

Study other compositions of epistemic information operators.

In [1,13], information explications of epistemic information operators is studied in uniform Mizzaro spaces.

Problem 8

Explore information explications of epistemic information operators in stratified M-spaces and Mizzaro spaces.

Figure 1. Interaction between components of personality.
Figure 1. Interaction between components of personality.
Information 02 00697f1 1024
Figure 2. A finite M-space stratification with the linear topology.
Figure 2. A finite M-space stratification with the linear topology.
Information 02 00697f2 1024
Figure 3. A finite M-space stratification with the cyclic topology.
Figure 3. A finite M-space stratification with the cyclic topology.
Information 02 00697f3 1024
Figure 4. The human memory hierarchy.
Figure 4. The human memory hierarchy.
Information 02 00697f4 1024

References

  1. Burgin, M. Theory of Information: Fundamentality, Diversity and Unification; World Scientific: Singapore, 2010. [Google Scholar]
  2. Bar-Hillel, Y.; Carnap, R. Semantic information. Br. J. Philos. Sci. 1958, 4, 147–157. [Google Scholar]
  3. Hintikka, J. The Varieties of Information and Scientific Explanation. In Logic, Methodology and Philosophy of Science III; van Rootselaar, B., Staal, J.F., Eds.; North-Holland Publishing Company: Amsterdam, The Netherlands, 1968; pp. 311–331. [Google Scholar]
  4. Hintikka, J. Surface Information and Depth Information. In Information and Inference; Synthese Library, Humanities Press: New York, NY, USA, 1970; pp. 263–297. [Google Scholar]
  5. Hintikka, J. On Defining Information. Ajatus 1971, 33, 271–273. [Google Scholar]
  6. Israel, D.; Perry, J. What is Information? In Information, Language and Cognition; University of British Columbia Press: Vancouver, BC, Canada, 1990; pp. 1–19. [Google Scholar]
  7. Shreider, Y.A. On the semantic characteristics of information. Inf. Storage Retr. 1965, 2, 221–233. [Google Scholar]
  8. MacKay, D.M. Information, Mechanism and Meaning; The MIT Press: Cambridge, MA, USA, 1969. [Google Scholar]
  9. Brookes, B.C. The foundations of information science, pt. 1, Philosophical aspects. J. Inf. Sci. 1980, 2, 125–133. [Google Scholar]
  10. Mizzaro, S. On the Foundations of Information Retrieval. Proceedings of the Atti del Congresso Nazionale (AICA'96), Roma, Italy, 25–27 September 1996; pp. 363–386.
  11. Mizzaro, S. How many relevances in information retrieval? Interact. Comput. 1998, 10, 303–320. [Google Scholar]
  12. Gackowski, Z.J. What to teach business students in mis courses about data and information. Issues Informing Sci. Inf. Technol. 2004, 1, 845–867. [Google Scholar]
  13. Mizzaro, S. Towards a Theory of Epistemic Information. In Information Modelling and Knowledge Bases; IOS Press: Amsterdam, The Netherlands, 2001; Volume 12, pp. 1–20. [Google Scholar]
  14. Baer, N.; Zeidman, B. Measuring Software Evolution with Changing Lines of Code. Proceedings of the 24th International Conference on Computers and Their Applications (CATA-2009), New Orleans, LA, USA, 8–10 April 2009; pp. 264–170.
  15. Burgin, M. Fundamental Structures of Knowledge and Information; Academy for Information Sciences: Kiev, Ukraine, 1997. (in Russian) [Google Scholar]
  16. Gardner, R.; Cory, G.A. The Evolutionary Neuroethology of Paul MacLean: Convergences and Frontiers; Praeger: New York, NY, USA, 2002. [Google Scholar]
  17. Russell, P. The Brain Book; Penguin Books: London, UK, 1992. [Google Scholar]
  18. Assagioli, R. Psychosynthesis: A Collection of Basic Writings; Penguin Books: London, UK, 1993. [Google Scholar]
  19. Kenny, A. Action, Emotion and Will; Routledge: London, UK, 2003. [Google Scholar]
  20. Burgin, M. Information and transformation. Transformation 1, 48.
  21. Von Weizsäcker, C.F. Die Einheit der Natur; Deutscher Taschenbuch Verlag: Munich, Germany, 1974. [Google Scholar]
  22. Von Weizsäcker, C.F. Aufbau der Physik; Hanser: Munich, Germany, 1985; (Eglish translation: The Structure of Physics; Springer: Berlin, Germany, Heidelberg, Germany and New York, NY, USA, 2006). [Google Scholar]
  23. Minsky, M. The Society of Mind; Simon and Schuster: New York, NY, USA, 1986. [Google Scholar]
  24. Capurro, R.; Fleissner, P.; Hofkirchner, W. Is a Unified Theory of Information Feasible? In The Quest for a unified theory of information; Routledge: London, UK, 1999; pp. 9–30. [Google Scholar]
  25. Melik-Gaikazyan, I.V. Information Processes Reality; Nauka: Moscow, Russia, 1997; (in Russian, English summary). [Google Scholar]
  26. Jung, C. On Psychic Energy. In On the Nature of the Psyche; Princeton University Press: Princeton, NJ, USA, 1928/1960. [Google Scholar]
  27. Von Grot, N. Die begriffe der seele und der psychischen energie in der psychologie. Arch. fur Syst. Philos. 1898, IV. [Google Scholar]
  28. Colby, K. Energy and Structure in Psychoanalysis; Ronald: New York, NY, USA, 1955. [Google Scholar]
  29. Freud, S. The Standard Edition of the Complete Psychological Works of Sigmund Freud; Hogarth and the Institute of Psycho-Analysis: London, UK, 1954. [Google Scholar]
  30. Lieberman, H.R. Cognitive methods for assessing mental energy. Nutr. Neurosci. 2007, 10, 229–242. [Google Scholar]
  31. O'Connor, P.J. Mental energy: Assessing the mood dimension. Nutr. Rev. 2006, 64, S7–S9. [Google Scholar]
  32. Dahl, H. The panel on “Psychoanalytic Theory of the Instinctual Drives in Relation to Recent Developments”. J. Am. Psychoanal. Assoc. 1968, XVI, 613–637. [Google Scholar]
  33. MacLean, P.D. A Triune Concept of the Brain and Behavior; University of Toronto Press: Toronto, ON, Canada, 1973. [Google Scholar]
  34. MacLean, P.D. On the Origin and Progressive Evolution of the Triune Brain. In Primate Brain Evolution; Plenum Press: New York, NY, USA, 1982. [Google Scholar]
  35. Smith, C.U.M. The triune brain in antiquity: Plato, Aristotle, Erasistratus. J. Hist. Neurosci. 2010, 19, 1–14. [Google Scholar]
  36. Herrmann, N. The Creative Brain; Brain Books: Lake Lure, NC, USA, 1990. [Google Scholar]
  37. Cory, G.A. The Reciprocal Modular Brain in Economics and Politics: Shaping the Rational and Moral Basis of Organization, Exchange, and Choice; Kluwer Academic/Plenum Publishers: New York, NY, USA, 1999. [Google Scholar]
  38. Zametkin, A.J. Cerebral glucose metabolism in adults with hyperactivity of childhood onset. N. Engl. J. Med. 1990, 323, 1361–1366. [Google Scholar]
  39. Levine, P.A. Waking the Tiger: Healing Trauma; North Atlantic Books: Berkeley, CA, USA, 1999. [Google Scholar]
  40. Patton, P. One world, many minds: Intelligence in the animal kingdom. Sci. Am. 2008, 19, 72–79. [Google Scholar]
  41. Anderson, R.C. The Notion of Schemata and the Educational Enterprise. In Schooling and the Acquisition of Knowledge; Anderson, R.C., Spiro, R.J., Montague, W.E., Eds.; Lawrence Erlbaum: Hillsdale, NJ, USA, 1977. [Google Scholar]
  42. Arbib, M. Schema Theory. In The Encyclopedia of AI; Wiley-Interscience: New York, NY, USA, 1992; pp. 1427–1443. [Google Scholar]
  43. Armbruster, B. Schema theory and the design of content-area textbooks. Educ. Psychol. 1996, 21, 253–276. [Google Scholar]
  44. Burgin, M. Mathematical Schema Theory for Modeling in Business and Industry. Proceedings of the 2006 Spring Simulation Multi Conference (Spring Sim '06), Huntsville, AL, USA; 2006; pp. 229–234. [Google Scholar]
  45. Baars, B.J.; Gage, N.M. Cognition, Brain, and Consciousness: introduction to Cognitive Neuroscience; Elsevier Science/Academic Press: Amsterdam, The Netherlands, 2007. [Google Scholar]
  46. Carter, R. Mapping the Brain; Phoenix Books: Junction, VT, USA, 2003. [Google Scholar]
  47. DeArmond, S.J.; Fusco, M.M.; Dewey, M. Structure of the Human Brain: A Photographic Atlas; Oxford University Press: New York, NY, USA, 1989. [Google Scholar]
  48. Dehaene, S. Reading in the Brain: The Science and Evolution of a Human invention; Viking: New York, NY, USA, 2009. [Google Scholar]
  49. Berrios, G.E.; Gili, M. Will and its disorders. A conceptual history. Hist. Psychiatry 1995, 6, 87–104. [Google Scholar]
  50. Spence, S.A. Between will and action. J. Neurol. Neurol. Psychiatry 2000, 69. [Google Scholar] [CrossRef]
  51. Sternberg, R.J. Beyond IQ: A Triarchic Theory of intelligence; Cambridge University Press: Cambridge, UK, 1985. [Google Scholar]
  52. Neisser, U. Cognition and Reality: Principles and implications of Cognitive Psychology; W.H. Freeman and Company: San Fransisco, CA, USA, 1976. [Google Scholar]
  53. Beck, J. Cognitive Therapy: Basics and Beyond; Guilford: New York, NY, USA, 1995. [Google Scholar]
  54. Freeman, A.; DeWolf, R. The 10 Dumbest Mistakes Smart People Make and How to Avoid Them; Harper Collins Publ.: New York, NY, USA, 1992. [Google Scholar]
  55. Kolb, B.; Whishaw, I.Q. Fundamentals of Human Neuropsychology; W.H. Freeman and Co.: New York, NY, USA, 1990. [Google Scholar]
  56. Benferhat, S.; Baida, R. A stratified first order logic approach for access control. Int. J. Int. Syst. 2004, 19, 817–836. [Google Scholar]
  57. Benferhat, S.; Garcia, L. Handling locally stratified inconsistent knowledge bases. Stud. Log. 2002, 70, 77–104. [Google Scholar]
  58. Lassez, C.; McAloon, K.; Port, G.S. Stratification and knowledge base management. J. Symb. Comput. 1989, 7, 509–522. [Google Scholar]
  59. Hunter, A.; Liu, W.R. Knowledge Base Stratification and Merging Based on Degree of Support. In Symbolic and Quantitative Approaches to Reasoning with Uncertainty; Springer-Verlag: Berlin, Germany, 2009; pp. 383–395. [Google Scholar]
  60. Yue, A.; Liu, W.; Hunter, A. Approaches to Constructing a Stratified Merged Knowledge Base. Proceedings of the 9th European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty (ECSQARU'07), Hammamet, Tunisia, 31 October–2 November 2007; pp. 54–65.
  61. Cholewinski, P. Stratified default logic. Proc. Comp. Sci. Logic'94 1994, 933, 456–470. [Google Scholar]
  62. Burgin, M. Structural Organization of Temporal Databases. Proceedings of the 17th International Conference on Software Engineering and Data Engineering (SEDE-2008), ISCA, Los Angeles, CA, USA, 30 June–2 July 2008; pp. 68–73.
  63. Snodgrass, R.T.; Jensen, C.S. Developing Time-Oriented Database Applications in SQL; Morgan Kaufmann: San Francisco, CA, USA, 1999. [Google Scholar]
  64. Burgin, M. Measuring Power of Algorithms, Programs, and Automata. In Artificial Intelligence and Computer Science; Nova Science Publishers: New York, NY, USA, 2005; pp. 1–61. [Google Scholar]
  65. Burgin, M. Super-Recursive Algorithms; Springer: Heidelberg, Germany, 2005. [Google Scholar]
  66. Hamkins, J.D.; Lewis, A. Infinite time turing machines. J. Symb. Log. 2000, 65, 567–604. [Google Scholar]
  67. Burgin, M. Grammars with Prohibition and Human-Computer Interaction. Proceedings of the Business and Industry Simulation Symposium, San Diego, CA, USA, 3–7 April 2005b; pp. 143–147.
  68. Okhotin, A. Boolean grammars. Inf. Comput. 2004, 194, 19–48. [Google Scholar]
  69. Burgin, M. Data, information, and knowledge. Information 2004, 7, 47–57. [Google Scholar]
  70. Barwise, J.; Perry, J. Situations and Attitudes; MIT Press: Cambridge, MA, USA, 1983. [Google Scholar]
  71. Dretske, F.I. Knowledge and the Flow of Information; Basil Blackwell: Oxford, UK, 1981. [Google Scholar]
  72. Burgin, M. Logical Tools for Program Integration and Interoperability. Proceedings of the IASTED International Conference on Software Engineering and Applications, MIT, CA, USA, 9–11 November 2004a; pp. 743–748.
  73. Cohn, P.M. Universal Algebra; Harper&Row, Publ.: New York, NY, USA, 1965. [Google Scholar]
  74. Kohlas, J.; Stärk, R.F. Information algebras and consequence operators. Log. Universalis 2007, 1, 139–165. [Google Scholar]
  75. Shoenfield, J.R. Mathematical Logic; Addison-Wesley: Reading, MA, USA, 1967. [Google Scholar]
  76. Burgin, M. Information algebras. Control Syst. Mach. 1997, 6, 5–16. (in Russian). [Google Scholar]
  77. Burgin, M. Functional equivalence of operators and parallel computations. Program. Comput. Softw. 1980, 6, 283–294. [Google Scholar]

Share and Cite

MDPI and ACS Style

Burgin, M. Epistemic Information in Stratified M-Spaces. Information 2011, 2, 697-726. https://doi.org/10.3390/info2040697

AMA Style

Burgin M. Epistemic Information in Stratified M-Spaces. Information. 2011; 2(4):697-726. https://doi.org/10.3390/info2040697

Chicago/Turabian Style

Burgin, Mark. 2011. "Epistemic Information in Stratified M-Spaces" Information 2, no. 4: 697-726. https://doi.org/10.3390/info2040697

Article Metrics

Back to TopTop