EXK-SC: A Semantic Communication Model Based on Information Framework Expansion and Knowledge Collision

Semantic communication is not focused on improving the accuracy of transmitted symbols, but is concerned with expressing the expected meaning that the symbol sequence exactly carries. However, the measurement of semantic messages and their corresponding codebook generation are still open issues. Expansion, which integrates simple things into a complex system and even generates intelligence, is truly consistent with the evolution of the human language system. We apply this idea to the semantic communication system, quantifying semantic transmission by symbol sequences and investigating the semantic information system in a similar way as Shannon’s method for digital communication systems. This work is the first to discuss semantic expansion and knowledge collision in the semantic information framework. Some important theoretical results are presented, including the relationship between semantic expansion and the transmission information rate. We believe such a semantic information framework may provide a new paradigm for semantic communications, and semantic expansion and knowledge collision will be the cornerstone of semantic information theory.


I. INTRODUCTION AND OVERVIEW
It is well known that Shannon's information theory [1] answers two fundamental questions in digital communication theory [2], namely about the ultimate data compression rate (answer: the entropy H) and the ultimate reliable transmission rate of communication (answer: the channel capacity C).It also addresses technical implementation problems in digital communication systems, enabling the end user to receive the same symbols as the sender.However, with the ever-increasing demand for intelligent wireless communications, communication architecture is evolving from only focusing on transmitting symbols to the intelligent interconnection of everything [3].In the 1950s, Weaver [4] discussed the semantic problem of communications, and categorized communications into the following three levels: Recently, semantic information theory including semantic communications has attracted much attention.However, how to set up a reasonable and feasible measure of semantic messages is still an open problem, which may be the greatest challenge for the new developments of semantic communication (SC) systems.On the other hand, currently in 6G networks, intelligent interconnections of everything will bring a new paradigm of the communication mode with semantic interaction.In this regard, finding a way to quickly reflect the semantic processing of messages will become one more promising pathway to 6G intelligent networking systems.This paper proposes a new communication system framework based on information framework expansion and knowledge collision.It extends Shannon's theory of communication (Level A) to a theory of semantic communication (Level B).This work is initially influenced by Choi and Park [5], but it provides further important contributions, which are as follows: • We generalize the work of Choi and Park from a single fixed semantic type to a dynamic expansion mode, associated with knowledge collision, which can reflect the asynchronous knowledge update processes between the sender and the receiver to some degree.In addition, it also takes into account the effect of channel noise in the new model.
• We present a new measure related to the semantic communication system based on the framework with semantic expansion and knowledge collision, called the Measure of Comprehension and Interpretation, which can be used to quantify the semantic entropy of discrete sources.
• We discuss the additional gains from semantic expansion and find the relationship between the semantic expansion and the transmission information rate.We demonstrate that knowledge matching of its asynchronous scaling up plays a key role in semantic communications.
As a primary work, the system proposed in this paper only focuses on discrete cases and makes some drastic simplifications.We assume that semantics are abstractly generated only from signals and knowledge, without considering other factors.In addition, semantic types, signals, responses, and knowledge instances (explained in Section II) are represented by random variables.Moreover, we do not focus on the effectiveness problem (Level C), which is beyond the scope of this paper.However, we believe that these simplifications are necessary for us to focus on the "core" issues to clearly explain and set up connections within a general semantic information theory.To the best of our knowledge, this is the first study to explore the semantic expansion and the asynchronous scaling up of semantic knowledge.The insights gained from this work may be of great assistance to the development of future intelligent semantic communication systems and 6G.
The rest of this paper is organized as follows.Section II introduces the related work of semantic communications.It summarizes the key concepts and theoretical results of Shannon's information theory and semantic information theory.In Section III, we present the new generalized model of semantic communications from two aspects, one is an attempt to set up a close relationship with the Shannon communication model, and the other is to find a feasible modification of the model so that it can exactly reflect the quick surge of semantic communications from the user requirements.In Section IV, we lay out the implementation details of the simulation and discuss the experimental results.Finally, we conclude the paper in Section V.

II. RELATED WORK
In this section, we introduce the framework of semantic information theory, specifically the key concept, semantic entropy.We then go on to explore the works that we are concerned with, semantic communication as a signaling game with correlated knowledge bases.

A. Preliminaries
Although Shannon's information did not address the semantic problem of communications, it provided important insights into the message-processing techniques associated with the focus of both the sender and the receiver's attention.Thus, in this subsection, we briefly introduce the main concepts and theoretical results of Shannon's information theory.
Entropy.It is a measure of the uncertainty of a random variable [2].Let X be a discrete random variable with alphabet X and probability mass function p(x) = Pr{X = x}, x ∈ X .The entropy H(X) of X is defined as follows: Mutual information.It is a measure of the amount of information that one random variable contains about another.That is, mutual information can be seen as the reduction in the uncertainty of one random variable due to the knowledge of the other.Let us consider two random variables X and Y with a joint probability mass function p(x, y) and the marginal probability mass functions p(x) and p(y).The mutual information between X and Y is defined as follows: Channel capacity.The channel capacity is the maximum amount of mutual information given the conditional transit probability from X to Y , p(y|x).It is defined by where X and Y are the input and output of the channel, respectively.
Source coding theorem.As N → ∞, N i.i.d.(independent identically distribution) random variables with entropy H(X) can be compressed into a little more than N H(X) bits to represent them completely, the information loss can be negligible in this case.Conversely, if they are compressed into fewer than N H(X) bits, there will be errors with a non-zero probability.
Channel coding theorem.For a discrete memoryless channel, all rates below capacity C are achievable.Conversely, any sequence of codes with a negligible error must obey the rule that its transmission rate R is not greater than the channel capacity, R ≤ C.

B. Semantic Information Theory
In the literature, there are some works that focus on the semantic information theory.Carnap and Bar-Hillel [6] were the first to propose the concept of semantic entropy, using logical probability rather than statistical probability to measure the semantic entropy of a sentence.For simplicity, it is necessary here to clarify that we use p and m to denote statistical probability and logical probability in this paper, respectively.The logical probability of a sentence is measured by the likelihood that the sentence is true in all possible situations [7].Then, the semantic information of the message e is defined as follows: where m(e) is the logical probability of e.However, this metric led to a paradox that any fact has an infinite amount of information when it contradicts itself, i.e., H S (e ∧ ¬e) = ∞.Floridi [8] solved the paradox in Carnap and Bar-Hillel's proposal [6], adopting the relative distance of semantics to measure the amount of information.
Bao et al. [7] defined the semantic entropy of a message x as follows: where the logical probability of x is given by W is the symbol set of a source, |= is the proposition satisfaction relation, and W x is the set of models for x.In addition, µ is a probability measure, Besides the logical probability, there are some definitions of semantic entropy based on different backgrounds [9] [10].D'Alfonso [11] utilized the notion of truthlikeness to quantify semantic information.Kolchinsky and Wolpert [12] defined semantic entropy as the syntactic information that a physical system has about its environment, which is necessary for the system to maintain its own existence.Kountouris and Pappas [13] advocated for assessing and extracting the semantic value of data at three different granularity levels, namely the microscopic scale, mesoscopic scale, and macroscopic scale.
Analogous to Shannon's information theory, some related theories at the semantic level, such as semantic channel capacity, semantic rate distortion, and information bottleneck [14], are also explored.Based on (5), Bao et al. further proposed the semantic channel coding theorem [7].
Moreover, Liu et al. [15] formulated rate distortion in semantic communication as follows: where D s is the semantic distortion between source, X, and recovered information, X, at the receiver.D w is the distortion between semantic representation, W , and received semantic representation, Ŵ .In addition, some works [16], [17] generalized it, leading to the development of semantic rate distortion theory.
In recent years, the fusion of semantic communication algorithms and learning theory [18]- [21] has driven the changes in communication architecture.Although these explorations did not fully open the door to semantic communication, they provided us with the motivation to move forward theoretically.

C. Semantic Communication as a Lewis Signaling Game with Knowledge Bases
Recently, Choi and Park [5] proposed an SC model based on the Lewis signaling game, which provided some meaningful results.Their work motivated us to conduct this study to some degree.
Further, we make a generalization of this model by adding the scaling up update module of the knowledge instance at both the sender and the receiver, which may be asynchronous, and closer to real semantic communications in the era of intelligent communication, i.e., human to robots, robots to robots, machines to machines and cells to cells in bio-molecular communications, etc.
Thus, we provide further details.
Suppose there is a semantic communication system where Alice is the sender and Bob is the receiver.Let T ∈ T denote the semantic type that includes semantic information or messages.
Alice wishes to transmit T to Bob by sending a signal S ∈ S, and Bob chooses its response R ∈ R. Both Alice and Bob utilize their local knowledge bases K A and K B , respectively.The semantic architecture can be described as follows: where K A ∈ K A and K B ∈ K B are knowledge instances.If R = T , the communication process can be seen as successful.Moreover, the reward for one communication is defined as follows: Our object is to maximize the average reward.We use the success rate of semantic agreement (SRSA) to represent it.
The knowledge base can be regarded as the side information.Alice has her knowledge base K A .
The instance of Alice's knowledge base at each time is K A ∈ K A , which affects the generation of semantic types.On the other hand, Bob has his knowledge base K B and its instance is K B is the "core", which is used to infer the intended message together with the received signal.
Based on these foundations, Choi and Park [5] derived some meaningful results.On the one hand, the similarity of the two knowledge instances plays a crucial role in SC.With limited bandwidth, the communication quality is regulated by I(K A ; K B |S).On the other hand, the SRSA is dependent on the similarity of knowledge bases at both ends of the communication, the sender and the receiver.However, a close relationship with Shannon's communication model was not established.In this work, we will construct a generalization of the model by adding the knowledge scaling up update module at both the sender and the receiver, which may be asynchronous, but it can closely reflect the quick surge of semantic communications in many potential applications.

III. SEMANTIC COMMUNICATION MODELS
In this section, we will present the generalized model of semantic communications from two different aspects, one is an attempt to set up a close relationship with the Shannon communication model, and the other is to find a feasible modification of the model so that it can exactly reflect the quick surge of semantic communications from the user requirements.In this regard, we first set up a basic model, and then extend it by adding the knowledge base update functional module with information framework expansion and knowledge collision.

A. Motivation
Shannon's communication system is concerned with the accurate transmission of symbols over channels.However, it does not take into account the differences in the knowledge backgrounds of the two parties, which affect whether the communication can convey the desired meaning.In other words, the matching of knowledge can also be treated as another special channel, although it may not actually exist in an explicit form.In this work, we denote it as a virtual channel.We want to fill this gap and propose an intelligent communication system with the coexistence of the real channel and virtual channel, making it easier to formulate the information framework.Thus, we consider not only the physical channel for symbol transmission but also the virtual channel for the transfer of knowledge between two parties.Moreover, the knowledge of both parties is also evolving, so we take into account the asynchronous scaling up updates of knowledge and semantic expansion.

B. Basic Model
Let us first observe the channel, which is the physical medium of message exchange, and plays a key role in the model setup of a complete semantic communication system.A communication channel is a system in which the output depends probabilistically on its input [2].For simplification, we still use Alice and Bob as the two parties participating in the communication.Suppose Alice and Bob are at opposite ends of a memoryless channel, which can transmit physical signals.
The channel is said to be memoryless if the probability distribution of the output depends only on the input at that time and it is conditionally independent of previous inputs or outputs.That is, the output of the channel is only related to the input at the current moment and no feedback processing is considered here.
We use S and Ŝ to represent the input and output of the channel.Under the interference of the noise, the signal S becomes Ŝ through channel transmission.Specifically, Alice observes the input S of the channel, and Bob observes the output Ŝ.Let K A and K B represent the knowledge instance possessed by Alice and Bob, respectively.On the one hand, Alice encodes S as T with K A .This means that Alice uses the knowledge instance K A to process the signal S, resulting in the semantic type T , i.e., (S, K A ) → T .On the other hand, Bob decodes Ŝ into the response R with K B , which can be expressed as ( Ŝ, K B ) → R. The architecture of this process can be expressed as follows: We denote the process (S, K A ) → T as semantic encoding and ( Ŝ, K B ) → T as semantic decoding.They indicate the understanding of meaning at the semantic level by both parties.The definitions of symbols are summarized as follows: • Semantic Types: T = t k , k = 1, . . ., |T |, is a random variable that is generated by Alice.
• Knowledge instances: Proof.Since (S, K A ) → T and ( Ŝ, K B ) → T , we obtain H(T ) = H(K A , S) and H( T ) = H(K B , Ŝ).With the characteristics of entropy and mutual information, it follows that

= I(S; Ŝ) + I(K
which completes the proof. The mutual information between T and T reflects the effectiveness of communication.It indicates the highest rate in bits per channel use at which information can be sent with an arbitrarily low probability of error.We note that I(T ; T ) consists of the following three terms: 1) I(S; Ŝ), the mutual information between the input and output of the channel.It corresponds to the channel capacity in Shannon's information theory.
2) I(K A ; K B |S, Ŝ), the mutual information between K A and K B given S and Ŝ.It indicates the amount of information that two knowledge instances contain about each other when signals are known.
3) I(S; K B | Ŝ) + I( Ŝ; K A |S), the conditional mutual information between S and K B , Ŝ and From Eq. ( 10), we know that if Bob wants to fully understand what Alice means, the communication process needs to meet two conditions, one is the accuracy of the received symbols during channel transmission, and the other is the matching degree of the knowledge instances of both parties.In other words, i) The transmission S → Ŝ should be reliable.It indicates the effect of channel noise on signal transmission.Moreover, the physical carrier of this process actually exists, which we call an explicit channel.ii) The two knowledge instances should be similar.
Although there is no actual transmission between K A and K B , we assume that there is a virtual channel that reflects the probabilistic relationship between knowledge instances, which we call an implicit channel.The explicit channel and implicit channel together form the transmission medium of a communication system.Specifically, the characteristics of the explicit channel determine the value of the first term in Eq. ( 10), and the implicit channel determines the second term.In addition, the last term is affected by both explicit and implicit channels.
If the communication system has only the explicit channel, it degenerates to the Shannon case, and I(T ; T ) = I(S; Ŝ).Furthermore, we are interested in the following three special cases (demonstrated in Appendix A).
I. The explicit channel is noiseless, i.e., S = Ŝ.In this setting, the formula (10) can be simplified to which is consistent with the result in [5].The mutual information between T and T is subjected to the entropy of the signal S and the conditional mutual information between K A and K B given S. In semantic communications or emergent communications, the signal S is usually limited.According to Eq. ( 17), one can increase the conditional mutual information between K A and K B to solve the problem from another perspective.In other words, in the case of limited technical bandwidth, the SC performance can be improved by increasing the similarity between two knowledge instances.
II.The implicit channel is noiseless, i.e.K A = K B .This means that Alice and Bob have the same knowledge instance, so the communication performance will not be affected by the difference in the background of both parties.I(T ; T ) can be expressed as follows: where and Ŝ.Even when the implicit channel is noiseless, I(T ; T ) is not exactly equal to I(S; Ŝ), and we cannot ignore the influence of knowledge in SC.Compared to Shannon's channel capacity, it adds a term, regulated by the relationship between the knowledge instance and the signal.
Moreover, if K A is a function of S and Ŝ, the left equation (a) holds; if K A is independent of the signal S or Ŝ, the right equation (b) holds.
Proof.The non-negativity of mutual information, I(K A ; S; Ŝ) ≥ 0, means that I(T ; T ) can be upper bounded by I(S; Ŝ) + H(K A ).Because the mutual information is lower than the entropy, we obtain H(K A ) − I(K A ; S; Ŝ) ≥ 0. By combining these results, we obtain which completes the proof of the bound of I(T ; T ).

III.
Both explicit and implicit channels are noiseless, i.e., S = Ŝ, K A = K B .In this setting, the mutual information between T and T equals the entropy of T or T , Eq. ( 22) indicates that Bob can perfectly understand the meaning of Alice when channels are noiseless.That is, no information is lost.
The following result shows the bounds of SRSA, constrained by the characteristics of the explicit and implicit channels.
Lemma 2. The SRSA satisfies Proof.(K B , Ŝ) can be seen as an estimator for (K A , S).Let P e = Pr{T = T }; then, with Fano's Inequality, we obtain Since SRSA= 1 − P e , we can obtain Eq. ( 23), which completes the proof.

C. EXK-SC Model
Let S 1 and S 2 denote the input signals of the explicit channel.Ŝ1 and Ŝ2 are their corresponding output signals.Similarly, K 1 A and K 2 A represent Alice's knowledge instances.Alice encodes S 1 as T 1 with K 1 A , and S 2 as T 2 with K 2 A .On the other hand, Bob uses K 1 B and K 2 B , decoding Ŝ1 and Ŝ2 into T1 and T2 , respectively.In the basic model, the encoding and decoding processes of the two transmissions are It is well known that the integration of lots of simple things can be expanded to a complex system and can even generate intelligence.Such an expansion process is consistent with the evolution of the human language system.Based on a similar idea, we extend the basic model As known, expansion often implies collision and fusion.Similarly, the expansion of signals corresponds to the collision of knowledge bases in this work.For Alice, we use to represent the collision process, where denotes the collision and α is the collision factor.
Eq. (31) indicates that besides the characteristics of explicit and implicit channels, the relationship between α and β also affects the performance of communication.
Lemma 4. When the explicit channel is noiseless, the gain brought by semantic expansion is given by where Proof.We note that Then, We have which completes the proof.
Lemma 4 shows that the semantic enhancement, I(T ; T )−I(T 1 ; T1 ), is determined not only by the knowledge scaling up but also by the cost of the information rate over the explicit physical channel, H(S) − H(S 1 ).It exactly illustrates the relationship between the semantic expansion and the transmission information rate.
Lemma 5. When the explicit channel is noiseless, the mutual information between T and T is bounded by Proof.Since expansion would create more possibilities, leading to an increase in uncertainty, we can obtain Because of the non-negativity of entropy and conditioning reduces entropy, Eq. ( 40) can be derived directly.We omit the proof.
Lemma 5 shows the bounds of the semantic communication rate in the perfect explicit channel mode.The upper bound is composed of the entropy rate over the explicit channel and the collision of knowledge instances of the sender, which is reasonable, as expected.

IV. EXPERIMENT AND NUMERICAL RESULTS
In this section, we use SRSA to measure the performance of SC, especially the impact that asynchronous knowledge scaling up has on the system.
Basic Model.We use Q-learning in [22] to complete semantic encoding and decoding, which may reflect the continuous semantic learning process.
In addition, S, K A and K B are uniformly distributed.We would like to discuss the impact of the characteristics of the explicit channel and implicit channel on SC.We use the binary symmetric channel (BSC) to denote the explicit channel.The BSC is shown in Figure 1, which shows the probabilistic relationship between S and Ŝ, with the error probability 1 .
For the implicit channel, we assume the correlation between K A and K B satisfies  where U ∼ Unif{1, L} is an independent random variable.2 illustrates the discrepancy between the knowledge bases at both parties of communications, 0 < 2 < 1.In Figure 2, we show the SRSA with 1 ∈ [0, 0.5] and 2 ∈ [0, 0.5].When 1 = 2 = 0, SRSA reaches the maximum 1.
As 1 and 2 increase, SASA will decrease in both directions.This indicates that two channels jointly determine the quality of communication.For high-quality SC, it is necessary to meet the condition that Ŝ = S and K B = K A with a high probability.
In addition, we discuss the impact of explicit and implicit channels on SC, respectively.We take the channel error 1 and 2 as the horizontal axis, and the SRSA entering the stable region as the vertical axis.Figure 3a shows the results of the explicit channel.As 1 increases, the mean value of SRSA first decreases and then has a gradual increase.When 1 = 0 or 1, the communication performance is the best and the mean is close to 1.Moreover, it falls to a low point when 1 = 0.5.On the other hand, the trend of variance is completely opposite to that of the mean.These results are consistent with the description of channels in Shannon's information theory.Figure 3b shows the results of the implicit channel.As 2 increases, the mean value of SRSA keeps decreasing, and the variance increases.It reflects the impact of knowledge misalignment on SC.
EXK-SC Model.We want to explore how semantic communication quality varies in the to set up the type class methods to provide more insights on the knowledge base formulation and the system design, which requires further study.Semantic communication is evolving towards intelligence.The insights gained from this work may be of assistance to the development of future semantic communication systems and 6G.It is also expected to pave the way for the design of next-generation real-time data networking and provide the foundational technology for a plethora of socially useful services, including autonomous transportation, consumer robotics, VR/AR, and the metaverse.

ACKNOWLEDGMENTS
We thank Prof. Choi for the help and source code in [5].
Level A. How accurately can the symbols of communication be transmitted?(The technical problem.)Level B. How precisely do the transmitted symbols convey the desired meaning?(The semantic problem.)Level C. How effectively does the received meaning affect conduct in the desired way?(The effectiveness problem.)

Lemma 1 .
the knowledge instance used by Alice and Bob during semantic encoding and decoding, respectively.Under the knowledge instances K A and K B , the mutual information between T and T is obtained as follows:I(T ; T ) = I(S; Ŝ) + I(K A ; K B |S, Ŝ) + I(S; K B | Ŝ) + I( Ŝ; K A |S)(10)where I(K A ; K B |S, Ŝ), I(S; K B | Ŝ) and I( Ŝ; K A |S) represent the conditional mutual information.

Corollary 1 .≤
If the implicit channel is noiseless, the mutual information between T and T satisfies I(S; Ŝ) I(S; Ŝ) + H(K A ).
proposed in Section II-C using expansion.In this context, Alice wishes to send an expansion of multiple signals to Bob.We first consider the case of two signals, which can be generalized to more instances.Now, what Alice sends is expanded from S = S 1 to S = S 1 ⊕ T 2 .We use ⊕ to denote semantic expansion.Specifically, the expansion of signals only represents their semantic combination.For instance, ' Shannon published a paper' expands to 'Shannon published a paper in The Bell system technical journal [1]'.

Fig. 2 .
Fig. 2. The simulation results of SRSA with the explicit channel error probability 1 and the implicit channel error probability 2.