You are currently viewing a new version of our website. To view the old version click .
Technologies
  • Article
  • Open Access

21 June 2025

Emotionally Controllable Text Steganography Based on Large Language Model and Named Entity

,
and
College of Communication Engineering, Rocket Force University of Engineering, Xi’an 710025, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Research on Security and Privacy of Data and Networks

Abstract

For the process of covert transmission of text information, in addition to the need to ensure the quality of the text at the same time, it is also necessary to make the text content match the current context. However, the existing text steganography methods excessively pursue the quality of the text, and lack constraints on the content and emotional expression of the generated steganographic text (stegotext). In order to solve this problem, this paper proposes an emotionally controllable text steganography based on large language model and named entity. The large language model is used for text generation to improve the quality of the generated stegotext. The named entity recognition is used to construct an entity extraction module to obtain the current context-centered text and constrain the text generation content. The sentiment analysis method is used to mine the sentiment tendency so that the stegotext contains rich sentiment information and improves its concealment. Through experimental validation on the generic domain movie reviews dataset IMDB, the results prove that the proposed method has significantly improved hiding capacity, perplexity, and security compared with the existing mainstream methods, and the stegotext has a strong connection with the current context.

1. Introduction

With the continuous development of network technology and information digitisation, digital communication occupies an important position in daily life, which puts forward higher requirements for information security in cyberspace. In order to protect the medium can be transmitted safely, at first, researchers protect the information by taking encryption. Among the types of media include text [1], pictures [2,3,4], and video [5]. However, encryption algorithms, while protecting the medium for transmission, expose the encrypted medium to potential attackers and make it vulnerable to targeted attacks [6]. To solve the above problem, steganography has received a lot of attention.
Steganography is a technique of embedding secret information in a digital medium with the aim of achieving covert communication of information [7]. Among them, text, as one of the commonly used carriers in digital communication, has led to numerous cases of malicious theft of text information due to the openness of the network. Text steganography is a text-based information hiding technology, which can realise the covert transmission of secret text in the open channel, and has gradually become a research hotspot in the field of information network security [8]. In recent years, with the development of natural language processing (NLP) and language models, generative text steganography has received extensive attention [9]. It uses the language model as the text generator, and constructs the correspondence between the predicted word and the secret information bit stream through the encoding strategy, so as to realize the stegotext generation while embedding the secret information.
However, generative text steganography generates stegotext through the probability distribution of the output of the language model [10]. This means that the generation of words at each step is determined by the words nearby the word and the secret information, rather than by the semantics and the current context [11]. If the language model selects a word that violates the context at the current time step, it will lead to a deviation in the semantics of subsequent text generation. The above problems of deviation between stegotext and context content and sentiment affect the imperceptibility of stegotext, making it easier to be detected and difficult to realize covert information transmission.
To solve the above problems, it is important to find a way to increase the semantic and emotional information we can obtain from the context. Therefore, this paper extracts the central text and emotional features in the current context information as the input of LLM to constrain the generation of stegotext, ensure that the text content meets the context requirements, and improve the quality, imperceptibility and security of stegotext.
In summary, the contributions of this work can be outlined as follows.
  • Proposes a controllable generative text steganography framework based on LLM. The LLM is used as the stegotext generator without extra training, and the controllable g eneration of stegotext is realized by only using the cue word template, avoiding the complex training process of the traditional deep neural network (DNN) model.
  • For the first time, named entity and sentiment analysis are integrated into the process of generating stegotext. By identifying and extracting the central text and emotional features of the current context, the generation of stegotext is constrained. On the basis of information hiding, the consistency between the stegotext and the current context is guaranteed.
  • Extensive experiments on different models. Compared with the mainstream advanced algorithms in recent years, the framework proposed in this paper has significantly improved in all aspects, indicating that the framework has better performance in text quality, concealment and security.
The rest of this paper involves the following parts: Section 2 introduces representative text steganography. Section 3 presents the background of text steganography. Section 4 describes the overall framework and design details of the proposed model in this paper. Section 5 is indicated the results and analysis through a series of experiments. Finally, the full text is summarized in Section 6.

3. Background of Text Steganography

3.1. The Framework for Text Steganography

The general framework of text steganography can be described vividly by the Simmons’ “prisoner’s problem” [37], as shown in Figure 1. Alice, as the sender, needs to hide the secret information and send it to the receiver Bob. Then Bob decrypts and restores the received secret information. At the same time of the transmission process, there is a “monitor” Eve, who is responsible for monitoring the transmitted text and judging whether it contains secret information.
Figure 1. The general framework of steganography.
The model was modeled with the equation shown in Equation (1) [25]:
E m b : C × K × M S E x t : S × K M
where Emb is information hiding, Ext is information extraction, C is the carrier text to hide the secret information, K is the key to encrypt or contain the shared information, M is the secret information to be hidden, S is the stegotext.

3.2. Generative Text Steganography

Generative text steganography is a technique that uses NLP to generate text containing secret information. In NLP, the process of text generation is usually viewed as the creation of a sequence x = [ x 1 , , x n ] based on the joint probability of the language model P L M ( x ) , which is formulated as follows [38]:
P L M ( x ) = t = 1 n P L M ( x t x 1 , , x t 1 ) = t = 1 n P L M ( x t x < t )
Different from the traditional text steganography, generative text steganography does not need the original cover, but directly embedded the secret information in the process of automatic text generation, which greatly improved the efficiency of secret information embedding [39]. With the development of neural network models, generative text steganography based on neural network models has become a hot topic in the current field of research. The generation process is shown in Figure 2.
Figure 2. Generative process of generative text steganography based on neural network model.
In the stegotext generation stage, the neural network model used as text generation is first trained. Then, according to the current text sequence, the trained neural network model is used to predict the word, and the candidate word pool is constructed. Finally, under the constraints of secret information bits, encoding strategy and key, the candidate words that meet the requirements are selected. The above process is repeated until the secret information is completely embedded, and the stegotext is obtained [40].
With the emergence of large-scale pre-trained language models based on transformers, especially the wide application of LLM, the quality of generated text is getting higher and higher, and generation-based text steganography is more and more applied in the field of text steganography [39].

3.3. Steganographic Coding Strategy

For generative text steganography, a steganographic coding strategy is necessary to ensure that the receiver can extract secret information from the stegotext with less shared information. Taking FLC and HC as examples [25], the specific structure is shown in Figure 3 and Figure 4.
Figure 3. Structure of fixed-length coding.
Figure 4. Structure of Huffman coding.
In the secret information embedding phase, the binary tree of the corresponding encoding is constructed, and the candidate words corresponding to the encoded values are selected. On the premise that the sender and the receiver share the text generation model and steganographic coding strategy, the receiver encodes the candidate pool in each step of the text generation process, and according to the stegotext, the secret information bit stream corresponding to the current word is obtained. In the stegotext generation process, the words with high prediction probability should be selected as much as possible. Therefore, the choice of encoding strategy has a great impact on the quality of stegotext [41].

4. Emotionally Controllable Text Steganography Framework Based on Large Language Model and Named Entity

In this paper, we propose an emotionally controllable text steganography based on LLM and named entity recognition, and the overall architecture is shown in Figure 5. The core of the framework can be split into three parts, which are the entity extraction module for obtaining named entities, the sentiment analysis module for obtaining the sentiment tendency of the current context, and the steganography generator for generating stegotext.
Figure 5. The overall architecture of the proposed method.

4.1. Entity Extraction Module

In order to better convey the secret information to the receiver and reduce the possibility of the generated stegotext being perceived by the “monitor” Eve, it is necessary to improve the semantic correlation between the current context and the stegotext. Therefore, before the LLM generates text based on secret information, it is provided with a set of cue words. Through the entity extraction method, the entities are extracted from the current context information, and the entity words with different numbers are randomly selected to fill in the cue words, and the encoding results are used as the input of the LLM.
In this paper, the retrieval-augmented named entity recognition (RaNER) model [42] is chosen as the entity extraction module. The specific structure is shown in Figure 6.
Figure 6. Structure of entity extraction module.
RaNER uses Transformer and conditional random field (CRF) as the base model and combines cross-lingual language model (XLM) [43] with robustly optimized BERT pretraining approach (RoBERTa) [44] is combined as a pretraining model base, by following the Transformer architecture and dynamic masking strategy of RoBERTa, while inheriting the XLM pair sentence structure, the model training efficiency becomes faster, the text generation is adapted to more linguistic scenarios, and the entity recognition is expanded from a monolingual BERT relies on local context to combine with global retrieval of relevant knowledge. External tools are used to retrieve external sentences related to the current context as additional context, which provides supplementary information to solve the problem of polysemy and improves the accuracy of entity recognition in the whole text. Finally, a multi-view training approach is adopted to improve the performance of named entity recognition by integrating the representations of the original text and the semantic information of external sentences retrieved by external tools. According to the generic domain named entities, four entity categories are preset, as shown in Table 2 [24], where MISC are all other entities [45].
Table 2. The predefined entity categories in RaNER.
The entity extraction module is a process from entity extraction to entity selection and finally to prompt template generation, as shown in Table 3, an example of entity extraction from secret text to backfill prompt words. The process of entity word selection is random, depending on the number of extraction results. If the number of entity identification is 0, then there is no selection. If 1, then select the entity word. When the entity word is greater than or equal to 2, then randomly select half the number of entity words.
Table 3. Example of entity extraction module.

4.2. Sentiment Analysis Module

The inclusion of a sentiment analysis module in the text steganography framework is to ensure that the generated stegotext and the current context can maintain consistency in terms of sentiment tendency, so as to improve the cognitive imperceptibility of the generated results. The pre-trained model StructBERT [46] is selected as the sentiment analysis model in the framework. The specific model structure is shown in Figure 7.
Figure 7. Structure of sentiment analysis module.
By disrupting some of the words in the sentence and requiring the model to restore the original order, the model is forced to understand the dependency relationship between words and the local structure, and to capture the sentiment cues in the complex sentences more accurately, so as to enhance the model’s capability of capturing the sentiment expression of the utterance.
As shown in Table 4, by classifying the current context into happy, sad, disgusted, and other emotions [46], the module backfills the identified emotion categories into the cue word template to realize the emotion injection into the implicitly written text.
Table 4. Example of sentiment analysis module.

4.3. Stego Generator

The steganography generator consists of the steganography encoding module and the text generation module. The steganographic coding module generates the corresponding coded values based on the conditional probability of the candidate word at each moment under a specific coding strategy. HC has been widely used in text steganography due to its adoption of variable-length encoding and its property of having a better compression ratio [25]. From [9,10], HC has an advantage in perceptual imperceptibility compared with coding methods such as AC and SAAC. The powerful text generation capability of LLM can effectively make up for the shortcomings in statistical and semantic imperceptibility. Therefore, in this paper, HC is chosen as the coding strategy for generating stegotext, and its implementation is shown as follows [47]: by constructing a special binary tree (called Huffman tree), the high frequency characters are mapped to shorter binary codes, and the low frequency characters are mapped to more extended binary codes by taking advantage of the difference in the frequency of characters’ appearances in order to achieve the purpose of compressing data. Each leaf node corresponds to a character in the tree, and non-leaf nodes represent the merging of character codes. Character encodings with high frequency are located on shorter paths, while character encodings with low frequency are located on longer paths, thus increasing the probability of selecting high frequency characters.
The text generation module uses an LLM as a language model for the auto-regressive generation of text sequences. Firstly, the prompts of the LLM are obtained based on the contextual information, entity extraction module and sentiment analysis module. The prompt template is shown in Table 5. The LLM generates a candidate pool at each moment under the constraint of the prompts. Subsequently, the steganographic encoding module generates an encoding value for each candidate word, and determines the candidate word based on the mapping relationship between secret information and coded value. Until the secret information is fully embedded, the stegotext is generated. The proposed information hiding algorithm are shown in Algorithm 1.
Algorithm 1 Information hiding algorithm
Input:  Sec ret   information   B ,   Contextual   information   I c t x
Output: Stegotext T
1: Initialize T
2: while B is not empty do
3:   P T Generate   text   prompts   based   on   I c t x
4:   w E Entity   words   are   obtained   from   I c t x through the entity extraction
5:   If   l e n ( w E ) = 1  then
6:    P E Generates   entity   prompt   based   on   w E
7:  else
8:    P E Generates   entity   prompts   based   on   l e n ( w E )     of   w E
9:  end if
10:  w S   Sentiment   words   are   obtained   from   I c t x through the sentiment analysis
11:  P S Generates   entity   prompt   based   on   w S
12:  Prompts   for   LLM   input   P = P T + P E + P S
13:  c p o o l Generate the candidate pool at current step t based on the P and T
14:  H T Build   a   Huffman   tree   based   on   c p o o l
15:  b Obtain the bitstream of the embedding at t based on B and HT
16:  x t Generate   the   output   word   based   on   c p o o l , HT, b
17:  B = B b   ,   T = T + x t   ,   t = t + 1
18: end while
19: return Stegotext T
Table 5. The prompt template of LLM.

5. Experiments and Analysis

5.1. Dataset and Experimental Setup

To be able to extract the central text and sentiment features from the text, this paper selects the IMDB movie review dataset [48] as the evaluation dataset. The IMDB movie review dataset is a widely used data resource for sentiment analysis and natural language processing tasks. After a series of preprocessing operations such as text cleaning and removing stop words, 33,000 pieces of data are obtained. The data set was divided 9:1:1, and 27,000 pieces of data were used as the training set for training, 3000 pieces of data were used as the test set to test the trained model, and the remaining 3000 pieces of data were used as the validation set. As shown in Figure 8, the median length of the 27,000 IMDB movie review dataset is 106, the average length is 131.7, and over 95% of the texts are less than 500 characters in length.
Figure 8. Length distribution of the IMDB dataset.
The experiments in this paper are conducted based on Pytorch 2.4.0 and Python 3.10, and the model training is accelerated by using NVIDIA A10 and CUDA 12.4. To ensure a fair comparison, we rebuild all baseline methods using the same details throughout the entire experiment. For RNN-Stega [25] and VAE-Stega [31] to be trained, the word vector dimension is 768, the hidden state dimension is 768, the dropout is set to 0.5, and the number of model training epochs is 50. The batch size is 128, the learning rate is 0.0001, and the maximum character length of the generated text does not exceed 256. Pretrained models (e.g., GPT-2, Qwen-2.5 and LLaMA-3) are obtained using Huggingface library.

5.2. Evaluation Metrics

5.2.1. Hidden Capacity

Hidden capacity is an important metric of text steganography. It describes the amount of secret information embedded in the stegotext. Hidden capacity is usually evaluated by Embedding Rate (ER) [34], which is calculated as shown in Equation (3):
E R = 1 N i = 1 N L i S i
where N represents the number of generated sentences, Li is the number of bits embedded in the ith sentence, and Si is the length of the ith sentence.

5.2.2. Perceptual Imperceptibility

In order to ensure the perceptual imperceptibility of the stegotext, the stegotext is required to be perceptually similar to the real text. The perceptual imperceptibility evaluation of text steganography can be carried out from both subjective and objective perspectives. Subjective evaluation means that people rate the stegotext based on its fluency and semantic coherence. The objective evaluation relies on the standard measure of text quality evaluation, perplexity (PPL) [25], to quantify the imperceptibility of the stegotext. The smaller the value, the higher the quality of the generated text. The specific calculation formula is shown in Equation (4):
p p l = 2 1 N i = 1 N l o g P i ( S i )
where N represents the total number of stegotext sentences, S i = { x i 1 , x i 2 , , x i n } is the ith sentence of the generated stegotext, n represents the total number of words in the ith sentence, and Pi(Si) represents the probability of the ith sentence.

5.2.3. Security

The difference in statistical distribution between the stegotext and the real text directly characterizes the security of the steganography algorithm. Kullback-Leibler divergence (KLD) [11] is usually used to calculate the difference between two probability distributions in terms of information entropy and other features, and the specific calculation formula is shown in Equation (5):
D K L ( P Q ) = i = 1 N [ p ( x i ) l o g p ( x i ) p ( x i ) l o g q ( x i ) ]
where P and Q represent the distribution before and after encoding, p(xi) and q(xi) represent the ith probability value in the distributions P and Q, respectively, and N represents the total number of stegotexts.

5.2.4. Cognitive Imperceptibility

When evaluating cognitive imperceptibility, it is necessary to determine whether the content and emotional features of the stegotext are related to the current context content. Therefore, this paper adopts bilingual evaluation understudy (BLEU) [49] to evaluate cognitive imperceptibility. It evaluates the quality of the generation by calculating the degree of n-gram overlap between the generated text and one or more reference texts. A higher BLEU value indicates a higher quality of the generated text. The calculation formula is shown in Equation (6) [50]:
B L E U = B P × e x p ( n = 1 N w n l o g ( P n ) )
where BP is the penalty factor, which is used to prevent the model from falsely improving the accuracy by repeatedly generating part of the content. wn is the weight used to weight the precision of different n-grams when calculating the BLEU score. Pn is the coincidence accuracy between the generated stegotext and the reference text. The specific calculation formulas of BP and Pn are shown in Equations (7) and (8):
B P = 1 , l c > l s , e 1 l s l c , l c l s
P n = i k m i n ( h k ( c i ) , m a x j m h k ( s i j ) ) i k m i n ( h k ( c i ) )
where lc is the length of the generated stegotext, ls is the length of the reference text, hk(ci) is the number of occurrences of the kth n-gram in the generated stegotext ci of reference text i, hk(sij) is the number of occurrences of the kth n-gram in the corresponding reference text s j of reference text i, and m is the set of reference texts.

5.3. Comparison Experiment

In this paper, experiments are carried out on the IMDB movie review dataset [48], and the representative LLaMA-3 and Qwen-2.5 in the large model are selected as text generation models for empirical research on text steganography. In the comparison experiments, two types of classical methods are selected respectively. One is represented by RNN-Stega [25] and VAE-Stega [31], which need to be trained deep learning model methods. NLS [33], SAAC [10], and LLsM [35] are pretrained methods that require no additional training.
In the comparison experiments, the dataset is divided into training, testing, and validation sets by 9:1:1. The models to be trained are trained for 50 rounds, and then experiments are conducted separately for all models using the validation set. In the stegotext generation process, stegotext is generated by randomly generating a bit stream of 30 to 256 bits as secret information. For the affective controlled text steganography approach fusing LLM and named entity proposed in this paper, based on hiding the above secret information, the validation set is used as the information of the current context for prompt word generation, which is tested on models of different parameter sizes, such as Qwen-2.5, LLaMA-3, and the specific results are shown in Table 6.
Table 6. The results of comparing experimental.
It can be seen from the comparison experiments that the LLaMA-3 (3B), which integrates the entity extraction and sentiment analysis modules, achieves the optimum in three evaluation metrics, ER, PPL, and BLEU, in the results of the text steganography experiments conducted. The optimal model for the other evaluation metrics, KLD, is Qwen-2.5 (3B), which integrates the entity extraction and sentiment analysis modules. The model with unconstrained semantics constructs the candidate pool by considering only the influence brought by the words around the predicted position, resulting in a deviation of the probability distribution of the candidate words from the probability distribution of the normal text at that position. After the introduction of the entity extraction module and the sentiment analysis module, the construction of the candidate pool at the predicted position is more comprehensively considered, and the probability distribution is closer to that of the normal text in terms of probability distribution, which makes the KLD smaller and improves the security of the text.
Taken together, the LLM integrating the entity extraction and sentiment analysis modules achieves results ahead of the comparative models in the text steganography methods, and it can be proved that the model has better results in embedding ability, hiddenness, and security.

5.4. Anti-Steganalysis

To assess the security of steganography, a steganalyzer is usually used to distinguish between normal text and stegotext. we use three steganalysis models: LS-CNN [23], TS-RNN [51] and BERT classifier [52] to test the anti-steganalysis ability of each model. The evaluation metrics are accuracy (Acc) and recall (R). The calculation formula is as follows:
Acc = T p + T n F p + F n + T p + T n
R = T p T p + F n
where Tp is the number of stegotexts correctly judged by steganalysis as stegotexts, Tn is the number of normal texts correctly judged by steganalysis as normal texts, Fp is the number of stegotexts incorrectly determined by steganalysis as normal texts, and Fn is the number of normal texts incorrectly determined by steganalysis as stegotexts.
Table 7 shows the anti-steganalysis comparison of the proposed model and the baseline model. According to Table 5, the proposed model shows a good performance in anti-steganalysis, which improves the security of stegotext.
Table 7. The results of anti-steganalysis experimental.

5.5. Ablation Experiment

In order to further validate the effectiveness of the entity extraction module and sentiment analysis module in the text steganography framework proposed in this paper, this paper conducts ablation experiments with LLaMA-3 as the language model, and the specific experimental results are shown in Table 8.
Table 8. The results of ablation experimental.
The comparison of the results of each model under different module combinations is shown in Figure 9. The horizontal axis represents the different combination methods, the left side of the vertical axis represents the scores of ER, KLD, and PPL, and the right side represents the BLEU scores. The results show that the LLaMA-3 (3B), which incorporates the entity extraction and sentiment analysis modules, dominates the evaluation metrics of PPL, KLD, and BLEU. In the experimental results, the introduction of entity extraction module and sentiment analysis module is mainly reflected in BLEU. Through entity words and sentiment tendency, the language model is guided to predict the candidate pool of the target location, and more consideration is given to the contextual information. The generated stegotext matches the context information better, which enhances the semantic concealment. The LLM without incorporating the entity extraction and sentiment analysis module consistently lagged the metrics compared to the other three methods. The above results show the effectiveness of the entity extraction module and the sentiment analysis module for generating controlled stegotext.
Figure 9. The results of ablation experimental.
To verify the generalization of the module proposed in this paper, LLaMA-3 (1B) with smaller parameters is tested and the results are shown in Table 9. From the results, it is easy to see that the introduction of the module achieves better results in all four metrics, ER, PPL, KLD and BLEU, compared to pure LLaMA-3 (1B). This shows that the module designed in this paper has better generalization and proves that our method still has better results at lower scales.
Table 9. Experimental results on module generalization.

5.6. Case Study

As shown in Table 10, we sample several stegotexts generated by baselines and our model. We can find that our stegotext performs better semantically and emotionally compared to the baseline. This indicates that our model has great potential for imperceptibility.
Table 10. Case Study on the IMDB datasets.
To assess the quality of the production of stegotexts, we randomly selected 10 sets of texts from the results and invited 10 PhDs to score them on a 5-point scale. The results are shown in Table 11. For the human evaluation, in general, the steganography based on LLM generation has achieved high scores. Compared with [35], the two modules proposed in this paper enable the generator to consider contextual information more comprehensively, and make the generated text maintain fluency while having better performance in semantics, so the score is slightly higher than it.
Table 11. The results of human evaluation.

6. Conclusions

In this paper, we focus on generative text steganography based on LLM and propose an emotionally controllable text steganography based on LLM and named entity. In this method, an entity extraction module composed of RaNER and a sentiment analysis module composed of StructBERT are designed to extract the central text and sentiment features from the current context information respectively. It is then used as the input of the large model, which ensures the controllable generation of stegotext. Compared with the existing mainstream algorithms under the public data set, the results show that the proposed method is effective in hiding capacity, concealment and security. Specifically, the LAMA-3 (3B) model which combines the entity extraction module and the sentiment analysis module has an embedding rate of 2.383 bit per word, a PPL of 16.304, a KLD of 17.952, and a BLEU of 68.864. Ablation experiments show that the entity extraction module and sentiment analysis module designed in this paper can effectively improve the hiding capacity, invisibility and security of stegotext.
Although the proposed method performs better overall, we recognize that there are still challenges with our method. The method currently achieves good performance in terms of semantics, but there are opportunities for improvement in terms of hiding capacity and security. Therefore, in the future research, one is to try to adopt the optimized steganographic coding strategy to improve the hiding capacity of the generated text. Secondly, the relationship between the distribution of real text and the distribution of stegotext can be considered, and the secure transmission of stegotext can be achieved by reducing the statistical gap between them. In addition, LLM itself brings huge computational overhead and has higher requirements for computing hardware. The introduction of entity extraction module and sentiment analysis module further increases the computational overhead. Whether the model can be improved to run on smaller devices becomes another direction for future research.

Author Contributions

Conceptualization, H.S. and W.G.; methodology, H.S. and W.G.; software, H.S. and S.G.; validation, H.S. and S.G.; formal analysis, H.S. and S.G.; investigation, H.S.; resources, W.G.; data curation, H.S. and S.G.; writing—original draft preparation, H.S.; writing—review and editing, W.G.; supervision, W.G.; project administration, W.G.; funding acquisition, W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in a GitHub repository at https://github.com/SHZXX109/dataset_IMDB (accessed on 5 May 2025).

Acknowledgments

The authors gratefully acknowledge the reviewers for their precious time and effort dedicated to this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

KeysContent
ACArithmetic coding
ACCAccuracy
BERTBidirectional encoder representations from transformers
BLEUBilingual evaluation understudy
CRFConditional random field
DLDeep learning
DNNDeep neural network
EREmbedding rate
FLCFixed-length coding
GANGenerative adversarial network
HCHuffman coding
KLDKullback-Leibler divergence
LLMLarge language model
LSTMLong short-term memory
NLPNatural language processing
PPLPerplexity
RRecall
RaNERRetrieval-augmented named entity recognition
RNNRecurrent neural network
RoBERTaRobustly optimized BERT pretraining approach
stegotextSteganographic text
VAEVariational auto-encoder
XLMcross-lingual language model

References

  1. Khan, I.; Ali, Q.E.; Hadi, H.J.; Ahmad, N.; Ali, G.; Cao, Y.; Alshara, M.A. Securing Blockchain-Based Supply Chain Management: Textual Data Encryption and Access Control. Technologies 2024, 12, 110. [Google Scholar] [CrossRef]
  2. Xu, Y.; Liu, J.; You, Z.; Zhang, T. A Novel Color Image Encryption Algorithm Based on Hybrid Two-Dimensional Hyperchaos and Genetic Recombination. Mathematics 2024, 12, 3457. [Google Scholar] [CrossRef]
  3. McAteer, I.; Ibrahim, A.; Zheng, G.; Yang, W.; Valli, C. Integration of Biometrics and Steganography: A Comprehensive Review. Technologies 2019, 7, 34. [Google Scholar] [CrossRef]
  4. Rashid, A.; Salamat, N.; Prasath, V.B.S. An Algorithm for Data Hiding in Radiographic Images and ePHI/R Application. Technologies 2018, 6, 7. [Google Scholar] [CrossRef]
  5. Selma, T.; Masud, M.M.; Bentaleb, A.; Harous, S. Inference Analysis of Video Quality of Experience in Relation with Face Emotion, Video Advertisement, and ITU-T P.1203. Technologies 2024, 12, 62. [Google Scholar] [CrossRef]
  6. Qin, Z.; Sun, C.; He, T.; He, Y.; Abdullah, A.; Samian, N.; Roslan, N.A. ADLM—stega: A Universal Adaptive Token Selection Algorithm for Improving Steganographic Text Quality via Information Entropy. arXiv 2024, arXiv:2410.20825. [Google Scholar]
  7. Cao, Y.; Zhou, Z.; Chakraborty, C.; Wang, M.; Wu, Q.M.J.; Sun, X. Generative Steganography Based on Long Readable Text Generation. IEEE Trans. Comput. Soc. Syst. 2024, 11, 4584–4594. [Google Scholar] [CrossRef]
  8. Kuznetsov, O.; Chernov, K.; Shaikhanova, A.; Iklassova, K.; Kozhakhmetova, D. DeepStego: Privacy-Preserving Natural Language Steganography Using Large Language Models and Advanced Neural Architectures. Computers 2025, 14, 165. [Google Scholar] [CrossRef]
  9. Wang, S.; Li, F.; Yu, J.; Lai, H.; Wu, S.; Zhou, W. Enhancing Semantic Consistency in Linguistic Steganography via Denosing Auto-Encoder and Semantic-Constrained Huffman Coding. In Proceedings of the 12th National CCF Conference on Natural Language Processing and Chinese Computing (NLPCC), Foshan, China, 12–15 October 2023; pp. 799–812. [Google Scholar]
  10. Shen, J.; Ji, H.; Han, J. Near-imperceptible Neural Linguistic Steganography via Self-Adjusting Arithmetic Coding. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Virtual, 16–20 November 2020; pp. 303–313. [Google Scholar]
  11. Ding, C.; Fu, Z.; Yang, Z.; Yu, Q.; Li, D.; Huang, Y. Context-Aware Linguistic Steganography Model Based on Neural Machine Translation. IEEE/ACM Trans. Audio Speech Lang. Process. 2024, 32, 868–878. [Google Scholar] [CrossRef]
  12. Badawy, I.L.; Nagaty, K.; Hamdy, A. A Comprehensive Review on Deep Learning-Based Generative Linguistic Steganography. In Proceedings of the 25th International Conference on Interactive Collaborative Learning (ICL), Vienna, Austria, 27–30 September 2022; pp. 651–660. [Google Scholar]
  13. Meral, H.M.; Sankur, B.; Sumru, Ö.A.; Güngör, T.; Sevinç, E. Natural language watermarking via morphosyntactic alterations. Comput. Speech Lang. 2009, 23, 107–125. [Google Scholar] [CrossRef]
  14. Xiang, L.; Wang, X.; Yang, C.; Liu, P. A novel linguistic steganography based on synonym run-length encoding. IEICE Trans. Inf. Syst. 2017, E100D, 313–322. [Google Scholar] [CrossRef]
  15. Lee, I.-S.; Tsai, W.-H. A new approach to covert communication via PDF files. Signal Process. 2010, 90, 557–565. [Google Scholar] [CrossRef]
  16. Xiang, L.; Guo, G.; Yu, J.; Sheng, V.S.; Yang, P. A convolutional neural network-based linguistic steganalysis for synonym substitution steganography. Math. Biosci. Eng. 2020, 17, 1041–1058. [Google Scholar] [CrossRef] [PubMed]
  17. Chen, X.; Sun, H.; Tobe, Y.; Zhou, Z.; Sun, X. Coverless information hiding method based on the Chinese mathematical expression. In Proceedings of the 1st International Conference on Cloud Computing and Security (ICCCS), Nanjing, China, 13–15 August 2015; pp. 133–143. [Google Scholar]
  18. Wang, Y.; Song, R.; Li, L.; Zhang, R.; Liu, J. Dynamically allocated interval-based generative linguistic steganography with roulette wheel. Appl. Soft Comput. 2025, 176, 113101. [Google Scholar] [CrossRef]
  19. Fang, T.; Jaggi, M.; Argyraki, K. Generating steganographic text with LSTMs. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), Vancouver, Canada, 30 July–4 August 2017; pp. 100–106. [Google Scholar]
  20. Gatt, A.; Krahmer, E. Survey of the state of the art in natural language generation: Core tasks, applications and evaluation. J. Artif. Intell. Res. 2018, 61, 1–64. [Google Scholar] [CrossRef]
  21. Chapman, M.; Davida, G. Hiding the hidden: A software system for concealing ciphertext as innocuous text. In Proceedings of the 1997 International Conference on Information and Communications Security (ICICS), Beijing, China, 11–14 November 1997; pp. 335–345. [Google Scholar]
  22. Dai, W.; Yu, Y.; Dai, Y.; Deng, B. Text steganography system using Markov chain source model and DES algorithm. J. Softw. 2010, 5, 785–792. [Google Scholar] [CrossRef]
  23. Wen, J.; Zhou, X.; Zhong, P.; Xue, Y. Convolutional Neural Network Based Text Steganalysis. IEEE Signal Process. Lett. 2019, 26, 460–464. [Google Scholar] [CrossRef]
  24. Chen, Y.; Li, Q.; Wu, X.; Li, H.; Chang, Q. A Character-based Diffusion Embedding Algorithm for Enhancing the Generation Quality of Generative Linguistic Steganographic Texts. arXiv 2025, arXiv:2505.00977. [Google Scholar]
  25. Yang, Z.; Guo, X.; Chen, Z.; Huang, Y.; Zhang, Y. RNN-Stega: Linguistic Steganography Based on Recurrent Neural Networks. IEEE Trans. Inf. Forensics Secur. 2019, 14, 1280–1295. [Google Scholar] [CrossRef]
  26. Xiang, L.; Yang, S.; Liu, Y.; Li, Q.; Zhu, C. Novel Linguistic Steganography Based on Character-Level Text Generation. Mathematics 2020, 8, 1558. [Google Scholar] [CrossRef]
  27. Yang, Z.; Wei, N.; Liu, Q.; Huang, Y.; Zhang, Y. GAN-TStega: Text Steganography Based on Generative Adversarial Networks. In Proceedings of the 18th International Workshop on Digital Forensics and Watermarking (IWDW), Chengdu, China, 2–4 November 2019; pp. 18–31. [Google Scholar]
  28. Zhou, X.; Peng, W.; Yang, B.; Wen, J.; Xue, Y.; Zhong, P. Linguistic Steganography Based on Adaptive Probability Distribution. IEEE Trans. Dependable Secure Comput. 2022, 19, 2982–2997. [Google Scholar] [CrossRef]
  29. Yang, Z.; Zhang, P.; Jiang, M.; Huang, Y.; Zhang, Y. RITS: Real-time interactive text steganography based on automatic dialogue model. In Proceedings of the 4th International Conference on Cloud Computing and Security (ICCCS), Haikou, China, 8–10 June 2018; pp. 253–264. [Google Scholar]
  30. Yang, Z.; Xu, Z.; Zhang, R.; Huang, Y. T-GRU: ConTextual Gated Recurrent Unit model for high quality Linguistic Steganography. In Proceedings of the 2022 IEEE International Workshop on Information Forensics and Security (WIFS), Shanghai, China, 12–16 December 2022; pp. 1–6. [Google Scholar]
  31. Yang, Z.; Zhang, S.; Hu, Y.; Hu, Z.; Huang, Y. VAE-Stega: Linguistic Steganography Based on Variational Auto-Encoder. IEEE Trans. Inf. Forensics Secur. 2021, 16, 880–895. [Google Scholar] [CrossRef]
  32. Tu, H.; Yang, Z.; Yang, J.; Zhou, L.; Huang, Y. FET-LM: Flow-Enhanced Variational Autoencoder for Topic-Guided Language Modeling. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11180–11193. [Google Scholar] [CrossRef]
  33. Ziegler, Z.M.; Deng, Y.; Rush, A.M. Neural linguistic steganography. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hongkong, China, 2–7 November 2019; pp. 1210–1215. [Google Scholar]
  34. Pang, K.; Bai, M.; Yang, J.; Wang, H.; Jiang, M.; Huang, Y. FREmax: A Simple Method Towards Truly Secure Generative Linguistic Steganography. In Proceedings of the 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 4755–4759. [Google Scholar]
  35. Wang, Y.; Song, R.; Zhang, R.; Liu, J.; Li, L. LLsM: Generative Linguistic Steganography with Large Language Model. arXiv 2024, arXiv:2401.15656. [Google Scholar]
  36. Qi, Y.; Chen, K.; Zeng, K.; Zhang, W.; Yu, N. Provably Secure Disambiguating Neural Linguistic Steganography. arXiv 2024, arXiv:2403.17524. [Google Scholar] [CrossRef]
  37. Simmons, G.J. The Prisoners’ Problem and the Subliminal Channel. In Advances in Cryptology; Proceedings of Crypto 83; Springer: Boston, MA, USA, 1984; pp. 51–67. [Google Scholar]
  38. Huang, Y.; Tian, C.; Narayanan, K.; Zheng, L. Relatively-Secure LLM-Based Steganography via Constrained Markov Decision Processes. arXiv 2025, arXiv:2502.01827. [Google Scholar]
  39. Xiang, L.; Wang, R.; Yang, Z.; Liu, Y. Generative Linguistic Steganography: A Comprehensive Review. KSII Trans. Internet Inf. Syst. 2022, 16, 986–1005. [Google Scholar] [CrossRef]
  40. Pang, K. FreStega: A Plug-and-Play Method for Boosting Imperceptibility and Capacity in Generative Linguistic Steganography for Real-World Scenarios. arXiv 2024, arXiv:2412.19652. [Google Scholar]
  41. Wang, Y.; Pei, G.; Chen, K.; Ding, J.; Pan, C.; Pang, W.; Hu, D.; Zhang, W. SparSamp: Efficient Provably Secure Steganography Based on Sparse Sampling. arXiv 2025, arXiv:2503.19499. [Google Scholar]
  42. Wang, X.; Shen, Y.; Cai, J.; Wang, T.; Wang, X.; Xie, P.; Huang, F.; Lu, W.; Zhuang, Y.; Tu, K.; et al. DAMO-NLP at SemEval-2022 Task 11: A Knowledge-based System for Multilingual Named Entity Recognition. In Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval), Seattle, WA, USA, 14–15 July 2022; pp. 1457–1468. [Google Scholar]
  43. Conneau, A.; Lample, G. Cross-lingual language model pretraining. In Proceedings of the 33rd Annual Conference on Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 8–14 December 2019; pp. 8–14. [Google Scholar]
  44. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  45. Wang, X.; Jiang, Y.; Bach, N.; Wang, T.; Huang, Z.; Huang, F.; Tu, K. Improving Named Entity Recognition by External Context Retrieving and Cooperative Learning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL-IJCNLP), Virtual, 1–8 August 2021; pp. 1800–1812. [Google Scholar]
  46. Wang, W.; Bi, B.; Yan, M.; Wu, C.; Xia, J.; Bao, Z.; Peng, L.; Si, L. StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding. In Proceedings of the 8th International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26–30 April 2020; pp. 1–10. [Google Scholar]
  47. Huffman, D.A. A Method for the Construction of Minimum-Redundancy Codes. Proc. IRE 1952, 40, 1098–1101. [Google Scholar] [CrossRef]
  48. Maas, A.L.; Daly, R.E.; Pham, P.T.; Huang, D.; Ng, A.Y.; Potts, C. Learning Word Vectors for Sentiment Analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies (ACL-HLT), Portland, OR, USA, 19–24 June 2011; pp. 142–150. [Google Scholar]
  49. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W. BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics (ACL), Philadelphia, PA, USA, 7–12 July 2002; pp. 311–318. [Google Scholar]
  50. Ding, C.; Fu, Z.; Yu, Q.; Wang, F.; Chen, X. Joint Linguistic Steganography With BERT Masked Language Model and Graph Attention Network. IEEE Trans. Cogn. Dev. Syst. 2024, 16, 772–781. [Google Scholar] [CrossRef]
  51. Yang, Z.; Wang, K.; Li, J.; Huang, Y.; Zhang, Y.-J. TS-RNN: Text Steganalysis Based on Recurrent Neural Networks. IEEE Signal Process. Lett. 2019, 26, 1743–1747. [Google Scholar] [CrossRef]
  52. Peng, W.; Zhang, J.; Xue, Y.; Yang, Z. Real-Time Text Steganalysis Based on Multi-Stage Transfer Learning. IEEE Signal Process. Lett. 2021, 28, 1510–1514. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.