Event Monitoring and Intelligence Gathering Using Twitter Based Real-Time Event Summarization and Pre-Trained Model Techniques

: Recently, an emerging application ﬁeld through Twitter messages and algorithmic computation to detect real-time world events has become a new paradigm in the ﬁeld of data science applications. During a high-impact event, people may want to know the latest information about the development of the event because they want to better understand the situation and possible trends of the event for making decisions. However, often in emergencies, the government or enterprises are usually unable to notify people in time for early warning and avoiding risks. A sensible solution is to integrate real-time event monitoring and intelligence gathering functions into their decision support system. Such a system can provide real-time event summaries, which are updated whenever important new events are detected. Therefore, in this work, we combine a developed Twitter-based real-time event detection algorithm with pre-trained language models for summarizing emergent events. We used an online text-stream clustering algorithm and self-adaptive method developed to gather the Twitter data for detection of emerging events. Subsequently we used the Xsum data set with a pre-trained language model, namely T5 model, to train the summarization model. The Rouge metrics were used to compare the summary performance of various models. Subsequently, we started to use the trained model to summarize the incoming Twitter data set for experimentation. In particular, in this work, we provide a real-world case study, namely the COVID-19 pandemic event, to verify the applicability of the proposed method. Finally, we conducted a survey on the example resulting summaries with human judges for quality assessment of generated summaries. From the case study and experimental results, we have demonstrated that our summarization method provides users with a feasible method to quickly understand the updates in the speciﬁc event intelligence based on the real-time summary of the event story.


Introduction
In emergencies with high market uncertainty and business interruption, real-time decision-making ability is usually a necessary condition to avoid risks.For example, the outbreak of COVID-19, earthquakes or floods have brought unexpected challenges to companies, making people realize that they should have real-time event information and resilience to immediately respond to changing external disasters.Furthermore, during such high-impact events, users may want to learn up-to-date information as the event develops because they want to know the possible trend of the event.In this case, it is almost impossible for anyone to browse through all of the news articles to understand the event stories required for decision-making.Automated summarization systems that access a large number of information sources in real time can help by providing event summaries, which are updated whenever important new events are detected.The update of the event will help those interested persons to understand the situation of the current event.With the real-time nature of social media, an automated event summary system using social media (such as Twitter) data can provide people with event updates while the situation is still evolving, without the need for individuals to manually analyze a large number of news articles.The term "real-time" here means that the summary system can expand or update the summary at any point in time, and the real-time computing property is not the focus of this work.In order to meet the above-mentioned increasing demand, in this work, we provide a solution for real-time event summarization.It combines the developed Twitter-based real-time event detection algorithm with a pre-trained language model for summarizing emergency events.
Automatically summarizing social message topics is a relatively unnoticed emerging research field.Essentially, tweet summaries can be viewed as an example of a more general automatic text summarization problem, which is an issue of automatically generating a condensed version of the most important content from user-generated text content.In addition, it is also a type of stream summarization.For text streams containing millions of documents covering various topics and events, traditional single-document and multidocument summarization methods cannot solve the challenge of information overload.Rudrapal [1] summarized the methods of automatic summarization on Twitter's most popular topics.For some highly structured and recurring events in social media messages, it is best to use more complex techniques to summarize relevant information.Chakrabarti [2] formalized the summary of event messages and proposed a hidden Markov model.To learn the solution of the non-recessive state representation of the event.Chua [3] proposed a framework for searching and summarizing on social media, extracting the most representative messages from time-series social media messages to generate continuous and concise summaries.Furthermore, the text summarization algorithms can be generally divided into two types of methods, including extractive summarization algorithms and abstractive summarization algorithms.Since our task requires extensive domain knowledge beyond the meaning of individual words and sentences, in this work, we limit our focus to abstractive summarization algorithms.
In order to generate more semantically rich sentence representations, we apply the concept of transfer learning to perform text summarization with a pre-trained language model.Transfer learning, where the model is first pre-trained on data-rich tasks, and then fine-tuned on downstream tasks, has become a powerful technology in natural language processing (NLP) field.In this work, the T5 pre-trained model [4] is used to conduct experiments for text summarization.The motivation for this work can be divided into the following two parts:

•
In a high-impact event, people may want to learn the latest information as the event develops.It is necessary to develop an effective real-time event summary system to obtain sufficient real-time intelligence information for effective decision-making.

•
In order to acquire extensive coverage of event details, using pre-trained language models can help us expand more knowledge and trending topics about how event stories develop.The benefit of this potential for event awareness deserves in-depth study.
The view is taken, therefore, in this work, we utilized the above-mentioned approaches to produce a streamlined version of a dynamically changing event data from the collected Twitter data set; when users view this version, they will be able to immediately capture the most important part of the latest event content in real-time for making appropriate decisions.It is worth mentioning that "real-time event summarization" is different from "real-time event detection" methods.The summarization system we developed can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection.The summary generated by our system can represent a simplified version of the event tweet message of interest, starting at any point in time of the ongoing event.
The rest of the paper is organized as follows.Section 2 surveys several related work in the literature.The proposed system model and system framework are described in Section 3. Section 4 reports on our pre-trained language model for text summarization.In particular, we introduce our developed text summarization method using a hybrid pre-trained model system.In Section 5, we describe our experiments in evaluation on (i) the performance of the text summarization method based on Rouge evaluation method, and (ii) the effectiveness of the real-time summary method developed for intelligence collection and event monitoring.In particular, in this work, we provide a real-world case study, namely the COVID-19 pandemic event, to verify the applicability of the proposed method.In Section 6, the discussion of experimental results is addressed in detail.Section 7 concludes the work by summarizing our research contributions.

Related Work
In this work, we attempt to build a real-time event summarization system which can generate a simplified version of a dynamically changing event data from the collected Twitter data set.When users read the generated sentences, they will be able to immediately capture the most important part of the latest event content in real time to make appropriate decisions.Using Twitter for event detection and event summarization is a vast subject.Twitter is also employed to better understand how the communication transfers during citizen movements and events.In our previous work [5,6] a real-time event detection approach has been well developed to gather tweets of detected events for awareness of emerging events in the real world.
Summarization of a single text summarization is the task of automatically generating a shorter version of a document and retaining its most important information.The types of automatic text summarization can be classified as extractive and abstractive.The extractive text summarization extracts sentences from the texts, while the abstractive ones may produce sentences and phrases which do not appear in the original documents.In practice, the techniques of abstractive text summarization emulate human summarization in that it utilizes a vocabulary beyond the specified text and summarizes key points in the text.Due to such approaches emulate how people summarize material, they normally require more computing power such as several GPUs to train the model over many days for learning algorithms, thus it is generally more difficult to acquire a cost-effective solution.In this work, we concentrate on abstractive summarization techniques.
While text summarization has been investigated for years, automatic summarization on social messages (e.g., tweets) is still in its infancy.Sharifi et al., used the Phrase Reinforcement algorithm to deal with tweet summarization [7].Inouye developed a Hybrid TF-IDF algorithm to produce multiple post summaries [8].
In order to make our model to summarize text accurately, it needs to learn to "understand" the input data.Google's BERT (Bidirectional Encoder Representations from Transformers) is the perfect tool for natural language understanding.Before the emergence of the BERT model, several important techniques related to the modeling method we adopted were developed.Recently, the Recurrent Neural Networks (RNNs; including Long Short Term Memory Networks) have been widely used for many natural language processing applications.However, it needs massive amounts of data, costly data sources, and several hours of training to achieve satisfactory results.Due to the drawback of suffering from poor performance with very long sequences of data, Vaswani proposed a model called the "Transformer", replacing the RNNs and Convolutional Neural Networks (CNNs) by utilizing an architecture comprised of feed forward networks and attention mechanisms [9].An unsupervised learning architecture called BERT (Bidirectional Encoder Representations from Transformers) has been developed and outperformed almost all existing models in the competitions for various types of NLP tasks [10].In addition, they published several pre-trained models which can be utilized for transfer learning for many tasks and applications [10].
Text summarization is a technology that extracts the important text or phrase information from the original text and then presents a short message to the user.According to the content and method of abstracting, it can be divided into Extractive and Abstractive.Extractive summarization is a technique that uses statistics that extracts the most important words and phrases in a text according to the given weight, and then arranges and combines these features to generate a summary of the text; abstractive summarization is a technique that uses machine learning methods to generate a summary of the text after understanding the content of the text [11][12][13][14].
Most of the extractive summarization are generated by extracting human-designed words or phrases as features in the original text [15][16][17], so that these extracted features are the most useful and important words and phrases in the text to achieve a more accurate final summary [18].These features are usually rated according to their importance in the title, word frequency, word position, sentence length, similarity between sentences, number of occurrences of words, and famous words [19], and then rearranged and combined according to this rating to generate the summary.The method has been studied in various applications [20][21][22].For example, Isonuma [23] proposed a framework for extractive summarization using the dataset NIKKEI financial report corpus and the New York Times annotation corpus.Nallapati [24] proposed a Recurrent Neural Network (RNN) to extract the words and phrases that are relevant to the topic.Nallapati [24] proposed SummaRuN-Ner, a language model based on Recurrent Neural Network (RNN), which has a different training mode from the usual models and can reduce the need for extractive feature labels during training.Gunawan [25] used sentence length, headline features, and keyword frequency to develop an unsupervised summarization model for 3 K articles.Narayan [26] summed up the concept of extracted summaries as a task for ranking sentence and presents a training method that uses reinforcement learning to optimize the reward shaping that related to the task by exploring the space of candidate summaries.Dong [27] presents a framework for extractive summarization through the neural networks and enhanced learning contexts, which did not require sentence-level extraction labels on extracted summaries and optimized the ROUGE score of model that generates the summary.
Rush [28] presented a Neural Attention Model for Abstractive Sentence Summarization, combining the chance model with the generation algorithm to produce an accurate summary, and acquires good results in the Gigaword and DUC datasets, on which Chopra [29] further improves the performance of the summary after replacing the decoder with an RNN.Song [30] proposed an LSTM-combined CNN abstractive text summarization framework (ATSDL), which extracts key phrases from the text and generates a smooth and grammatical summary by deep learning models.Paulus [31] proposes a neural network model for dealing with longer textual data, and the state-of-the-art is achieved in the experiment using 12 data sets.Chen [32] proposed a sentence-level learning method that allows abstractive summarization models to learn the hierarchical nature of words and sentences.Masum [33] used a bidirectional RNN with an LSTM as the encoding layer and an attention model as the decoding layer to create a short, fluent and comprehensible summary of a text document.Cai [34] proposed an abstractive text summarization model RC-Transformer (RCT), which adds an RNN encoder to capture the context of the sequence and adds a convolutional layer to extract important sentences and understand the semantics to generate a paragraph of extremely high quality summary.
In recent years, transfer learning methods using pre-trained models have achieved good results in most natural language processing tasks [35].The self-supervised pretraining of the model on a large amount of unlabeled data will affect the effect of transfer learning, such as building a language model or filling in the blanks of words.Therefore, finetuning a small data set in the pre-trained model will be more effective.Devlin [10] proposed the BERT (Bidirectional Encoder Representations from Transformers) model in 2018, which has had a great impact on the field of natural language processing.BERT provides a pretrained bidirectional feature for unlabeled text data in the context of all levels.Automatic text summarization is a hot research topic in the field of natural language processing.Many studies use deep learning technology to obtain important information from text data, so that the machine can understand the text and generate a text-related summary.Therefore, many recent studies have applied pre-training models to text summarization.Liu [36] proposed a framework built using BERT.The framework can provide extractive and abstractive types of text summarization.The experimental results on three data sets (CNN/DailyMail, NYT and XSum) show that their model achieves the state-of-the-art result under the automatic and human-based evaluation protocol.Kieuvongngam [37] used the BERT model and GPT-2 model to conduct text summarization on the open research data set of COVID-19 and extracted the keywords and semantic meaning of the sentences from the original text data through the above-mentioned pre-trained model.Their system provides a concise and meaningful summary for the medical literature.Khandelwal [38] used the pre-trained Transformer language model to generate summaries from the original text data.Compared with the pre-trained Transformer encoder-decoder network, this model has significantly improved results.Zhang [39] proposed a pre-trained model HIBERT (Hierarchical Bidirectional Encoder Representations from Transformers) for pre-training on text data and unlabeled data.This model can achieve higher results on the CNN/Dailymail data set and the New York Times data set.Therefore, the experimental results show that the use of pre-trained models can make large open data sets acquire better results.Farahani [40] mentioned two approaches to deal with the task of text summarization using mT5 model and ParsBERT model and got good results on a data set named pn-summary for Persian abstractive text summarization.Ma [41] proposed a pre-trained model T-BERTSum for text summarization, which captures the key words of the topic information of social media, understands the meaning of the sentence and judges the topic of the message discussion, and then generates a high-quality section Summary.Kerui [42] uses BERT, Seq2seq and reinforcement learning to form a text summary model.Garg [43] uses T5, one of the most advanced pre-trained models, to perform a summary task on a data set with 80,000 news articles, and the results indicate that the summary generated by T5 has better quality than those generated by other models.Daiya [44] has developed a pre-trained language model ENEMAbst that can be used in extractive and abstractive summarization techniques.Both abstractive summaries and extractive summaries on the CNN/DailyMail dataset obtain the most excellent results.Fecht [45] proposed a pre-trained language model that can be used for abstractive text summarization, and combined with the use of BERT to experiment on the German summarization data set.After performing the contextual word embedding task through BERT, the model can improve the model's ability to generate high-quality German summaries.In Table 1, some important related work is shown for comparison.
In summary, in our work, social media (such as Twitter), pre-trained models, and text summaries (including extractive and abstractive methods) are the three key elements of event summarization.From the above literature review, the text summaries of specific event tweets on social media have gradually become one of the most popular summarization research topics in recent years [1][2][3]41,42].However, extractive summarization cannot obtain many key sentences in the tweet-form data, resulting in unsatisfactory quality of the generated summaries.Furthermore, due to our task requires extensive domain knowledge beyond the meaning of individual words and sentences, in this work we limit our focus to abstractive summarization algorithms.On the other hand, in the recent references of text summarization, there are more and more work using pre-training language models [35][36][37][38][39][40][41][42][43][44].Among them, BERT is the pre-trained model mainly used by most research teams.Most researchers use the improved BERT model architecture in their work [36,37,[39][40][41].However, because the sentences in the pre-training data set of BERT have correct grammar, BERT does not perform well for tweets with less grammar.In order to simplify the summarization process, the text-to-text pre-trained model (e.g., T5 model) has become a good choice.Unfortunately, for our problem domain, so far little attention has been paid to the development of a summary method that supports real-time functions to achieve the requirements of real-time event monitoring and intelligence collection.Motivated by this, in this work, we developed a novel model combining a real-time event summarization system and a Twitter-based event detection method.The developed summarization system can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection.To the best of our knowledge, this is the first attempt to develop such an event summarization system with multiple functional requirements to meet the challenge of intelligence gathering in emergency situations.T-BERTSum Yes Abstractive Zhang et al. [39] HIBERT Yes Abstractive Liu et al. [36] BERT Yes Abstractive and Extractive Daiya [44] ENEMABst Yes Abstractive and Extractive

System Framework
In this work, the pre-trained models are mainly used to perform summarization function on social media data to fine-tune the text summaries.With the rapid development of social media, many messages flow and gather on social media, forming various event topics.However, identifying "event" is one of the most difficult challenges of current decision support systems.If the system encounters an emergency, it is usually unable to promptly notify users of the warning to avoid risks.A sensible solution is to employ the functions of the real-time event summarization to obtain emergent event information.As a result, the system framework of this study utilizes the method of event detection on social-media and messages to extract their main content, and then summarized to understand the evolution of important events.Figure 1 illustrates the conceptual diagram of the process followed to extract real-time event intelligence for both the event detection and summarization.The system is designed as a two-pass process, consisted of two modules, say, real-time event detection module and event-message accumulation and summarization module.More details on real-time event summarization are presented in the following.Firstly, in this work, we combine a developed Twitter-based real-time event detection algorithm with pre-trained language models for summarizing emergent events.We used an online text-stream clustering algorithm and self-adaptive method developed to gather the Twitter data for detection of major international events.The unimportant messages on the data set are removed through the trained SBERT based classifier.After that, the DBSCAN algorithm is used to cluster similar topics in each day into hot topics (i.e., event topics).These event topics in Twitter data are generated by using the algorithm we developed to generate storylines.When the hot topics of each day are summarized, the specific topic and other hot topics are semantically correlated.When the correlation exceeds a certain threshold, it is added to the storyline.After daily updates, through the Twitter based real-time intelligence data collection, a storyline of a specific event is jointly generated.The next work is to deal with text summarization of specific topics.In this work, we use the pre-trained model (i.e., T5; Text-to Text Transfer Transformer) to experiment with the Xsum data set to perform text summarization, and the Rouge metrics are used to evaluate the resulting summaries.The Xsum [46] data set contains 226,711 news articles and their summaries.News articles come from the BBC's content from 2010 to 2017.The articles cover many fields such as news, politics, sports, weather, business, technology, science, health, family, education, entertainment and art, etc.We split the data set into 90% training data set, 5% validation data set and 5% test data set.The pre-trained language models and pre-training process will be described later in more detail.
Subsequently, we started to use the trained model to summarize the incoming Twitter data set for experimentation.The large amount of events and related tweets detected by the real-time event detection system are grouped into about hundreds of event topics every day.Taking the Twitter data in 2020 as an example, there are about 36,000 event topics on Twitter.Using the resulting models of text summarization on popular topics of social media in this work, the system is used to perform real-world event summarization tasks and analyze the results.More details about pre-trained language models for text summarization are presented in the following section.

Proposed Approach for Collectively Summurizing Text Streams
The first part is the real-time event detection algorithm, which analyzes the life cycle of events and the distribution of word weights through dynamic term weighting [47] in the instant message stream.In the early stage of the event, potential major events are captured, and then the real-time event cluster is combined into an event sequence through the self-adaptive stream clustering method.The algorithm for real-time event detection is shown in Algorithm 1. for each TS{i} 3: BursT ← Dynamic Term Weighting(i) 4: RTED ← Self-Adaptive Stream Clustering(BursT, i) 5: end for 6: End Dynamic Term Weighting can be divided into two factor values: Burst Score, BS and Term Importance, TI, as shown in Formula (1).In Formula (2), Burst Score is to calculate the burstiness of a keyword according to the rate at which the keyword arrives.In Formula (3), the key word weight is the logarithm of the ratio of the number of documents in the global time window to the number of messages containing the keyword to determine the popularity of the keyword [47].
Self-adaptive stream clustering algorithm [48] is used to represent the situation that, when a new message flows into the system, the message will be compared with the existing online group for similarity.When the similarity comparison result is greater than the threshold, the new message will be added to the most similar group in the online group.If it is less than the threshold, the system will create a new group and add the message.At the same time, the system will remove non-informative messages and aging online groups based on the adjustable sliding window module and the time decay function.The proposed self-adaptive stream clustering algorithm is shown in Algorithm 2.
Once the adaptive stream clustering algorithm is used to process stream messages in real time, the number and weight of features will be automatically adjusted within the thread time width tw.As shown in the Algorithm 2, when a new message flows into the system, the message will be compared with the existing online cluster for similarity evaluation.When the similarity comparison result is greater than the threshold, the new message will be added to the most similar cluster.If it is less than the threshold, the system will create a new cluster and add the message.At the same time, the system will remove non-informative messages and aging online clusters based on the adjustable sliding window.The second part of the event-message accumulation and summarization algorithm is to filter the event information through the BERT classifier, and then use the vector of the filtered message obtained by the Sentence-BERT model and put it into the DBSCAN algorithm to cluster the event information into For each sub-event, the summary content is finally obtained through the T5 model.The proposed method for event-message accumulation and summarization algorithm is shown in Algorithm 3. Text ← BERT Classifier Model(TM{i}) 4: Sentence Embedding ← Sentence-BERT Model(Text) 5: HT ← DBSCAN(Sentence Embedding) 6: for each HT{j} 7: HTS ← T5 Text Summarization Model(j) 8: end for 9: end for 10: End

Proposed Pre-Trained Language Models for Text Summarization
In this study, the pre-trained language model is applied to social media data to finetune the text summaries.With the development of the Internet generation, people prefer to express their feelings or the content they are interested in on social media, and many messages flow on social media every day.When there is more and more information on a specific topic, it is easy to form a hot topic.This research will focus on summarizing popular topics when social media content forms as the main core content, and then summarize them to explore the evolution of popular topics.
Transfer learning, where the model is first pre-trained on data-rich tasks, and then fine-tuned on downstream tasks, has become a powerful technology in natural language processing (NLP) field.In this work, the T5 [4] pre-trained model is used to conduct experiments and the Rouge metrics are used to evaluate the resulting summaries.In addition, we also use the BERT pre-training model for experiments as a baseline for performance comparison with the T5 model.The concepts of T5 and BERT pre-trained models are briefly described in the following subsections.

Bert Model and Sentence-BERT Model
The establishment of the BERT model uses other techniques for training contextdependent representation models, including Semi-supervised Sequence Learning, GPT [49], ELMo [50] and ULMFit [51], but these models are unidirectional or shallow Bidirectional, which makes these models only use the words on their left or right to learn context, while BERT uses a deep bidirectional architecture so that the model can learn the words on its left and right.The BERT model also has good results in the task of semantic similarity, but when the BERT model calculates the semantic similarity, it needs to put two sentences into the model at the same time and calculate the information of the two sentences.This approach requires a lot of For calculation and time, if there are N sentences, it needs to calculate (N * (N + 1))/2 times, and the similarity retrieval of 10,000 sentences requires about 50 million calculations, and it takes about 50 million times to execute using the BERT model.65 h, which makes the BERT model unsuitable for semantic similarity retrieval and cluster analysis.
Reimers proposed a model Sentence-BERT (SBERT) to improve the BERT architecture, using the twin and triplets network architecture to generate a sentence embedding that can compare semantics with cosine similarity, making it possible to calculate the similarity of 10,000 sentences The time required for the BERT is shortened from 65 h to 5 s, and the accuracy of the similarity calculation using the BERT model can still be maintained.SBERT added a layer of pooling operation behind the output layer of BERT to generate a revised sentence embedding.The maximum value is used as the sentence vector of the entire sentence.In Reimers' experiment, the Mean-Pooling method is used as the sentence vector to acquire the best results.

T5 (Text-to-Text Transfer Transformer) Model
T5 is a transformer-based model from Google that is trained in an end-to-end manner with text as input and modified text as output.It achieves state-of-the-art results on multiple NLP tasks such as summarization, question answering, machine translation using a text-to-text transformer trained on a large text corpus.The T5 model is composed of the encoder-decoder scheme of transformer model.The encoder and decoder on the T5-Small model are composed of 6-layer Transformer, the T5-Base model is composed of the BERT-base encoder and decoder, and the T5-Large model is composed of the BERT-large encoder and decoder group.T5-3B and T5-11B each increase the size of the model by 4 times.One of the goals of pre-training process is to provide general knowledge for the model to improve its performance on downstream tasks.T5 uses Transformer's Encoder-Decoder model during pre-training process.In the transfer learning of natural language processing, previous studies have used large unlabeled data sets for unsupervised learning.T5 attempts to measure the impact of the quality, characteristics and size of these unlabeled data sets, using C4 (Colosal Clean Crawled Corpus) corpus is used as the training data.

Developed Text Summarization Method Using a Hybrid Pre-Trained Model System
Although there are a lot of content from social media platforms, there are often some unimportant daily message content.Therefore, when data is processed, tweets are also filtered through classification models to filter out unimportant daily tweets.We use about 200,000 HuffPost news headline data sets from 2012 to 2018 [52].This data set contains 40 different categories of news headline categories, plus the unimportant daily tweet content from our collected tweets and mark it as the 41st category 'OTHERS'.The text data are classified into 41 categories in total.The classification data is divided into three parts, 60% of the data is used to train the model, 20% of the data is used to verify whether the model has learned useful information, and 20% of the data is used to test the accuracy of the model.Since there are about 40,000 to 500,000 non-retweeted tweets in the daily collected tweets, which contain a large number of daily trivial tweets, such as happy birthday, and holiday wishes, in this work the messages are filtered to avoid too much unnecessary data and reduce the calculation time.In order to alleviate this problem, we decided to use the characteristics of the BERT model as a classier to perform the function of tweet filtering.Subsequently, the filtered data is clustered through an unsupervised learning algorithm, and the content of the tweet is retained by the density of the content of the tweet, so as to ensure that the tweet is a hot topic on the day.
As a result, in this system, the tweets are aggregated into a daily form and screened and filtered the content of tweets that were not reposted in English, and filtered the tweets through the BERT classification method, using the HuffPost data set plus the tweet content from the tweets.The nonessential daily tweets are filtered out and divided into 41 categories in total.The filtered tweets are processed through the SBERT (Sentence-BERT) model to obtain sentence vectors, and the Euclidean distance between the vectors is used to calculate the similarity between sentences.Here we use the DBSCAN algorithm to find more intensive discussion content, and finally form a daily hot topic.This study will learn text summaries by fine-tuning the pre-trained model in the Xsum data sets.In the experiment, the T5 pre-trained model is firstly used for text summarization, as shown in Figure 2. Subsequently, we compare the T5 model with the BERT pre-trained model for summarization performance comparison.

Evaluation Method for Text Summarization
This research work mainly uses Rouge (Recall-Oriented Understudy for Gisting Evaluation) [53] as the evaluation method of text summarization.It uses automatically generated summaries and a set of reference summaries to do calculations to measure the similarity between automatically generated summaries and reference summaries.Rouge-N is used to calculate the number of n-grams shared by the generated summary and the reference summary, as shown in Formula (4).
Rouge-L is used to match the longest common subsequence (Longest Common Subsequence, LCS) between two texts.

Experiment and Case Study
We designed two sets of experiments to evaluate (i) the performance of the text summarization method based on Rouge evaluation method, and (ii) the effectiveness of the real-time summary method developed for intelligence collection and event monitoring, through the implementation of a case study in the real world.

Experiment and Performance Comparison
We first performed experiments on the Xsum dataset on the BERT and T5 pre-trained models.After the evaluation of the text summaries, it can be found that the performance and output length of different pre-training models used in Tables 2-4 are different.Compared with BERT, T5 is more excellent in the performance of text summarization.The training time spent is much less than that of BERT, so the T5-base pre-trained model and the output length of 150 parameters are finally used in the text summary tool.Table 5 shows the result with the complete training data and most suitable parameters in the experiment.In order to save the training cost of the pre-trained model, the training data is compressed to find the most suitable parameters.Table 6 illustrates a comparison with other models.

Xsum and Tweet Data Sets for Fine-Turning and Downstream Task
In this study, we use the Xsum data set and a pre-trained language model, the T5 model, to train the summary model.Since the downstream task is to generate a summary of the tweet data set, the features required for the summary generated in the proposed system are mainly learned by fine-tuning the pre-trained T5 model in the Xsum data set.The Xsum dataset contains 226,711 BBC news articles and their summaries.Articles cover news, politics, sports, weather, business, technology, science, health, family, education, entertainment, art and other fields.The coverage and diversity of Xsum are very suitable for our downstream task, which is to summarize a large number of tweets.Therefore, after the fine-tuning process, the well-trained model is used in massive tweet data sets for summarization.

Detected Hot Topics with Twitter Data
In the hot topic detection system, a large amount of social media content is integrated into about hundreds of hot topics every day.Taking the Twitter data in 2020 as an example, there are about 36,000 hot topics in total, as shown in Figure 3. Based on the method of text summarization on popular topics in social media, the technique is applied to the task of text summarization in various fields on social media and specific data analysis.

Case Study:Tweet Summary of the "COVID-19 Pandemic" Event
In 2020, COVID-19 is undoubtedly one of the hottest topics.The COVID-19 pandemic has exposed many serious errors in global public policy preparation and response.Although it has been affected by other epidemics and pandemics for more than a decade, many countries have failed to incorporate lessons learned into their pandemic preparedness and response plans.One of the main factors is that we lack effective real-time event-related intelligence collection systems which can provide a summary of updates related to the COVID-19 pandemic in all countries.Hence, we cannot learn from the successful experience of other countries in fighting the epidemic in time.Since the COVID-19 outbreak, very few studies that respond to practical needs, such as through rapid advice guidelines or application modules, have been conducted.However, many professionals, such as health workers, are in urgent need of effective ways to help people improve their well-being during this pandemic.According to the US Centers for Disease Control and Prevention (CDC), information on the outbreak of the coronavirus disease may increase the levels of distress among specific population-especially in those who were already struggling with mental health issues before the pandemic.They may be likely to experience higher levels of worry and stress regarding issues surrounding COVID-19.This trepidation may worsen over time and cause a long-lasting mental disturbance [54].We wish that the proposed system can help the professionals detect and track COVID-19 events at an early stage as well as monitoring the emergency events within specific groups.
The view is taken, therefore, in this case study, we combine our developed Twitterbased real-time event detection algorithm with pre-trained language models for summarizing COVID-19 events and sub-events.
In Figure 4, it can be found that there are many detected hot-event topics around COVID-19 in 2020.In order to satisfy users' information needs, we pick up all of the real-time event topics with a comprehensive spatio-temporal viewpoint.By analyzing the events detected by our system, the spatial and temporal impacts of the emerging events can be estimated.Therefore, we first build a representation of the spatial distribution by mapping the summary of events related to COVID-19 in the world in 2020, as shown in Figure 5.After that, in terms of temporal aspects, we constructed a storyline retrospection mechanism that demonstrates the story telling capability for users to understand the context of event development in a timeline, as illustrated in Figures 6 and A1-A3.In Figure 5, we present hot topics in the form of a map to illustrate global events to understand the spatial impact of the COVID-19 pandemic on the world.
Furthermore, in this work, we developed a storyline retrospection mechanism that demonstrates the story telling capability for users to understand the context of event development in an efficient way.It can be used to identify events and their relations under an event topic and then composes a storyline which gives users the sketch of event evolution in the topic, by computing daily related tweets and summaries of formulated event topics.The main storyline construction removes irrelevant events and presents a main theme that exhibits the sequence of events in a hot topic.
In Figures 6 and A1-A3, we can see the storyline of COVID-19 in 2020, from the beginning of the outbreak to the emergence of the final vaccine.Figure 7 shows the example resulting summaries of COVID-19 event related tweets in 2020, starting with a new virus related to the SARS outbreak that may cause Chinese health officials to discuss on social media Pneumonia.From the time trend shown in Figures A1-A3, when the COVID-19 pandemic first broke out, social media were mainly discussing where the pandemic occurred and how to properly protect themselves from the pandemic.After a period of time, the social media began to gradually appear on the topic of what drugs may be able to treat the symptoms of COVID-19 or the topic of vaccine-related research and development.As time progressed, social media began to see related topics such as whether the vaccine was safe.

Human Evaluation for Quality Assessment of Generated Summaries
order to verify the quality of the text summary results, we conducted a survey on the example resulting summaries (in Figures 7 and A4-A11) with human judges for quality assessment of generated summaries.In this work, we invited 39 participants from University graduate students, including 19 males and 20 females.Their age ranges from 20 to 30 years old.Participants were asked to rate the summaries on a 1-5 scale, with higher being better.The assessed summaries were selected using random sampling.Thus, we randomly selected 10 sample summaries from our summarization system for evaluation.In this work, the summaries are evaluated from three aspects.The first aspect is to assess the consistency of the topic between the tweets and the summary.The second aspect is to evaluate the closeness to the "facts" of the summary.The third aspect is to assess the degree of grammatical correctness of the summary.The summaries are reviewed on a scale of 1 to 5 points, and the evaluation results are shown in Table 7. Options for grading summaries are as follows: 1: Bad 2: Below Average 3: Average 4: Good 5: Excellent.In Table 7, we find that the mean values (i.e., 3.92, 3.89, 3.92, respectively) are close to 4 points, which show that the experimental result is acceptable and very close to the performance of human.In addition, we used Cronbach's alpha test to check whether the multi-question Likert scale survey is reliable and acquires results (Cronbach's alpha coefficients are 0.725, 0.812, 0.846).It shows that Cronbach's alpha coefficient is greater than 0.7, and the resulting coefficient indicates that the results are reliable.

Discussion
In this work we use natural language processing techniques to deal with social media data, extract event topics from daily social messages, and summarize them to quickly grasp various event intelligence information.In the experiment, we adopted a pre-trained model (i.e., T5) method to quickly fine-tune the model to learn text summaries through the Xsum data set, which was achieved a score of 38.13 in the evaluation of Rouge-1.After that, the trained model is actually applied to real social media data.We started to use the trained model to summarize the incoming Twitter data set for experimentation.The model summarizes the collected hot topics of social media into a short sentence.The resulting summary can help users quickly understand topic trends on social media, most of which are in line with the development of real-world events.Furthermore, in order to verify the quality of the text summary results, we conducted a survey on the resulting summaries with human judges for quality assessment of generated summaries.The experimental result is discussed in detail as follows: (1) The T5 model is trained in an end-to-end and text-to-text manner, with text as input and summary text as output.In this work, we demonstrated that T5 has achieved state-of-the-art results in many NLP benchmarks and can be fine-tuned to perform automatic summarization tasks well.(2) As mentioned earlier, due to the real-time nature of social media, an automatic event summary system using Twitter data was developed in this work to provide people with an updated summary of the event while the situation is still evolving.However, it is worth mentioning that the "real-time" here means that the summary system can expand or update the summary at any point in time, rather than real-time computing.(3) Going further, "real-time event summarization" is different from "real-time event detection" methods.The summarization system we developed can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection.Compared to real-time event detection methods, detected events on Twitter are often short-term.The summary generated by our system can represent a simplified version of the event tweet message of interest, starting at any point in time of the ongoing event.(4) Generally speaking, abstractive summarization models are prone to produce grammatically correct incoherent sentences.In the work of summarizing text, it is easy for human researchers to make language errors.The advantages of using abstractive summarization models enable us to alleviate some of the grammatical problems of real people.Despite the results of the system evaluation show that the summaries we got are not perfect in terms of grammatical fluency, the results are acceptable and close to the performance of human researchers in the real world.( 5) Although this developed model can quickly summarize the text, it still has certain limitations in generality.The social media data used in this study only uses English data on Twitter to generate summaries.Although English is the common language of most people in the world, it must be only through English to understand the various events in the world, there are still some prejudices.In the future, we consider using multilingual models and multilingual data sets (made by different native speakers) to train the model so that we can more accurately grasp various world events from different perspectives.(6) In this work, we implemented a method of generating storylines using the automatic summarization technique we developed.The generated storyline is composed of sub-events or key events related to a given major event.The sub-events show the status of the main event in progress.Automated storyline generation has been a research issue since nearly the inception of artificial intelligence.In particular, the storylines of events relate to identify entities and summarize events that leads to the event of interest.However, the proposed system model has not yet performed the full function of the "automated storyline generation" paradigm for understanding event development.This is because in this work, we have not provided methods for analyzing the evolution of specific event entities and exploring the entities and their relationships in event stories.We will leave the work of constructing event storyline for future work to complete.

Conclusions
The uncertainty of emergencies and natural disasters may affect companies and everyone's lives.During such high-impact events, people may want to stay up-to-date as the event develops because they want to better understand the situation and possible trend of the event.In this case, it is almost impossible for anyone to browse through all of the news articles to understand the event stories required for decision-making.Automated summarization systems that access a large number of information sources in real time can help by providing event summaries, which are updated whenever important sub-events are detected.The update of the event will help those interested persons to understand the situation of the current event.
In this work, we developed a hybrid method to combine a real-time event summarization system and a self-adaptive Twitter-based event detection method, using pre-trained language models on the tweet data set for online machine learning.Hence, each user can obtain real-time information about specific incidents (or events) and the agility to make appropriate decisions quickly.The goal of this work can be divided into the following two parts: (1).In a high-impact event, people may want to learn the latest information as the event develops.It is necessary to develop an effective real-time event summary system to obtain sufficient real-time intelligence information for effective decision-making; (2).In order to acquire extensive coverage of event details, using pre-trained language models can help us expand more knowledge and trending topics about how event stories develop.Our research is intended as a starting point for exploring the benefits of this potential for event awareness.
The novelty and contributions of the work are summarized as follows: • We implemented an end-to-end and text-to-text summarization system model, in which text streams as input and summary text as output.

•
Our system model combined a real-time event summarization system and a selfadaptive Twitter-based event detection method.The developed summarization system can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection.

•
The developed system used pre-trained language models to help users obtain more knowledge and trending topics about how event stories develop.

•
The system can support users to obtain sufficient real-time intelligence information in specific events to make effective decisions.
For future work, it will be interesting to further incorporate the sentiment of the tweet into the summary of related events to predict the world market trends affected by the event.On the other hand, as mentioned earlier, building a summary system in the "automatic storyline generation" paradigm is very useful for understanding the details of event development.In this way, it is possible to explore the evolution of a specific event entity in the event story and the relationship between the entity and its event.In addition, by taking advantage of the potential of pre-trained language models, it will be very useful to gather intelligence on multilingual event summaries from multilingual social media messages.Finally, the use of automatic video summarization methods to extend the developed real-time text summarization methods to generate event summaries is also a promising application.Figure A10 shows that it was discussed in social media on 16 November 2020.American drug maker Moderna stated that its COVID vaccine can effectively prevent the coronavirus.

Figure 1 .
Figure 1.Proposed Twitter based real-time event summarization and pre-trained models.

Figure 2 .
Figure 2. Developed text summarization method using a hybrid pre-trained model system.

Figure 3 .
Figure 3. Daily detected event hot topics in 2020.

Figure 5 .
Figure 5. Representation of a map of generated summaries related to COVID-19 events in 2020.

Figure
FigureA6illustrates the topics that will be postponed for the 2020 Olympic Games discussed on 30 March 2020.

Figure
FigureA7shows that the possible variants of the new coronavirus were discussed in social media on 6 May 2020, and they may become more toxic and more contagious than the original.

Figure A7 .
Figure A7.A sample resulting summary of formulated event topics (6 May 2020).

Figure
FigureA8shows that the Dexamethasone drug was discussed in social media on 17 June 2020 to help save the lives of patients with the new coronavirus.

Figure
Figure A9 shows that on 15 July 2020, in social media, the Center for Disease and Control of the United States called on people to wear masks to effectively prevent the new coronavirus.

Figure
FigureA11shows that Pfizer's COVID vaccine was discussed in social media on 9 December 2020, in compliance with the safety and effectiveness guidelines for emergency use.

Table 1 .
Related work of automated text summarization.

Table 5 .
Performance evaluation (using complete training data).

Table 7 .
Quality assessment of generated summaries.