Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Event Monitoring and Intelligence Gathering Using Twitter Based Real-Time Event Summarization and Pre-Trained Model Techniques

Event Monitoring and Intelligence Gathering Using Twitter Based Real-Time Event Summarization and... applied sciences Article Event Monitoring and Intelligence Gathering Using Twitter Based Real-Time Event Summarization and Pre-Trained Model Techniques 1 , 2 3 1 Chung-Hong Lee *, Hsin-Chang Yang , Yenming J. Chen and Yung-Lin Chuang Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung 807618, Taiwan; assh70057@gmail.com Department of Information Management, National University of Kaohsiung, Kaohsiung 811726, Taiwan; yanghc@nuk.edu.tw Management School, National Kaohsiung University of Science and Technology, Kaohsiung 824005, Taiwan; yjjchen@nkust.edu.tw * Correspondence: leechung@mail.ee.nkust.edu.tw Abstract: Recently, an emerging application field through Twitter messages and algorithmic com- putation to detect real-time world events has become a new paradigm in the field of data science applications. During a high-impact event, people may want to know the latest information about the development of the event because they want to better understand the situation and possible trends of the event for making decisions. However, often in emergencies, the government or enterprises are usually unable to notify people in time for early warning and avoiding risks. A sensible solution is to integrate real-time event monitoring and intelligence gathering functions into their decision support system. Such a system can provide real-time event summaries, which are updated whenever Citation: Lee, C.-H.; Yang, H.-C.; important new events are detected. Therefore, in this work, we combine a developed Twitter-based Chen, Y.J.; Chuang, Y.-L. Event real-time event detection algorithm with pre-trained language models for summarizing emergent Monitoring and Intelligence events. We used an online text-stream clustering algorithm and self-adaptive method developed Gathering Using Twitter Based to gather the Twitter data for detection of emerging events. Subsequently we used the Xsum data Real-Time Event Summarization and set with a pre-trained language model, namely T5 model, to train the summarization model. The Pre-Trained Model Techniques. Appl. Rouge metrics were used to compare the summary performance of various models. Subsequently, we Sci. 2021, 11, 10596. https://doi.org/ started to use the trained model to summarize the incoming Twitter data set for experimentation. In 10.3390/app112210596 particular, in this work, we provide a real-world case study, namely the COVID-19 pandemic event, to verify the applicability of the proposed method. Finally, we conducted a survey on the example Academic Editor: Andrea Prati resulting summaries with human judges for quality assessment of generated summaries. From the case study and experimental results, we have demonstrated that our summarization method provides Received: 25 September 2021 users with a feasible method to quickly understand the updates in the specific event intelligence Accepted: 5 November 2021 Published: 11 November 2021 based on the real-time summary of the event story. Publisher’s Note: MDPI stays neutral Keywords: text summarization; event detection; machine learning; natural language processing; Twitter with regard to jurisdictional claims in published maps and institutional affil- iations. 1. Introduction In emergencies with high market uncertainty and business interruption, real-time decision-making ability is usually a necessary condition to avoid risks. For example, the Copyright: © 2021 by the authors. outbreak of COVID-19, earthquakes or floods have brought unexpected challenges to Licensee MDPI, Basel, Switzerland. companies, making people realize that they should have real-time event information and This article is an open access article resilience to immediately respond to changing external disasters. Furthermore, during distributed under the terms and such high-impact events, users may want to learn up-to-date information as the event conditions of the Creative Commons develops because they want to know the possible trend of the event. In this case, it is Attribution (CC BY) license (https:// almost impossible for anyone to browse through all of the news articles to understand the creativecommons.org/licenses/by/ event stories required for decision-making. Automated summarization systems that access 4.0/). Appl. Sci. 2021, 11, 10596. https://doi.org/10.3390/app112210596 https://www.mdpi.com/journal/applsci Appl. Sci. 2021, 11, 10596 2 of 29 a large number of information sources in real time can help by providing event summaries, which are updated whenever important new events are detected. The update of the event will help those interested persons to understand the situation of the current event. With the real-time nature of social media, an automated event summary system using social media (such as Twitter) data can provide people with event updates while the situation is still evolving, without the need for individuals to manually analyze a large number of news articles. The term “real-time” here means that the summary system can expand or update the summary at any point in time, and the real-time computing property is not the focus of this work. In order to meet the above-mentioned increasing demand, in this work, we provide a solution for real-time event summarization. It combines the developed Twitter-based real-time event detection algorithm with a pre-trained language model for summarizing emergency events. Automatically summarizing social message topics is a relatively unnoticed emerging research field. Essentially, tweet summaries can be viewed as an example of a more general automatic text summarization problem, which is an issue of automatically generating a condensed version of the most important content from user-generated text content. In addition, it is also a type of stream summarization. For text streams containing millions of documents covering various topics and events, traditional single-document and multi- document summarization methods cannot solve the challenge of information overload. Rudrapal [1] summarized the methods of automatic summarization on Twitter ’s most pop- ular topics. For some highly structured and recurring events in social media messages, it is best to use more complex techniques to summarize relevant information. Chakrabarti [2] formalized the summary of event messages and proposed a hidden Markov model. To learn the solution of the non-recessive state representation of the event. Chua [3] proposed a framework for searching and summarizing on social media, extracting the most represen- tative messages from time-series social media messages to generate continuous and concise summaries. Furthermore, the text summarization algorithms can be generally divided into two types of methods, including extractive summarization algorithms and abstractive sum- marization algorithms. Since our task requires extensive domain knowledge beyond the meaning of individual words and sentences, in this work, we limit our focus to abstractive summarization algorithms. In order to generate more semantically rich sentence representations, we apply the concept of transfer learning to perform text summarization with a pre-trained language model. Transfer learning, where the model is first pre-trained on data-rich tasks, and then fine-tuned on downstream tasks, has become a powerful technology in natural language processing (NLP) field. In this work, the T5 pre-trained model [4] is used to conduct experiments for text summarization. The motivation for this work can be divided into the following two parts: In a high-impact event, people may want to learn the latest information as the event develops. It is necessary to develop an effective real-time event summary system to obtain sufficient real-time intelligence information for effective decision-making. In order to acquire extensive coverage of event details, using pre-trained language models can help us expand more knowledge and trending topics about how event sto- ries develop. The benefit of this potential for event awareness deserves in-depth study. The view is taken, therefore, in this work, we utilized the above-mentioned approaches to produce a streamlined version of a dynamically changing event data from the collected Twitter data set; when users view this version, they will be able to immediately capture the most important part of the latest event content in real-time for making appropriate decisions. It is worth mentioning that “real-time event summarization” is different from “real-time event detection” methods. The summarization system we developed can sum- marize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. The summary generated by our system can represent a simplified version of the event tweet message of interest, starting at any point in time of the ongoing event. Appl. Sci. 2021, 11, 10596 3 of 29 The rest of the paper is organized as follows. Section 2 surveys several related work in the literature. The proposed system model and system framework are described in Section 3. Section 4 reports on our pre-trained language model for text summarization. In particular, we introduce our developed text summarization method using a hybrid pre-trained model system. In Section 5, we describe our experiments in evaluation on (i) the performance of the text summarization method based on Rouge evaluation method, and (ii) the effectiveness of the real-time summary method developed for intelligence collection and event monitoring. In particular, in this work, we provide a real-world case study, namely the COVID-19 pandemic event, to verify the applicability of the proposed method. In Section 6, the discussion of experimental results is addressed in detail. Section 7 concludes the work by summarizing our research contributions. 2. Related Work In this work, we attempt to build a real-time event summarization system which can generate a simplified version of a dynamically changing event data from the collected Twitter data set. When users read the generated sentences, they will be able to immediately capture the most important part of the latest event content in real time to make appropriate decisions. Using Twitter for event detection and event summarization is a vast subject. Twitter is also employed to better understand how the communication transfers during citizen movements and events. In our previous work [5,6] a real-time event detection approach has been well developed to gather tweets of detected events for awareness of emerging events in the real world. Summarization of a single text summarization is the task of automatically generating a shorter version of a document and retaining its most important information. The types of automatic text summarization can be classified as extractive and abstractive. The extractive text summarization extracts sentences from the texts, while the abstractive ones may produce sentences and phrases which do not appear in the original documents. In practice, the techniques of abstractive text summarization emulate human summarization in that it utilizes a vocabulary beyond the specified text and summarizes key points in the text. Due to such approaches emulate how people summarize material, they normally require more computing power such as several GPUs to train the model over many days for learning algorithms, thus it is generally more difficult to acquire a cost-effective solution. In this work, we concentrate on abstractive summarization techniques. While text summarization has been investigated for years, automatic summarization on social messages (e.g., tweets) is still in its infancy. Sharifi et al., used the Phrase Reinforcement algorithm to deal with tweet summarization [7]. Inouye developed a Hybrid TF-IDF algorithm to produce multiple post summaries [8]. In order to make our model to summarize text accurately, it needs to learn to “un- derstand” the input data. Google’s BERT (Bidirectional Encoder Representations from Transformers) is the perfect tool for natural language understanding. Before the emergence of the BERT model, several important techniques related to the modeling method we adopted were developed. Recently, the Recurrent Neural Networks (RNNs; including Long Short Term Memory Networks) have been widely used for many natural language processing applications. However, it needs massive amounts of data, costly data sources, and several hours of training to achieve satisfactory results. Due to the drawback of suffering from poor performance with very long sequences of data, Vaswani proposed a model called the “Transformer”, replacing the RNNs and Convolutional Neural Networks (CNNs) by utilizing an architecture comprised of feed forward networks and attention mechanisms [9]. An unsupervised learning architecture called BERT (Bidirectional Encoder Representations from Transformers) has been developed and outperformed almost all existing models in the competitions for various types of NLP tasks [10]. In addition, they published several pre-trained models which can be utilized for transfer learning for many tasks and applications [10]. Appl. Sci. 2021, 11, 10596 4 of 29 Text summarization is a technology that extracts the important text or phrase infor- mation from the original text and then presents a short message to the user. According to the content and method of abstracting, it can be divided into Extractive and Abstractive. Extractive summarization is a technique that uses statistics that extracts the most important words and phrases in a text according to the given weight, and then arranges and combines these features to generate a summary of the text; abstractive summarization is a technique that uses machine learning methods to generate a summary of the text after understanding the content of the text [11–14]. Most of the extractive summarization are generated by extracting human-designed words or phrases as features in the original text [15–17], so that these extracted features are the most useful and important words and phrases in the text to achieve a more accurate final summary [18]. These features are usually rated according to their importance in the title, word frequency, word position, sentence length, similarity between sentences, num- ber of occurrences of words, and famous words [19], and then rearranged and combined according to this rating to generate the summary. The method has been studied in various applications [20–22]. For example, Isonuma [23] proposed a framework for extractive summarization using the dataset NIKKEI financial report corpus and the New York Times annotation corpus. Nallapati [24] proposed a Recurrent Neural Network (RNN) to extract the words and phrases that are relevant to the topic. Nallapati [24] proposed SummaRuN- Ner, a language model based on Recurrent Neural Network (RNN), which has a different training mode from the usual models and can reduce the need for extractive feature labels during training. Gunawan [25] used sentence length, headline features, and keyword frequency to develop an unsupervised summarization model for 3 K articles. Narayan [26] summed up the concept of extracted summaries as a task for ranking sentence and presents a training method that uses reinforcement learning to optimize the reward shaping that related to the task by exploring the space of candidate summaries. Dong [27] presents a framework for extractive summarization through the neural networks and enhanced learn- ing contexts, which did not require sentence-level extraction labels on extracted summaries and optimized the ROUGE score of model that generates the summary. Rush [28] presented a Neural Attention Model for Abstractive Sentence Summariza- tion, combining the chance model with the generation algorithm to produce an accu- rate summary, and acquires good results in the Gigaword and DUC datasets, on which Chopra [29] further improves the performance of the summary after replacing the decoder with an RNN. Song [30] proposed an LSTM-combined CNN abstractive text summarization framework (ATSDL), which extracts key phrases from the text and generates a smooth and grammatical summary by deep learning models. Paulus [31] proposes a neural network model for dealing with longer textual data, and the state-of-the-art is achieved in the experiment using 12 data sets. Chen [32] proposed a sentence-level learning method that allows abstractive summarization models to learn the hierarchical nature of words and sentences. Masum [33] used a bidirectional RNN with an LSTM as the encoding layer and an attention model as the decoding layer to create a short, fluent and comprehensi- ble summary of a text document. Cai [34] proposed an abstractive text summarization model RC-Transformer (RCT), which adds an RNN encoder to capture the context of the sequence and adds a convolutional layer to extract important sentences and understand the semantics to generate a paragraph of extremely high quality summary. In recent years, transfer learning methods using pre-trained models have achieved good results in most natural language processing tasks [35]. The self-supervised pre- training of the model on a large amount of unlabeled data will affect the effect of transfer learning, such as building a language model or filling in the blanks of words. Therefore, fine- tuning a small data set in the pre-trained model will be more effective. Devlin [10] proposed the BERT (Bidirectional Encoder Representations from Transformers) model in 2018, which has had a great impact on the field of natural language processing. BERT provides a pre- trained bidirectional feature for unlabeled text data in the context of all levels. Automatic text summarization is a hot research topic in the field of natural language processing. Appl. Sci. 2021, 11, 10596 5 of 29 Many studies use deep learning technology to obtain important information from text data, so that the machine can understand the text and generate a text-related summary. Therefore, many recent studies have applied pre-training models to text summarization. Liu [36] proposed a framework built using BERT. The framework can provide extractive and abstractive types of text summarization. The experimental results on three data sets (CNN/DailyMail, NYT and XSum) show that their model achieves the state-of-the-art result under the automatic and human-based evaluation protocol. Kieuvongngam [37] used the BERT model and GPT-2 model to conduct text summarization on the open research data set of COVID-19 and extracted the keywords and semantic meaning of the sentences from the original text data through the above-mentioned pre-trained model. Their system provides a concise and meaningful summary for the medical literature. Khandelwal [38] used the pre-trained Transformer language model to generate summaries from the original text data. Compared with the pre-trained Transformer encoder-decoder network, this model has significantly improved results. Zhang [39] proposed a pre-trained model HIBERT (Hierarchical Bidirectional Encoder Representations from Transformers) for pre-training on text data and unlabeled data. This model can achieve higher results on the CNN/Dailymail data set and the New York Times data set. Therefore, the experimental results show that the use of pre-trained models can make large open data sets acquire better results. Farahani [40] mentioned two approaches to deal with the task of text summarization using mT5 model and ParsBERT model and got good results on a data set named pn-summary for Persian abstractive text summarization. Ma [41] proposed a pre-trained model T-BERTSum for text summarization, which captures the key words of the topic information of social media, understands the meaning of the sentence and judges the topic of the message discussion, and then generates a high-quality section Summary. Kerui [42] uses BERT, Seq2seq and reinforcement learning to form a text summary model. Garg [43] uses T5, one of the most advanced pre-trained models, to perform a summary task on a data set with 80,000 news articles, and the results indicate that the summary generated by T5 has better quality than those generated by other models. Daiya [44] has developed a pre-trained language model ENEMAbst that can be used in extractive and abstractive summarization techniques. Both abstractive summaries and extractive summaries on the CNN/DailyMail dataset obtain the most excellent results. Fecht [45] proposed a pre-trained language model that can be used for abstractive text summarization, and combined with the use of BERT to experiment on the German summarization data set. After performing the contextual word embedding task through BERT, the model can improve the model’s ability to generate high-quality German summaries. In Table 1, some important related work is shown for comparison. In summary, in our work, social media (such as Twitter), pre-trained models, and text summaries (including extractive and abstractive methods) are the three key elements of event summarization. From the above literature review, the text summaries of specific event tweets on social media have gradually become one of the most popular summariza- tion research topics in recent years [1–3,41,42]. However, extractive summarization cannot obtain many key sentences in the tweet-form data, resulting in unsatisfactory quality of the generated summaries. Furthermore, due to our task requires extensive domain knowledge beyond the meaning of individual words and sentences, in this work we limit our focus to abstractive summarization algorithms. On the other hand, in the recent references of text summarization, there are more and more work using pre-training language models [35–44]. Among them, BERT is the pre-trained model mainly used by most research teams. Most researchers use the improved BERT model architecture in their work [36,37,39–41]. How- ever, because the sentences in the pre-training data set of BERT have correct grammar, BERT does not perform well for tweets with less grammar. In order to simplify the sum- marization process, the text-to-text pre-trained model (e.g., T5 model) has become a good choice. Unfortunately, for our problem domain, so far little attention has been paid to the development of a summary method that supports real-time functions to achieve the requirements of real-time event monitoring and intelligence collection. Motivated by this, in this work, we developed a novel model combining a real-time event summarization Appl. Sci. 2021, 11, 10596 6 of 29 system and a Twitter-based event detection method. The developed summarization sys- tem can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. To the best of our knowledge, this is the first attempt to develop such an event summarization system with multiple functional requirements to meet the challenge of intelligence gathering in emergency situations. Table 1. Related work of automated text summarization. Using Pre-Trained Method of Text Citing the Author Model Data Set Model Summarization NIKKEI Financial Isonuma et al. [23] CNN-LSTM No Extractive Report CNN/Daily Mail and Nallapati et al. [24] SummaRuNNer No Extractive DUC 2002 Amazon Fine Food Masum et al. [33] Bidirectional LSTM No Abstractive Reviews German Wikipedia Fecht et al. [45] BERT Yes Abstractive articles Kerui et al. [42] BERT Yes Abstractive LCSTS Garg et al. [43] T5 Yes Abstractive and Extractive 80 K News Articles Farahani et al. [40] mT5 Yes Abstractive Pn-summary Cai et al. [34] RC-Transformer Yes Abstractive Gigaword and Conditional DUC 2004 Chopra et al. [29] No Abstractive Generation by RNN Song et al. [30] CNN-LSTM No Abstractive Paulus et al. [31] Attention-RNN No Abstractive Ma et al. [41] T-BERTSum Yes Abstractive CNN/Daily Mail Zhang et al. [39] HIBERT Yes Abstractive Liu et al. [36] BERT Yes Abstractive and Extractive Daiya [44] ENEMABst Yes Abstractive and Extractive 3. Models and Methods 3.1. System Framework In this work, the pre-trained models are mainly used to perform summarization function on social media data to fine-tune the text summaries. With the rapid development of social media, many messages flow and gather on social media, forming various event topics. However, identifying “event” is one of the most difficult challenges of current decision support systems. If the system encounters an emergency, it is usually unable to promptly notify users of the warning to avoid risks. A sensible solution is to employ the functions of the real-time event summarization to obtain emergent event information. As a result, the system framework of this study utilizes the method of event detection on social-media and messages to extract their main content, and then summarized to understand the evolution of important events. Figure 1 illustrates the conceptual diagram of the process followed to extract real-time event intelligence for both the event detection and summarization. The system is designed as a two-pass process, consisted of two modules, say, real-time event detection module and event-message accumulation and summarization module. More details on real-time event summarization are presented in the following. Firstly, in this work, we combine a developed Twitter-based real-time event detection algorithm with pre-trained language models for summarizing emergent events. We used an online text-stream clustering algorithm and self-adaptive method developed to gather the Twitter data for detection of major international events. The unimportant messages on the data set are removed through the trained SBERT based classifier. After that, the DBSCAN algorithm is used to cluster similar topics in each day into hot topics (i.e., event topics). These event topics in Twitter data are generated by using the algorithm we developed to generate storylines. When the hot topics of each day are summarized, the specific topic and other hot topics are semantically correlated. When the correlation exceeds a certain threshold, it is added to the storyline. After daily updates, through the Appl. Sci. 2021, 11, 10596 7 of 29 Twitter based real-time intelligence data collection, a storyline of a specific event is jointly generated. Figure 1. Proposed Twitter based real-time event summarization and pre-trained models. The next work is to deal with text summarization of specific topics. In this work, we use the pre-trained model (i.e., T5; Text-to Text Transfer Transformer) to experiment with the Xsum data set to perform text summarization, and the Rouge metrics are used to evaluate the resulting summaries. The Xsum [46] data set contains 226,711 news articles and their summaries. News articles come from the BBC’s content from 2010 to 2017. The articles cover many fields such as news, politics, sports, weather, business, technology, science, health, family, education, entertainment and art, etc. We split the data set into 90% training data set, 5% validation data set and 5% test data set. The pre-trained language models and pre-training process will be described later in more detail. Subsequently, we started to use the trained model to summarize the incoming Twitter data set for experimentation. The large amount of events and related tweets detected by the real-time event detection system are grouped into about hundreds of event topics every day. Taking the Twitter data in 2020 as an example, there are about 36,000 event topics on Twitter. Using the resulting models of text summarization on popular topics of social media in this work, the system is used to perform real-world event summarization tasks and analyze the results. More details about pre-trained language models for text summarization are presented in the following section. Appl. Sci. 2021, 11, 10596 8 of 29 3.2. Proposed Approach for Collectively Summurizing Text Streams The first part is the real-time event detection algorithm, which analyzes the life cycle of events and the distribution of word weights through dynamic term weighting [47] in the instant message stream. In the early stage of the event, potential major events are captured, and then the real-time event cluster is combined into an event sequence through the self-adaptive stream clustering method. The algorithm for real-time event detection is shown in Algorithm 1. Algorithm 1 1st Pass real-time event detection INPUT: TS{1, 2, . . . , m} // Twitter Streams OUTPUT: RTED{1, 2, . . . , "}; //Real-Time Event Detection 1: Begin 2: for each TS{i} 3: BursT Dynamic Term Weighting(i) 4: RTED Self-Adaptive Stream Clustering(BursT, i) 5: end for 6: End BursT = BS  TI (1) w,t w,t w,t 8 9 < = ar q E ar q w ,t w ,t r r BS q = max   , 0 (2) w ,t : ; E ar w ,t TI q = 1  ln (3) w ,t [t , t] g [t , t] ln M m : w 2 W Dynamic Term Weighting can be divided into two factor values: Burst Score, BS and Term Importance, TI, as shown in Formula (1). In Formula (2), Burst Score is to calculate the burstiness of a keyword according to the rate at which the keyword arrives. In Formula (3), the key word weight is the logarithm of the ratio of the number of documents in the global time window to the number of messages containing the keyword to determine the popularity of the keyword [47]. Self-adaptive stream clustering algorithm [48] is used to represent the situation that, when a new message flows into the system, the message will be compared with the existing online group for similarity. When the similarity comparison result is greater than the threshold, the new message will be added to the most similar group in the online group. If it is less than the threshold, the system will create a new group and add the message. At the same time, the system will remove non-informative messages and aging online groups based on the adjustable sliding window module and the time decay function. The proposed self-adaptive stream clustering algorithm is shown in Algorithm 2. Once the adaptive stream clustering algorithm is used to process stream messages in real time, the number and weight of features will be automatically adjusted within the thread time width tw. As shown in the Algorithm 2, when a new message flows into the system, the message will be compared with the existing online cluster for similarity evaluation. When the similarity comparison result is greater than the threshold, the new message will be added to the most similar cluster. If it is less than the threshold, the system will create a new cluster and add the message. At the same time, the system will remove non-informative messages and aging online clusters based on the adjustable sliding window. Appl. Sci. 2021, 11, 10596 9 of 29 Algorithm 2 Self-adaptive stream clustering INPUT: MS{1, 2, . . . , m} // Message Streams OUTPUT: RTEC{1, 2, . . . , "}; HEC{1, 2, . . . , n} //Real-Time Event Cluster 1: Begin 2: for each MS{i} 3: for each RTEC{j} 4: if(d() – Q < 0) then 5: add RTEC{j} into HEC and revom RTEC{j}; 6: else 7: simOfMS&RTEC sim(MS{i}, RTEC{j})*d(); 8: if(simOfMS&RTEC > maxSimOfMS&RTEC) then 9: maxSimOfMS&RTEC simOfMS&RTEC; 10: idOfMaxC j; 11: else continue; 12: end for 13: if(maxSimOfMS&RTEC < Q) then 14: create a new RTEC for MS{i}; 15: else 16: merge MS{i} to RTEC{idOfMaxC}; 17: end for 18: until no messages is posted The second part of the event-message accumulation and summarization algorithm is to filter the event information through the BERT classifier, and then use the vector of the filtered message obtained by the Sentence-BERT model and put it into the DBSCAN algorithm to cluster the event information into For each sub-event, the summary con- tent is finally obtained through the T5 model. The proposed method for event-message accumulation and summarization algorithm is shown in Algorithm 3. Algorithm 3 2nd Pass event-message accumulation & summarization INPUT: TM{1, 2, . . . , n} // Text Message OUTPUT: HT{1, 2, . . . , "}; HTS{1, 2, . . . , m}; EHT{1, 2, . . . , k} // Hot Topic 1: Begin 2: for each TM{i} 3: Text BERT Classifier Model(TM{i}) 4: Sentence Embedding Sentence-BERT Model(Text) 5: HT DBSCAN(Sentence Embedding) 6: for each HT{j} 7: HTS T5 Text Summarization Model(j) 8: end for 9: end for 10: End 4. Proposed Pre-Trained Language Models for Text Summarization In this study, the pre-trained language model is applied to social media data to fine- tune the text summaries. With the development of the Internet generation, people prefer to express their feelings or the content they are interested in on social media, and many messages flow on social media every day. When there is more and more information on a specific topic, it is easy to form a hot topic. This research will focus on summarizing popular topics when social media content forms as the main core content, and then summarize them to explore the evolution of popular topics. Transfer learning, where the model is first pre-trained on data-rich tasks, and then fine-tuned on downstream tasks, has become a powerful technology in natural language processing (NLP) field. In this work, the T5 [4] pre-trained model is used to conduct experiments and the Rouge metrics are used to evaluate the resulting summaries. In addition, we also use the BERT pre-training model for experiments as a baseline for Appl. Sci. 2021, 11, 10596 10 of 29 performance comparison with the T5 model. The concepts of T5 and BERT pre-trained models are briefly described in the following subsections. 4.1. Bert Model and Sentence-BERT Model The establishment of the BERT model uses other techniques for training context- dependent representation models, including Semi-supervised Sequence Learning, GPT [49], ELMo [50] and ULMFit [51], but these models are unidirectional or shallow Bidirectional, which makes these models only use the words on their left or right to learn context, while BERT uses a deep bidirectional architecture so that the model can learn the words on its left and right. The BERT model also has good results in the task of semantic similarity, but when the BERT model calculates the semantic similarity, it needs to put two sentences into the model at the same time and calculate the information of the two sentences. This approach requires a lot of For calculation and time, if there are N sentences, it needs to calculate (N(N + 1))/2 times, and the similarity retrieval of 10,000 sentences requires about 50 million calculations, and it takes about 50 million times to execute using the BERT model. 65 h, which makes the BERT model unsuitable for semantic similarity retrieval and cluster analysis. Reimers proposed a model Sentence-BERT (SBERT) to improve the BERT architecture, using the twin and triplets network architecture to generate a sentence embedding that can compare semantics with cosine similarity, making it possible to calculate the similarity of 10,000 sentences The time required for the BERT is shortened from 65 h to 5 s, and the accuracy of the similarity calculation using the BERT model can still be maintained. SBERT added a layer of pooling operation behind the output layer of BERT to generate a revised sentence embedding. The maximum value is used as the sentence vector of the entire sentence. In Reimers’ experiment, the Mean-Pooling method is used as the sentence vector to acquire the best results. 4.2. T5 (Text-to-Text Transfer Transformer) Model T5 is a transformer-based model from Google that is trained in an end-to-end manner with text as input and modified text as output. It achieves state-of-the-art results on multiple NLP tasks such as summarization, question answering, machine translation using a text-to-text transformer trained on a large text corpus. The T5 model is composed of the encoder-decoder scheme of transformer model. The encoder and decoder on the T5- Small model are composed of 6-layer Transformer, the T5-Base model is composed of the BERT-base encoder and decoder, and the T5-Large model is composed of the BERT-large encoder and decoder group. T5-3B and T5-11B each increase the size of the model by 4 times. One of the goals of pre-training process is to provide general knowledge for the model to improve its performance on downstream tasks. T5 uses Transformer ’s Encoder- Decoder model during pre-training process. In the transfer learning of natural language processing, previous studies have used large unlabeled data sets for unsupervised learning. T5 attempts to measure the impact of the quality, characteristics and size of these unlabeled data sets, using C4 (Colosal Clean Crawled Corpus) corpus is used as the training data. 4.3. Developed Text Summarization Method Using a Hybrid Pre-Trained Model System Although there are a lot of content from social media platforms, there are often some unimportant daily message content. Therefore, when data is processed, tweets are also filtered through classification models to filter out unimportant daily tweets. We use about 200,000 HuffPost news headline data sets from 2012 to 2018 [52]. This data set contains 40 different categories of news headline categories, plus the unimportant daily tweet content from our collected tweets and mark it as the 41st category ‘OTHERS’. The text data are classified into 41 categories in total. The classification data is divided into three parts, 60% of the data is used to train the model, 20% of the data is used to verify whether the model has learned useful information, and 20% of the data is used to test the accuracy of the model. Since there are about 40,000 to 500,000 non-retweeted tweets in the daily collected Appl. Sci. 2021, 11, 10596 11 of 29 tweets, which contain a large number of daily trivial tweets, such as happy birthday, and holiday wishes, in this work the messages are filtered to avoid too much unnecessary data and reduce the calculation time. In order to alleviate this problem, we decided to use the characteristics of the BERT model as a classier to perform the function of tweet filtering. Subsequently, the filtered data is clustered through an unsupervised learning algorithm, and the content of the tweet is retained by the density of the content of the tweet, so as to ensure that the tweet is a hot topic on the day. As a result, in this system, the tweets are aggregated into a daily form and screened and filtered the content of tweets that were not reposted in English, and filtered the tweets through the BERT classification method, using the HuffPost data set plus the tweet content from the tweets. The nonessential daily tweets are filtered out and divided into 41 categories in total. The filtered tweets are processed through the SBERT (Sentence-BERT) model to obtain sentence vectors, and the Euclidean distance between the vectors is used to calculate the similarity between sentences. Here we use the DBSCAN algorithm to find more intensive discussion content, and finally form a daily hot topic. This study will learn text summaries by fine-tuning the pre-trained model in the Xsum data sets. In the experiment, the T5 pre-trained model is firstly used for text summarization, as shown in Figure 2. Subsequently, we compare the T5 model with the BERT pre-trained model for summarization performance comparison. Figure 2. Developed text summarization method using a hybrid pre-trained model system. 4.4. Evaluation Method for Text Summarization This research work mainly uses Rouge (Recall-Oriented Understudy for Gisting Evalu- ation) [53] as the evaluation method of text summarization. It uses automatically generated summaries and a set of reference summaries to do calculations to measure the similarity between automatically generated summaries and reference summaries. Rouge-N is used to calculate the number of n-grams shared by the generated summary and the reference summary, as shown in Formula (4). Count (gram ) å å match S2fRe f eremceSummariesg gram 2S n Rouge N = (4) Count(gram ) å å S2fRe f eremceSummariesg gram 2S Rouge-L is used to match the longest common subsequence (Longest Common Subse- quence, LCS) between two texts. 5. Experiment and Case Study We designed two sets of experiments to evaluate (i) the performance of the text summarization method based on Rouge evaluation method, and (ii) the effectiveness of the real-time summary method developed for intelligence collection and event monitoring, through the implementation of a case study in the real world. Appl. Sci. 2021, 11, 10596 12 of 29 5.1. Experiment and Performance Comparison We first performed experiments on the Xsum dataset on the BERT and T5 pre-trained models. After the evaluation of the text summaries, it can be found that the performance and output length of different pre-training models used in Tables 2–4 are different. Com- pared with BERT, T5 is more excellent in the performance of text summarization. The training time spent is much less than that of BERT, so the T5-base pre-trained model and the output length of 150 parameters are finally used in the text summary tool. Table 5 shows the result with the complete training data and most suitable parameters in the experiment. In order to save the training cost of the pre-trained model, the training data is compressed to find the most suitable parameters. Table 6 illustrates a comparison with other models. Table 2. Performance evaluation (output length: 80). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 28.17 8.08 21.76 T5-base 33.01 11.46 25.76 BERT 23.50 6.51 17.62 Table 3. Performance evaluation (output length: 100). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 28.21 8.37 21.66 T5-base 33.64 11.80 26.34 BERT 23.42 6.56 17.43 Table 4. Performance evaluation (output length: 150). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 28.97 8.26 22.07 T5-base 33.70 12.06 26.47 BERT 23.66 6.82 17.75 Table 5. Performance evaluation (using complete training data). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 32.64 11.37 25.66 T5-base 38.13 16.39 30.88 BERT 29.58 11.14 22.22 Table 6. Performance evaluation (model comparison). Model Dataset Rouge Score T5-base Xsum 38.13 UniLm CNN/DailyMail 43.08 BERT CNN/DailyMail 44.16 5.2. Xsum and Tweet Data Sets for Fine-Turning and Downstream Task In this study, we use the Xsum data set and a pre-trained language model, the T5 model, to train the summary model. Since the downstream task is to generate a summary of the tweet data set, the features required for the summary generated in the proposed system are mainly learned by fine-tuning the pre-trained T5 model in the Xsum data set. The Xsum dataset contains 226,711 BBC news articles and their summaries. Articles cover news, politics, sports, weather, business, technology, science, health, family, education, entertainment, art and other fields. The coverage and diversity of Xsum are very suitable for our downstream task, which is to summarize a large number of tweets. Therefore, Appl. Sci. 2021, 11, 10596 13 of 29 after the fine-tuning process, the well-trained model is used in massive tweet data sets for summarization. 5.3. Detected Hot Topics with Twitter Data In the hot topic detection system, a large amount of social media content is integrated into about hundreds of hot topics every day. Taking the Twitter data in 2020 as an example, there are about 36,000 hot topics in total, as shown in Figure 3. Based on the method of text summarization on popular topics in social media, the technique is applied to the task of text summarization in various fields on social media and specific data analysis. Figure 3. Daily detected event hot topics in 2020. Appl. Sci. 2021, 11, 10596 14 of 29 5.4. Case Study:Tweet Summary of the “COVID-19 Pandemic” Event In 2020, COVID-19 is undoubtedly one of the hottest topics. The COVID-19 pandemic has exposed many serious errors in global public policy preparation and response. Al- though it has been affected by other epidemics and pandemics for more than a decade, many countries have failed to incorporate lessons learned into their pandemic preparedness and response plans. One of the main factors is that we lack effective real-time event-related intelligence collection systems which can provide a summary of updates related to the COVID-19 pandemic in all countries. Hence, we cannot learn from the successful experi- ence of other countries in fighting the epidemic in time. Since the COVID-19 outbreak, very few studies that respond to practical needs, such as through rapid advice guidelines or application modules, have been conducted. However, many professionals, such as health workers, are in urgent need of effective ways to help people improve their well-being during this pandemic. According to the US Centers for Disease Control and Prevention (CDC), information on the outbreak of the coronavirus disease may increase the levels of distress among specific population—especially in those who were already struggling with mental health issues before the pandemic. They may be likely to experience higher levels of worry and stress regarding issues surrounding COVID-19. This trepidation may worsen over time and cause a long-lasting mental disturbance [54]. We wish that the proposed system can help the professionals detect and track COVID-19 events at an early stage as well as monitoring the emergency events within specific groups. The view is taken, therefore, in this case study, we combine our developed Twitter- based real-time event detection algorithm with pre-trained language models for summariz- ing COVID-19 events and sub-events. In Figure 4, it can be found that there are many detected hot-event topics around COVID-19 in 2020. In order to satisfy users’ information needs, we pick up all of the real-time event topics with a comprehensive spatio-temporal viewpoint. By analyzing the events detected by our system, the spatial and temporal impacts of the emerging events can be estimated. Therefore, we first build a representation of the spatial distribution by mapping the summary of events related to COVID-19 in the world in 2020, as shown in Figure 5. After that, in terms of temporal aspects, we constructed a storyline retrospection mechanism that demonstrates the story telling capability for users to understand the context of event development in a timeline, as illustrated in Figures 6 and A1–A3. Figure 4. Detected event topics related to COVID-19 in 2020. Appl. Sci. 2021, 11, 10596 15 of 29 Figure 5. Representation of a map of generated summaries related to COVID-19 events in 2020. In Figure 5, we present hot topics in the form of a map to illustrate global events to understand the spatial impact of the COVID-19 pandemic on the world. Furthermore, in this work, we developed a storyline retrospection mechanism that demonstrates the story telling capability for users to understand the context of event development in an efficient way. It can be used to identify events and their relations under an event topic and then composes a storyline which gives users the sketch of event evolution in the topic, by computing daily related tweets and summaries of formulated event topics. The main storyline construction removes irrelevant events and presents a main theme that exhibits the sequence of events in a hot topic. In Figures 6 and A1–A3, we can see the storyline of COVID-19 in 2020, from the beginning of the outbreak to the emergence of the final vaccine. Figure 7 shows the example resulting summaries of COVID-19 event related tweets in 2020, starting with a new virus related to the SARS outbreak that may cause Chinese health officials to discuss on social media Pneumonia. From the time trend shown in Figures A1–A3, when the COVID-19 pandemic first broke out, social media were mainly discussing where the pandemic occurred and how to properly protect themselves from the pandemic. After a period of time, the social media began to gradually appear on the topic of what drugs may be able to treat the symptoms of COVID-19 or the topic of vaccine-related research and development. As time progressed, social media began to see related topics such as whether the vaccine was safe. Appl. Sci. 2021, 11, 10596 16 of 29 Figure 6. Generated storyline related to COVID-19 from January to March 2020. Appl. Sci. 2021, 11, 10596 17 of 29 Figure 7. A sample resulting summary of formulated event topics (9 January 2020). 5.5. Human Evaluation for Quality Assessment of Generated Summaries In order to verify the quality of the text summary results, we conducted a survey on the example resulting summaries (in Figures 7 and A4–A11) with human judges for quality assessment of generated summaries. In this work, we invited 39 participants from University graduate students, including 19 males and 20 females. Their age ranges from 20 to 30 years old. Participants were asked to rate the summaries on a 1–5 scale, with higher being better. The assessed summaries were selected using random sampling. Thus, we randomly selected 10 sample summaries from our summarization system for evaluation. In this work, the summaries are evaluated from three aspects. The first aspect is to assess the consistency of the topic between the tweets and the summary. The second aspect is to evaluate the closeness to the “facts” of the summary. The third aspect is to assess the degree of grammatical correctness of the summary. The summaries are reviewed on a scale of 1 to 5 points, and the evaluation results are shown in Table 7. Options for grading summaries are as follows: 1: Bad 2: Below Average 3: Average 4: Good 5: Excellent. In Table 7, we find that the mean values (i.e., 3.92, 3.89, 3.92, respectively) are close to 4 points, which show that the experimental result is acceptable and very close to the performance of human. In addition, we used Cronbach’s alpha test to check whether the multi-question Likert scale survey is reliable and acquires results (Cronbach’s alpha coefficients are 0.725, 0.812, 0.846). It shows that Cronbach’s alpha coefficient is greater than 0.7, and the resulting coefficient indicates that the results are reliable. Table 7. Quality assessment of generated summaries. Consistency of the Closeness to the Grammatical Topic Facts Correctness No. of Participant 39 39 39 Minimum 2 2 1 Maximum 5 5 5 Mean 3.92 3.89 3.92 Median 4 4 4 Std. Deviation 0.824 0.819 0.928 Std. Error 0.132 0.131 0.148 Cronbach’s alpha 0.725 0.812 0.846 Appl. Sci. 2021, 11, 10596 18 of 29 6. Discussion In this work we use natural language processing techniques to deal with social media data, extract event topics from daily social messages, and summarize them to quickly grasp various event intelligence information. In the experiment, we adopted a pre-trained model (i.e., T5) method to quickly fine-tune the model to learn text summaries through the Xsum data set, which was achieved a score of 38.13 in the evaluation of Rouge-1. After that, the trained model is actually applied to real social media data. We started to use the trained model to summarize the incoming Twitter data set for experimentation. The model summarizes the collected hot topics of social media into a short sentence. The resulting summary can help users quickly understand topic trends on social media, most of which are in line with the development of real-world events. Furthermore, in order to verify the quality of the text summary results, we conducted a survey on the resulting summaries with human judges for quality assessment of generated summaries. The experimental result is discussed in detail as follows: (1) The T5 model is trained in an end-to-end and text-to-text manner, with text as input and summary text as output. In this work, we demonstrated that T5 has achieved state-of-the-art results in many NLP benchmarks and can be fine-tuned to perform automatic summarization tasks well. (2) As mentioned earlier, due to the real-time nature of social media, an automatic event summary system using Twitter data was developed in this work to provide people with an updated summary of the event while the situation is still evolving. However, it is worth mentioning that the “real-time” here means that the summary system can expand or update the summary at any point in time, rather than real-time computing. (3) Going further, “real-time event summarization” is different from “real-time event detection” methods. The summarization system we developed can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. Compared to real-time event detection methods, detected events on Twitter are often short-term. The summary generated by our system can represent a simplified version of the event tweet message of interest, starting at any point in time of the ongoing event. (4) Generally speaking, abstractive summarization models are prone to produce gram- matically correct incoherent sentences. In the work of summarizing text, it is easy for human researchers to make language errors. The advantages of using abstractive summarization models enable us to alleviate some of the grammatical problems of real people. Despite the results of the system evaluation show that the summaries we got are not perfect in terms of grammatical fluency, the results are acceptable and close to the performance of human researchers in the real world. (5) Although this developed model can quickly summarize the text, it still has certain limitations in generality. The social media data used in this study only uses English data on Twitter to generate summaries. Although English is the common language of most people in the world, it must be only through English to understand the various events in the world, there are still some prejudices. In the future, we consider using multilingual models and multilingual data sets (made by different native speakers) to train the model so that we can more accurately grasp various world events from different perspectives. (6) In this work, we implemented a method of generating storylines using the automatic summarization technique we developed. The generated storyline is composed of sub-events or key events related to a given major event. The sub-events show the status of the main event in progress. Automated storyline generation has been a research issue since nearly the inception of artificial intelligence. In particular, the storylines of events relate to identify entities and summarize events that leads to the event of interest. However, the proposed system model has not yet performed the full function of the “automated storyline generation” paradigm for understanding event development. This is because in this work, we have not provided methods for Appl. Sci. 2021, 11, 10596 19 of 29 analyzing the evolution of specific event entities and exploring the entities and their relationships in event stories. We will leave the work of constructing event storyline for future work to complete. 7. Conclusions The uncertainty of emergencies and natural disasters may affect companies and everyone’s lives. During such high-impact events, people may want to stay up-to-date as the event develops because they want to better understand the situation and possible trend of the event. In this case, it is almost impossible for anyone to browse through all of the news articles to understand the event stories required for decision-making. Automated summarization systems that access a large number of information sources in real time can help by providing event summaries, which are updated whenever important sub-events are detected. The update of the event will help those interested persons to understand the situation of the current event. In this work, we developed a hybrid method to combine a real-time event summariza- tion system and a self-adaptive Twitter-based event detection method, using pre-trained language models on the tweet data set for online machine learning. Hence, each user can obtain real-time information about specific incidents (or events) and the agility to make appropriate decisions quickly. The goal of this work can be divided into the following two parts: (1). In a high-impact event, people may want to learn the latest information as the event develops. It is necessary to develop an effective real-time event summary system to obtain sufficient real-time intelligence information for effective decision-making; (2). In order to acquire extensive coverage of event details, using pre-trained language models can help us expand more knowledge and trending topics about how event stories develop. Our research is intended as a starting point for exploring the benefits of this potential for event awareness. The novelty and contributions of the work are summarized as follows: We implemented an end-to-end and text-to-text summarization system model, in which text streams as input and summary text as output. Our system model combined a real-time event summarization system and a self- adaptive Twitter-based event detection method. The developed summarization system can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. The developed system used pre-trained language models to help users obtain more knowledge and trending topics about how event stories develop. The system can support users to obtain sufficient real-time intelligence information in specific events to make effective decisions. For future work, it will be interesting to further incorporate the sentiment of the tweet into the summary of related events to predict the world market trends affected by the event. On the other hand, as mentioned earlier, building a summary system in the “automatic storyline generation” paradigm is very useful for understanding the details of event development. In this way, it is possible to explore the evolution of a specific event entity in the event story and the relationship between the entity and its event. In addition, by taking advantage of the potential of pre-trained language models, it will be very useful to gather intelligence on multilingual event summaries from multilingual social media messages. Finally, the use of automatic video summarization methods to extend the developed real-time text summarization methods to generate event summaries is also a promising application. Appl. Sci. 2021, 11, 10596 20 of 29 Author Contributions: Conceptualization, C.-H.L. and Y.-L.C.; methodology, C.-H.L.; software, Y.- L.C.; validation, H.-C.Y., Y.J.C. and Y.-L.C.; formal analysis, H.-C.Y.; investigation, C.-H.L.; resources, Y.J.C.; data curation, H.-C.Y.; writing—original draft preparation, C.-H.L.; writing—review and editing, C.-H.L.; visualization, Y.-L.C.; supervision, C.-H.L.; project administration, C.-H.L.; funding acquisition, C.-H.L. All authors have read and agreed to the published version of the manuscript. Funding: This research received no external funding. Institutional Review Board Statement: Not applicable. Informed Consent Statement: Not applicable. Data Availability Statement: The study did not report any data. Conflicts of Interest: The authors declare no conflict of interest. Appendix A. Examples of Generated Storyline Related to COVID-19 Pandemic Figure A1. Generated storyline related to COVID-19 from April to June 2020. Appl. Sci. 2021, 11, 10596 21 of 29 Figure A2. Generated storyline related to COVID-19 from July to September 2020. Appl. Sci. 2021, 11, 10596 22 of 29 Figure A3. Generated storyline related to COVID-19 from October to December 2020. Appl. Sci. 2021, 11, 10596 23 of 29 Appendix B. Example Model Outputs: Tweet Summaries of the “COVID-19 Pandemic” Sub-Events Figure A4 shows that on 24 January 2020, the World Health Organization issued a global warning about the new coronavirus that was discussed in social media. Figure A4. A sample resulting summary of formulated event topics (24 January 2020). Figure A5 shows that on 12 February 2020, the World Health Organization officially named the new coronavirus COVID and was discussed in social media. Figure A5. A sample resulting summary of formulated event topics (12 February 2020). Appl. Sci. 2021, 11, 10596 24 of 29 Figure A6 illustrates the topics that will be postponed for the 2020 Olympic Games discussed on 30 March 2020. Figure A6. A sample resulting summary of formulated event topics (30 March 2020). Figure A7 shows that the possible variants of the new coronavirus were discussed in social media on 6 May 2020, and they may become more toxic and more contagious than the original. Figure A7. A sample resulting summary of formulated event topics (6 May 2020). Appl. Sci. 2021, 11, 10596 25 of 29 Figure A8 shows that the Dexamethasone drug was discussed in social media on 17 June 2020 to help save the lives of patients with the new coronavirus. Figure A8. A sample resulting summary of formulated event topics (17 June 2020). Figure A9 shows that on 15 July 2020, in social media, the Center for Disease and Control of the United States called on people to wear masks to effectively prevent the new coronavirus. Figure A9. A sample resulting summary of formulated event topics (15 July 2020). Appl. Sci. 2021, 11, 10596 26 of 29 Figure A10 shows that it was discussed in social media on 16 November 2020. American drug maker Moderna stated that its COVID vaccine can effectively prevent the coronavirus. Figure A10. A sample resulting summary of formulated event topics (16 November 2020). Figure A11 shows that Pfizer’s COVID vaccine was discussed in social media on 9 December 2020, in compliance with the safety and effectiveness guidelines for emergency use. Figure A11. A sample resulting summary of formulated event topics (9 December 2020). Appl. Sci. 2021, 11, 10596 27 of 29 References 1. Rudrapal, D.; Das, A.; Bhattacharya, B. A survey on automatic Twitter event summarization. J. Inf. Process. Syst. 2018, 14, 79–100. 2. Chakrabarti, D.; Punera, K. Event summarization using tweets. In Proceedings of the International AAAI Conference on Web and Social Media, Barcelona, Spain, 17–21 July 2011; Volume 5. 3. Chua, F.C.T.; Asur, S. Automatic summarization of events from social media. In Proceedings of the Seventh International AAAI Conference on Weblogs and Social Media, Cambridge, MA, USA, 8–11 July 2013. 4. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv 2019, arXiv:1910.10683. 5. Lee, C.-H. Mining spatio-temporal information on microblogging streams using a density-based online clustering method. Expert Syst. Appl. 2012, 39, 9623–9641. [CrossRef] 6. Lee, C.-H.; Chien, T.-F. Leveraging microblogging big data with a modified density-based clustering approach for event awareness and topic ranking. J. Inf. Sci. 2013, 39, 523–543. [CrossRef] 7. Sharifi, B.; Hutton, M.A.; Kalita, J. Summarizing microblogs automatically. In Proceedings of the 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Los Angeles, CA, USA, 2–4 June 2010; pp. 685–688. 8. Inouye, D.; Kalita, J.K. Comparing Twitter Summarization Algorithms for Multiple Post Summaries. In Proceedings of the 2011 IEEE Third Int’l Conference on Privacy, Security, Risk and Trust and 2011 IEEE Third International Conference on Social Computing, Boston, MA, USA, 9–11 October 2011; pp. 298–306. 9. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. 10. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. 11. Rachabathuni, P.K. A survey on abstractive summarization techniques. In Proceedings of the 2017 International Conference on Inventive Computing and Informatics (ICICI), Coimbatore, India, 23–24 November 2017; pp. 762–765. 12. Gambhir, M.; Gupta, V. Recent automatic text summarization techniques: A survey. Artif. Intell. Rev. 2017, 47, 1–66. [CrossRef] 13. Neto, J.L.; Freitas, A.A.; Kaestner, C.A.A. Automatic Text Summarization Using a Machine Learning Approach. In Proceedings of the 16th Brazilian Symposium on Artificial Intelligence, Porto de Galinhas/Recife, Brazil, 11–14 November 2002; pp. 205–215. 14. Tas, O.; Kiyani, F. A survey automatic text summarization. J. PressAcademia Procedia 2017, 5, 205–213. [CrossRef] 15. Radev, D.R.; Allison, T.; Blair-Goldensohn, S.; Blitzer, J.; Celebi, A.; Dimitrov, S.; Drabeck, E.; Hakim, A.; Lam, W.; Liu, D.; et al. MEAD-a platform for multidocument multilingual text summarization. In Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC’04), Lisbon, Portugal, 26–28 May 2004; pp. 699–702. 16. Nenkova, A.; Vanderwende, L.; McKeown, K. A compositional context sensitive multi-document summarizer: Exploring the factors that influence summarization. In Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle, WA, USA, 6–11 August 2006; pp. 573–580. 17. Filatova, E.; Hatzivassiloglou, V. Event-based extractive summarization. In Proceedings of the ACL-04 Workshop, Association for Computational Linguistics, Barcelona, Spain, 21–26 July 2004; pp. 104–111. 18. Moratanch, N.; Chitrakala, S. A survey on extractive text summarization. In Proceedings of the 2017 International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 10–11 January 2017; pp. 1–6. 19. Mutlu, B.; Sezer, E.A.; Akcayol, M.A. Multi-document extractive text summarization: A comparative assessment on features. J. Knowl.-Based Syst. 2019, 183, 104848. [CrossRef] 20. Kupiec, J.; Pedersen, J.; Chen, F. A trainable document summarizer. In Proceedings of the 18th Annual International ACM SIGIR conference on Research and Development in Information Retrieval–SIGIR ’95, Seattle, WA, USA, 9–13 July 1995; pp. 68–73. 21. Conroy, J.M.; O’Leary, D.P. Text summarization via hidden Markov models. In Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval—SIGIR ’01, New Orleans, LA, USA, 9–13 September 2001; pp. 406–407. 22. Woodsend, K.; Lapata, M. Automatic generation of story highlights. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 11–16 July 2010; pp. 565–574. 23. Isonuma, M.; Fujino, T.; Mori, J.; Matsuo, Y.; Sakata, I. Extractive Summarization Using Multi-Task Learning with Document Classification. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2101–2110. 24. Nallapati, R.; Zhai, F.; Zhou, B. Summarunner: A recurrent neural network based sequence model for extractive summarization of documents. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 25. Gunawan, D.; Pasaribu, A.; Rahmat, R.F.; Budiarto, R. Automatic Text Summarization for Indonesian Language Using TextTeaser. In IOP Conference Series: Materials Science and Engineering; IOP Publishing Ltd.: Semarang, Indonesia, 2016; Volume 190, p. 012048. 26. Narayan, S.; Cohen, S.B.; Lapata, M. Ranking Sentences for Extractive Summarization with Reinforcement Learning. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; pp. 1747–1759. Appl. Sci. 2021, 11, 10596 28 of 29 27. Dong, Y.; Shen, Y.; Crawford, E.; Van Hoof, H.; Cheung, J.C.K. BanditSum: Extractive Summarization as a Contextual Bandit. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 3739–3748. 28. Rush, A.M.; Chopra, S.; Weston, J. A Neural Attention Model for Abstractive Sentence Summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 379–389. 29. Chopra, S.; Auli, M.; Rush, A.M. Abstractive Sentence Summarization with Attentive Recurrent Neural Networks. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 93–98. 30. Song, S.; Huang, H.; Ruan, T. Abstractive text summarization using LSTM-CNN based deep learning. J. Multimed. Tools Appl. 2018, 78, 857–875. [CrossRef] 31. Paulus, R.; Xiong, C.; Socher, R. A deep reinforced model for abstractive summarization. arXiv 2017, arXiv:1705.04304. 32. Chen, Y.-C.; Bansal, M. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 675–686. 33. Masum, A.K.M.; Abujar, S.; Talukder, A.I.; Rabby, A.S.A.; Hossain, S.A. Abstractive method of text summarization with sequence to sequence RNNs. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; pp. 1–5. 34. Cai, T.; Shen, M.; Peng, H.; Jiang, L.; Dai, Q. Improving Transformer with Sequential Context Representations for Abstractive Text Summarization. In Proceedings of the CCF International Conference on Natural Language Processing and Chinese Computing, Dunhuang, China, 9–14 October 2019; pp. 512–524. 35. Qiu, X.; Sun, T.; Xu, Y.; Shao, Y.; Dai, N.; Huang, X. Pre-trained models for natural language processing: A survey. J. Sci. China Technol. Sci. 2020, 63, 1872–1897. [CrossRef] 36. Liu, Y.; Lapata, M. Text Summarization with Pretrained Encoders. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 3721–3731. 37. Kieuvongngam, V.; Tan, B.; Niu, Y. Automatic text summarization of Covid-19 medical research articles using bert and gpt-2. arXiv 2020, arXiv:2006.01997. 38. Khandelwal, U.; Clark, K.; Jurafsky, D.; Kaiser, L. Sample efficient text summarization using a single pre-trained transformer. arXiv 2019, arXiv:1905.08836. 39. Zhang, X.; Wei, F.; Zhou, M. HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 5059–5069. 40. Farahani, M.; Gharachorloo, M.; Manthouri, M. Leveraging ParsBERT and Pretrained mT5 for Persian Abstractive Text Summa- rization. In Proceedings of the 2021 26th International Computer Conference, Computer Society of Iran (CSICC), Tehran, Iran, 3–4 March 2021; pp. 1–6. 41. Ma, T.; Pan, Q.; Rong, H.; Qian, Y.; Tian, Y.; Al-Nabhan, N. T-BERTSum: Topic-Aware Text Summarization Based on BERT. J. IEEE Trans. Comput. Social Syst. 2021, 1–12. [CrossRef] 42. Kerui, Z.; Haichao, H.; Yuxia, L.; Zhang Kerui Lazada Group, China; Hu Haichao SF Technology, Shenzhen China; Liu Yuxia SF Technology, Shenzhen China. Automatic Text Summarization on Social Media. In Proceedings of the 2020 4th International Symposium on Computer Science and Intelligent Control, Newcastle upon Tyne, UK, 17–19 November 2020; pp. 1–5. 43. Garg, A.; Adusumilli, S.; Yenneti, S.; Badal, T.; Garg, D.; Pandey, V.; Nigam, A.; Gupta, Y.K.; Mittal, G.; Agarwal, R. NEWS Article Summarization with Pretrained Transformer. In Communications in Computer and Information Science, Proceedings of the Advanced Computing, IACC 2020, Panaji, Goa, India, 5–6 December 2020; Springer: Singapore, 2020; pp. 203–211. [CrossRef] 44. Daiya, D. Combining Temporal Event Relations and Pre-Trained Language Models for Text Summarization. In Proceedings of the 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 14–17 December 2020; pp. 641–646. 45. Fecht, P.; Blank, S.; Zorn, H.P. Sequential Transfer Learning in NLP for German Text Summarization. In Proceedings of the 4th edition of the Swiss Text Analytics Conference, Winterthur, Switzerland, 18–19 June 2019. 46. Narayan, S.; Cohen, S.B.; Lapata, M. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 1797–1807. 47. Lee, C.-H.; Wu, C.-H.; Chien, T.-F. BursT: A Dynamic Term Weighting Scheme for Mining Microblogging Messages. J. Int. Symp. Neural Netw. 2011, 6677, 548–557. [CrossRef] 48. Lee, C.-H.; Wu, C.-H. A Self-adaptive Clustering Scheme with a Time-Decay Function for Microblogging Text Mining. J. Inf. Technol. 2011, 185, 62–71. [CrossRef] 49. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Proceedings of the 27th Interna-tional Conference on Neural Information Processing Systems, Montréal, QC, Canada, 6–12 December 2014; pp. 3104–3112. Appl. Sci. 2021, 11, 10596 29 of 29 50. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; Volume 1, pp. 2227–2237. 51. Howard, J.; Ruder, S. Universal Language Model Fine-tuning for Text Classification. arXiv 2018, arXiv:1801.06146. 52. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv 2019, arXiv:1908.10084. 53. Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Workshop on Text Summarization Branches Out (WAS 2004), Barcelona, Spain, 21–26 July 2004; pp. 74–81. 54. Shigemura, J.; Ursano, R.J.; Morganstein, J.C.; Kurosawa, M.; Benedek, D.M. Public responses to the novel 2019 coronavirus (2019-nCoV) in Japan: Mental health consequences and target populations. J. Psychiatry Clin. Neurosci. 2020, 74, 281–282. [CrossRef] [PubMed] http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Applied Sciences Multidisciplinary Digital Publishing Institute

Event Monitoring and Intelligence Gathering Using Twitter Based Real-Time Event Summarization and Pre-Trained Model Techniques

Loading next page...
 
/lp/multidisciplinary-digital-publishing-institute/event-monitoring-and-intelligence-gathering-using-twitter-based-real-jmgRAdkO0G

References (54)

Publisher
Multidisciplinary Digital Publishing Institute
Copyright
© 1996-2021 MDPI (Basel, Switzerland) unless otherwise stated Disclaimer The statements, opinions and data contained in the journals are solely those of the individual authors and contributors and not of the publisher and the editor(s). MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. Terms and Conditions Privacy Policy
ISSN
2076-3417
DOI
10.3390/app112210596
Publisher site
See Article on Publisher Site

Abstract

applied sciences Article Event Monitoring and Intelligence Gathering Using Twitter Based Real-Time Event Summarization and Pre-Trained Model Techniques 1 , 2 3 1 Chung-Hong Lee *, Hsin-Chang Yang , Yenming J. Chen and Yung-Lin Chuang Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung 807618, Taiwan; assh70057@gmail.com Department of Information Management, National University of Kaohsiung, Kaohsiung 811726, Taiwan; yanghc@nuk.edu.tw Management School, National Kaohsiung University of Science and Technology, Kaohsiung 824005, Taiwan; yjjchen@nkust.edu.tw * Correspondence: leechung@mail.ee.nkust.edu.tw Abstract: Recently, an emerging application field through Twitter messages and algorithmic com- putation to detect real-time world events has become a new paradigm in the field of data science applications. During a high-impact event, people may want to know the latest information about the development of the event because they want to better understand the situation and possible trends of the event for making decisions. However, often in emergencies, the government or enterprises are usually unable to notify people in time for early warning and avoiding risks. A sensible solution is to integrate real-time event monitoring and intelligence gathering functions into their decision support system. Such a system can provide real-time event summaries, which are updated whenever Citation: Lee, C.-H.; Yang, H.-C.; important new events are detected. Therefore, in this work, we combine a developed Twitter-based Chen, Y.J.; Chuang, Y.-L. Event real-time event detection algorithm with pre-trained language models for summarizing emergent Monitoring and Intelligence events. We used an online text-stream clustering algorithm and self-adaptive method developed Gathering Using Twitter Based to gather the Twitter data for detection of emerging events. Subsequently we used the Xsum data Real-Time Event Summarization and set with a pre-trained language model, namely T5 model, to train the summarization model. The Pre-Trained Model Techniques. Appl. Rouge metrics were used to compare the summary performance of various models. Subsequently, we Sci. 2021, 11, 10596. https://doi.org/ started to use the trained model to summarize the incoming Twitter data set for experimentation. In 10.3390/app112210596 particular, in this work, we provide a real-world case study, namely the COVID-19 pandemic event, to verify the applicability of the proposed method. Finally, we conducted a survey on the example Academic Editor: Andrea Prati resulting summaries with human judges for quality assessment of generated summaries. From the case study and experimental results, we have demonstrated that our summarization method provides Received: 25 September 2021 users with a feasible method to quickly understand the updates in the specific event intelligence Accepted: 5 November 2021 Published: 11 November 2021 based on the real-time summary of the event story. Publisher’s Note: MDPI stays neutral Keywords: text summarization; event detection; machine learning; natural language processing; Twitter with regard to jurisdictional claims in published maps and institutional affil- iations. 1. Introduction In emergencies with high market uncertainty and business interruption, real-time decision-making ability is usually a necessary condition to avoid risks. For example, the Copyright: © 2021 by the authors. outbreak of COVID-19, earthquakes or floods have brought unexpected challenges to Licensee MDPI, Basel, Switzerland. companies, making people realize that they should have real-time event information and This article is an open access article resilience to immediately respond to changing external disasters. Furthermore, during distributed under the terms and such high-impact events, users may want to learn up-to-date information as the event conditions of the Creative Commons develops because they want to know the possible trend of the event. In this case, it is Attribution (CC BY) license (https:// almost impossible for anyone to browse through all of the news articles to understand the creativecommons.org/licenses/by/ event stories required for decision-making. Automated summarization systems that access 4.0/). Appl. Sci. 2021, 11, 10596. https://doi.org/10.3390/app112210596 https://www.mdpi.com/journal/applsci Appl. Sci. 2021, 11, 10596 2 of 29 a large number of information sources in real time can help by providing event summaries, which are updated whenever important new events are detected. The update of the event will help those interested persons to understand the situation of the current event. With the real-time nature of social media, an automated event summary system using social media (such as Twitter) data can provide people with event updates while the situation is still evolving, without the need for individuals to manually analyze a large number of news articles. The term “real-time” here means that the summary system can expand or update the summary at any point in time, and the real-time computing property is not the focus of this work. In order to meet the above-mentioned increasing demand, in this work, we provide a solution for real-time event summarization. It combines the developed Twitter-based real-time event detection algorithm with a pre-trained language model for summarizing emergency events. Automatically summarizing social message topics is a relatively unnoticed emerging research field. Essentially, tweet summaries can be viewed as an example of a more general automatic text summarization problem, which is an issue of automatically generating a condensed version of the most important content from user-generated text content. In addition, it is also a type of stream summarization. For text streams containing millions of documents covering various topics and events, traditional single-document and multi- document summarization methods cannot solve the challenge of information overload. Rudrapal [1] summarized the methods of automatic summarization on Twitter ’s most pop- ular topics. For some highly structured and recurring events in social media messages, it is best to use more complex techniques to summarize relevant information. Chakrabarti [2] formalized the summary of event messages and proposed a hidden Markov model. To learn the solution of the non-recessive state representation of the event. Chua [3] proposed a framework for searching and summarizing on social media, extracting the most represen- tative messages from time-series social media messages to generate continuous and concise summaries. Furthermore, the text summarization algorithms can be generally divided into two types of methods, including extractive summarization algorithms and abstractive sum- marization algorithms. Since our task requires extensive domain knowledge beyond the meaning of individual words and sentences, in this work, we limit our focus to abstractive summarization algorithms. In order to generate more semantically rich sentence representations, we apply the concept of transfer learning to perform text summarization with a pre-trained language model. Transfer learning, where the model is first pre-trained on data-rich tasks, and then fine-tuned on downstream tasks, has become a powerful technology in natural language processing (NLP) field. In this work, the T5 pre-trained model [4] is used to conduct experiments for text summarization. The motivation for this work can be divided into the following two parts: In a high-impact event, people may want to learn the latest information as the event develops. It is necessary to develop an effective real-time event summary system to obtain sufficient real-time intelligence information for effective decision-making. In order to acquire extensive coverage of event details, using pre-trained language models can help us expand more knowledge and trending topics about how event sto- ries develop. The benefit of this potential for event awareness deserves in-depth study. The view is taken, therefore, in this work, we utilized the above-mentioned approaches to produce a streamlined version of a dynamically changing event data from the collected Twitter data set; when users view this version, they will be able to immediately capture the most important part of the latest event content in real-time for making appropriate decisions. It is worth mentioning that “real-time event summarization” is different from “real-time event detection” methods. The summarization system we developed can sum- marize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. The summary generated by our system can represent a simplified version of the event tweet message of interest, starting at any point in time of the ongoing event. Appl. Sci. 2021, 11, 10596 3 of 29 The rest of the paper is organized as follows. Section 2 surveys several related work in the literature. The proposed system model and system framework are described in Section 3. Section 4 reports on our pre-trained language model for text summarization. In particular, we introduce our developed text summarization method using a hybrid pre-trained model system. In Section 5, we describe our experiments in evaluation on (i) the performance of the text summarization method based on Rouge evaluation method, and (ii) the effectiveness of the real-time summary method developed for intelligence collection and event monitoring. In particular, in this work, we provide a real-world case study, namely the COVID-19 pandemic event, to verify the applicability of the proposed method. In Section 6, the discussion of experimental results is addressed in detail. Section 7 concludes the work by summarizing our research contributions. 2. Related Work In this work, we attempt to build a real-time event summarization system which can generate a simplified version of a dynamically changing event data from the collected Twitter data set. When users read the generated sentences, they will be able to immediately capture the most important part of the latest event content in real time to make appropriate decisions. Using Twitter for event detection and event summarization is a vast subject. Twitter is also employed to better understand how the communication transfers during citizen movements and events. In our previous work [5,6] a real-time event detection approach has been well developed to gather tweets of detected events for awareness of emerging events in the real world. Summarization of a single text summarization is the task of automatically generating a shorter version of a document and retaining its most important information. The types of automatic text summarization can be classified as extractive and abstractive. The extractive text summarization extracts sentences from the texts, while the abstractive ones may produce sentences and phrases which do not appear in the original documents. In practice, the techniques of abstractive text summarization emulate human summarization in that it utilizes a vocabulary beyond the specified text and summarizes key points in the text. Due to such approaches emulate how people summarize material, they normally require more computing power such as several GPUs to train the model over many days for learning algorithms, thus it is generally more difficult to acquire a cost-effective solution. In this work, we concentrate on abstractive summarization techniques. While text summarization has been investigated for years, automatic summarization on social messages (e.g., tweets) is still in its infancy. Sharifi et al., used the Phrase Reinforcement algorithm to deal with tweet summarization [7]. Inouye developed a Hybrid TF-IDF algorithm to produce multiple post summaries [8]. In order to make our model to summarize text accurately, it needs to learn to “un- derstand” the input data. Google’s BERT (Bidirectional Encoder Representations from Transformers) is the perfect tool for natural language understanding. Before the emergence of the BERT model, several important techniques related to the modeling method we adopted were developed. Recently, the Recurrent Neural Networks (RNNs; including Long Short Term Memory Networks) have been widely used for many natural language processing applications. However, it needs massive amounts of data, costly data sources, and several hours of training to achieve satisfactory results. Due to the drawback of suffering from poor performance with very long sequences of data, Vaswani proposed a model called the “Transformer”, replacing the RNNs and Convolutional Neural Networks (CNNs) by utilizing an architecture comprised of feed forward networks and attention mechanisms [9]. An unsupervised learning architecture called BERT (Bidirectional Encoder Representations from Transformers) has been developed and outperformed almost all existing models in the competitions for various types of NLP tasks [10]. In addition, they published several pre-trained models which can be utilized for transfer learning for many tasks and applications [10]. Appl. Sci. 2021, 11, 10596 4 of 29 Text summarization is a technology that extracts the important text or phrase infor- mation from the original text and then presents a short message to the user. According to the content and method of abstracting, it can be divided into Extractive and Abstractive. Extractive summarization is a technique that uses statistics that extracts the most important words and phrases in a text according to the given weight, and then arranges and combines these features to generate a summary of the text; abstractive summarization is a technique that uses machine learning methods to generate a summary of the text after understanding the content of the text [11–14]. Most of the extractive summarization are generated by extracting human-designed words or phrases as features in the original text [15–17], so that these extracted features are the most useful and important words and phrases in the text to achieve a more accurate final summary [18]. These features are usually rated according to their importance in the title, word frequency, word position, sentence length, similarity between sentences, num- ber of occurrences of words, and famous words [19], and then rearranged and combined according to this rating to generate the summary. The method has been studied in various applications [20–22]. For example, Isonuma [23] proposed a framework for extractive summarization using the dataset NIKKEI financial report corpus and the New York Times annotation corpus. Nallapati [24] proposed a Recurrent Neural Network (RNN) to extract the words and phrases that are relevant to the topic. Nallapati [24] proposed SummaRuN- Ner, a language model based on Recurrent Neural Network (RNN), which has a different training mode from the usual models and can reduce the need for extractive feature labels during training. Gunawan [25] used sentence length, headline features, and keyword frequency to develop an unsupervised summarization model for 3 K articles. Narayan [26] summed up the concept of extracted summaries as a task for ranking sentence and presents a training method that uses reinforcement learning to optimize the reward shaping that related to the task by exploring the space of candidate summaries. Dong [27] presents a framework for extractive summarization through the neural networks and enhanced learn- ing contexts, which did not require sentence-level extraction labels on extracted summaries and optimized the ROUGE score of model that generates the summary. Rush [28] presented a Neural Attention Model for Abstractive Sentence Summariza- tion, combining the chance model with the generation algorithm to produce an accu- rate summary, and acquires good results in the Gigaword and DUC datasets, on which Chopra [29] further improves the performance of the summary after replacing the decoder with an RNN. Song [30] proposed an LSTM-combined CNN abstractive text summarization framework (ATSDL), which extracts key phrases from the text and generates a smooth and grammatical summary by deep learning models. Paulus [31] proposes a neural network model for dealing with longer textual data, and the state-of-the-art is achieved in the experiment using 12 data sets. Chen [32] proposed a sentence-level learning method that allows abstractive summarization models to learn the hierarchical nature of words and sentences. Masum [33] used a bidirectional RNN with an LSTM as the encoding layer and an attention model as the decoding layer to create a short, fluent and comprehensi- ble summary of a text document. Cai [34] proposed an abstractive text summarization model RC-Transformer (RCT), which adds an RNN encoder to capture the context of the sequence and adds a convolutional layer to extract important sentences and understand the semantics to generate a paragraph of extremely high quality summary. In recent years, transfer learning methods using pre-trained models have achieved good results in most natural language processing tasks [35]. The self-supervised pre- training of the model on a large amount of unlabeled data will affect the effect of transfer learning, such as building a language model or filling in the blanks of words. Therefore, fine- tuning a small data set in the pre-trained model will be more effective. Devlin [10] proposed the BERT (Bidirectional Encoder Representations from Transformers) model in 2018, which has had a great impact on the field of natural language processing. BERT provides a pre- trained bidirectional feature for unlabeled text data in the context of all levels. Automatic text summarization is a hot research topic in the field of natural language processing. Appl. Sci. 2021, 11, 10596 5 of 29 Many studies use deep learning technology to obtain important information from text data, so that the machine can understand the text and generate a text-related summary. Therefore, many recent studies have applied pre-training models to text summarization. Liu [36] proposed a framework built using BERT. The framework can provide extractive and abstractive types of text summarization. The experimental results on three data sets (CNN/DailyMail, NYT and XSum) show that their model achieves the state-of-the-art result under the automatic and human-based evaluation protocol. Kieuvongngam [37] used the BERT model and GPT-2 model to conduct text summarization on the open research data set of COVID-19 and extracted the keywords and semantic meaning of the sentences from the original text data through the above-mentioned pre-trained model. Their system provides a concise and meaningful summary for the medical literature. Khandelwal [38] used the pre-trained Transformer language model to generate summaries from the original text data. Compared with the pre-trained Transformer encoder-decoder network, this model has significantly improved results. Zhang [39] proposed a pre-trained model HIBERT (Hierarchical Bidirectional Encoder Representations from Transformers) for pre-training on text data and unlabeled data. This model can achieve higher results on the CNN/Dailymail data set and the New York Times data set. Therefore, the experimental results show that the use of pre-trained models can make large open data sets acquire better results. Farahani [40] mentioned two approaches to deal with the task of text summarization using mT5 model and ParsBERT model and got good results on a data set named pn-summary for Persian abstractive text summarization. Ma [41] proposed a pre-trained model T-BERTSum for text summarization, which captures the key words of the topic information of social media, understands the meaning of the sentence and judges the topic of the message discussion, and then generates a high-quality section Summary. Kerui [42] uses BERT, Seq2seq and reinforcement learning to form a text summary model. Garg [43] uses T5, one of the most advanced pre-trained models, to perform a summary task on a data set with 80,000 news articles, and the results indicate that the summary generated by T5 has better quality than those generated by other models. Daiya [44] has developed a pre-trained language model ENEMAbst that can be used in extractive and abstractive summarization techniques. Both abstractive summaries and extractive summaries on the CNN/DailyMail dataset obtain the most excellent results. Fecht [45] proposed a pre-trained language model that can be used for abstractive text summarization, and combined with the use of BERT to experiment on the German summarization data set. After performing the contextual word embedding task through BERT, the model can improve the model’s ability to generate high-quality German summaries. In Table 1, some important related work is shown for comparison. In summary, in our work, social media (such as Twitter), pre-trained models, and text summaries (including extractive and abstractive methods) are the three key elements of event summarization. From the above literature review, the text summaries of specific event tweets on social media have gradually become one of the most popular summariza- tion research topics in recent years [1–3,41,42]. However, extractive summarization cannot obtain many key sentences in the tweet-form data, resulting in unsatisfactory quality of the generated summaries. Furthermore, due to our task requires extensive domain knowledge beyond the meaning of individual words and sentences, in this work we limit our focus to abstractive summarization algorithms. On the other hand, in the recent references of text summarization, there are more and more work using pre-training language models [35–44]. Among them, BERT is the pre-trained model mainly used by most research teams. Most researchers use the improved BERT model architecture in their work [36,37,39–41]. How- ever, because the sentences in the pre-training data set of BERT have correct grammar, BERT does not perform well for tweets with less grammar. In order to simplify the sum- marization process, the text-to-text pre-trained model (e.g., T5 model) has become a good choice. Unfortunately, for our problem domain, so far little attention has been paid to the development of a summary method that supports real-time functions to achieve the requirements of real-time event monitoring and intelligence collection. Motivated by this, in this work, we developed a novel model combining a real-time event summarization Appl. Sci. 2021, 11, 10596 6 of 29 system and a Twitter-based event detection method. The developed summarization sys- tem can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. To the best of our knowledge, this is the first attempt to develop such an event summarization system with multiple functional requirements to meet the challenge of intelligence gathering in emergency situations. Table 1. Related work of automated text summarization. Using Pre-Trained Method of Text Citing the Author Model Data Set Model Summarization NIKKEI Financial Isonuma et al. [23] CNN-LSTM No Extractive Report CNN/Daily Mail and Nallapati et al. [24] SummaRuNNer No Extractive DUC 2002 Amazon Fine Food Masum et al. [33] Bidirectional LSTM No Abstractive Reviews German Wikipedia Fecht et al. [45] BERT Yes Abstractive articles Kerui et al. [42] BERT Yes Abstractive LCSTS Garg et al. [43] T5 Yes Abstractive and Extractive 80 K News Articles Farahani et al. [40] mT5 Yes Abstractive Pn-summary Cai et al. [34] RC-Transformer Yes Abstractive Gigaword and Conditional DUC 2004 Chopra et al. [29] No Abstractive Generation by RNN Song et al. [30] CNN-LSTM No Abstractive Paulus et al. [31] Attention-RNN No Abstractive Ma et al. [41] T-BERTSum Yes Abstractive CNN/Daily Mail Zhang et al. [39] HIBERT Yes Abstractive Liu et al. [36] BERT Yes Abstractive and Extractive Daiya [44] ENEMABst Yes Abstractive and Extractive 3. Models and Methods 3.1. System Framework In this work, the pre-trained models are mainly used to perform summarization function on social media data to fine-tune the text summaries. With the rapid development of social media, many messages flow and gather on social media, forming various event topics. However, identifying “event” is one of the most difficult challenges of current decision support systems. If the system encounters an emergency, it is usually unable to promptly notify users of the warning to avoid risks. A sensible solution is to employ the functions of the real-time event summarization to obtain emergent event information. As a result, the system framework of this study utilizes the method of event detection on social-media and messages to extract their main content, and then summarized to understand the evolution of important events. Figure 1 illustrates the conceptual diagram of the process followed to extract real-time event intelligence for both the event detection and summarization. The system is designed as a two-pass process, consisted of two modules, say, real-time event detection module and event-message accumulation and summarization module. More details on real-time event summarization are presented in the following. Firstly, in this work, we combine a developed Twitter-based real-time event detection algorithm with pre-trained language models for summarizing emergent events. We used an online text-stream clustering algorithm and self-adaptive method developed to gather the Twitter data for detection of major international events. The unimportant messages on the data set are removed through the trained SBERT based classifier. After that, the DBSCAN algorithm is used to cluster similar topics in each day into hot topics (i.e., event topics). These event topics in Twitter data are generated by using the algorithm we developed to generate storylines. When the hot topics of each day are summarized, the specific topic and other hot topics are semantically correlated. When the correlation exceeds a certain threshold, it is added to the storyline. After daily updates, through the Appl. Sci. 2021, 11, 10596 7 of 29 Twitter based real-time intelligence data collection, a storyline of a specific event is jointly generated. Figure 1. Proposed Twitter based real-time event summarization and pre-trained models. The next work is to deal with text summarization of specific topics. In this work, we use the pre-trained model (i.e., T5; Text-to Text Transfer Transformer) to experiment with the Xsum data set to perform text summarization, and the Rouge metrics are used to evaluate the resulting summaries. The Xsum [46] data set contains 226,711 news articles and their summaries. News articles come from the BBC’s content from 2010 to 2017. The articles cover many fields such as news, politics, sports, weather, business, technology, science, health, family, education, entertainment and art, etc. We split the data set into 90% training data set, 5% validation data set and 5% test data set. The pre-trained language models and pre-training process will be described later in more detail. Subsequently, we started to use the trained model to summarize the incoming Twitter data set for experimentation. The large amount of events and related tweets detected by the real-time event detection system are grouped into about hundreds of event topics every day. Taking the Twitter data in 2020 as an example, there are about 36,000 event topics on Twitter. Using the resulting models of text summarization on popular topics of social media in this work, the system is used to perform real-world event summarization tasks and analyze the results. More details about pre-trained language models for text summarization are presented in the following section. Appl. Sci. 2021, 11, 10596 8 of 29 3.2. Proposed Approach for Collectively Summurizing Text Streams The first part is the real-time event detection algorithm, which analyzes the life cycle of events and the distribution of word weights through dynamic term weighting [47] in the instant message stream. In the early stage of the event, potential major events are captured, and then the real-time event cluster is combined into an event sequence through the self-adaptive stream clustering method. The algorithm for real-time event detection is shown in Algorithm 1. Algorithm 1 1st Pass real-time event detection INPUT: TS{1, 2, . . . , m} // Twitter Streams OUTPUT: RTED{1, 2, . . . , "}; //Real-Time Event Detection 1: Begin 2: for each TS{i} 3: BursT Dynamic Term Weighting(i) 4: RTED Self-Adaptive Stream Clustering(BursT, i) 5: end for 6: End BursT = BS  TI (1) w,t w,t w,t 8 9 < = ar q E ar q w ,t w ,t r r BS q = max   , 0 (2) w ,t : ; E ar w ,t TI q = 1  ln (3) w ,t [t , t] g [t , t] ln M m : w 2 W Dynamic Term Weighting can be divided into two factor values: Burst Score, BS and Term Importance, TI, as shown in Formula (1). In Formula (2), Burst Score is to calculate the burstiness of a keyword according to the rate at which the keyword arrives. In Formula (3), the key word weight is the logarithm of the ratio of the number of documents in the global time window to the number of messages containing the keyword to determine the popularity of the keyword [47]. Self-adaptive stream clustering algorithm [48] is used to represent the situation that, when a new message flows into the system, the message will be compared with the existing online group for similarity. When the similarity comparison result is greater than the threshold, the new message will be added to the most similar group in the online group. If it is less than the threshold, the system will create a new group and add the message. At the same time, the system will remove non-informative messages and aging online groups based on the adjustable sliding window module and the time decay function. The proposed self-adaptive stream clustering algorithm is shown in Algorithm 2. Once the adaptive stream clustering algorithm is used to process stream messages in real time, the number and weight of features will be automatically adjusted within the thread time width tw. As shown in the Algorithm 2, when a new message flows into the system, the message will be compared with the existing online cluster for similarity evaluation. When the similarity comparison result is greater than the threshold, the new message will be added to the most similar cluster. If it is less than the threshold, the system will create a new cluster and add the message. At the same time, the system will remove non-informative messages and aging online clusters based on the adjustable sliding window. Appl. Sci. 2021, 11, 10596 9 of 29 Algorithm 2 Self-adaptive stream clustering INPUT: MS{1, 2, . . . , m} // Message Streams OUTPUT: RTEC{1, 2, . . . , "}; HEC{1, 2, . . . , n} //Real-Time Event Cluster 1: Begin 2: for each MS{i} 3: for each RTEC{j} 4: if(d() – Q < 0) then 5: add RTEC{j} into HEC and revom RTEC{j}; 6: else 7: simOfMS&RTEC sim(MS{i}, RTEC{j})*d(); 8: if(simOfMS&RTEC > maxSimOfMS&RTEC) then 9: maxSimOfMS&RTEC simOfMS&RTEC; 10: idOfMaxC j; 11: else continue; 12: end for 13: if(maxSimOfMS&RTEC < Q) then 14: create a new RTEC for MS{i}; 15: else 16: merge MS{i} to RTEC{idOfMaxC}; 17: end for 18: until no messages is posted The second part of the event-message accumulation and summarization algorithm is to filter the event information through the BERT classifier, and then use the vector of the filtered message obtained by the Sentence-BERT model and put it into the DBSCAN algorithm to cluster the event information into For each sub-event, the summary con- tent is finally obtained through the T5 model. The proposed method for event-message accumulation and summarization algorithm is shown in Algorithm 3. Algorithm 3 2nd Pass event-message accumulation & summarization INPUT: TM{1, 2, . . . , n} // Text Message OUTPUT: HT{1, 2, . . . , "}; HTS{1, 2, . . . , m}; EHT{1, 2, . . . , k} // Hot Topic 1: Begin 2: for each TM{i} 3: Text BERT Classifier Model(TM{i}) 4: Sentence Embedding Sentence-BERT Model(Text) 5: HT DBSCAN(Sentence Embedding) 6: for each HT{j} 7: HTS T5 Text Summarization Model(j) 8: end for 9: end for 10: End 4. Proposed Pre-Trained Language Models for Text Summarization In this study, the pre-trained language model is applied to social media data to fine- tune the text summaries. With the development of the Internet generation, people prefer to express their feelings or the content they are interested in on social media, and many messages flow on social media every day. When there is more and more information on a specific topic, it is easy to form a hot topic. This research will focus on summarizing popular topics when social media content forms as the main core content, and then summarize them to explore the evolution of popular topics. Transfer learning, where the model is first pre-trained on data-rich tasks, and then fine-tuned on downstream tasks, has become a powerful technology in natural language processing (NLP) field. In this work, the T5 [4] pre-trained model is used to conduct experiments and the Rouge metrics are used to evaluate the resulting summaries. In addition, we also use the BERT pre-training model for experiments as a baseline for Appl. Sci. 2021, 11, 10596 10 of 29 performance comparison with the T5 model. The concepts of T5 and BERT pre-trained models are briefly described in the following subsections. 4.1. Bert Model and Sentence-BERT Model The establishment of the BERT model uses other techniques for training context- dependent representation models, including Semi-supervised Sequence Learning, GPT [49], ELMo [50] and ULMFit [51], but these models are unidirectional or shallow Bidirectional, which makes these models only use the words on their left or right to learn context, while BERT uses a deep bidirectional architecture so that the model can learn the words on its left and right. The BERT model also has good results in the task of semantic similarity, but when the BERT model calculates the semantic similarity, it needs to put two sentences into the model at the same time and calculate the information of the two sentences. This approach requires a lot of For calculation and time, if there are N sentences, it needs to calculate (N(N + 1))/2 times, and the similarity retrieval of 10,000 sentences requires about 50 million calculations, and it takes about 50 million times to execute using the BERT model. 65 h, which makes the BERT model unsuitable for semantic similarity retrieval and cluster analysis. Reimers proposed a model Sentence-BERT (SBERT) to improve the BERT architecture, using the twin and triplets network architecture to generate a sentence embedding that can compare semantics with cosine similarity, making it possible to calculate the similarity of 10,000 sentences The time required for the BERT is shortened from 65 h to 5 s, and the accuracy of the similarity calculation using the BERT model can still be maintained. SBERT added a layer of pooling operation behind the output layer of BERT to generate a revised sentence embedding. The maximum value is used as the sentence vector of the entire sentence. In Reimers’ experiment, the Mean-Pooling method is used as the sentence vector to acquire the best results. 4.2. T5 (Text-to-Text Transfer Transformer) Model T5 is a transformer-based model from Google that is trained in an end-to-end manner with text as input and modified text as output. It achieves state-of-the-art results on multiple NLP tasks such as summarization, question answering, machine translation using a text-to-text transformer trained on a large text corpus. The T5 model is composed of the encoder-decoder scheme of transformer model. The encoder and decoder on the T5- Small model are composed of 6-layer Transformer, the T5-Base model is composed of the BERT-base encoder and decoder, and the T5-Large model is composed of the BERT-large encoder and decoder group. T5-3B and T5-11B each increase the size of the model by 4 times. One of the goals of pre-training process is to provide general knowledge for the model to improve its performance on downstream tasks. T5 uses Transformer ’s Encoder- Decoder model during pre-training process. In the transfer learning of natural language processing, previous studies have used large unlabeled data sets for unsupervised learning. T5 attempts to measure the impact of the quality, characteristics and size of these unlabeled data sets, using C4 (Colosal Clean Crawled Corpus) corpus is used as the training data. 4.3. Developed Text Summarization Method Using a Hybrid Pre-Trained Model System Although there are a lot of content from social media platforms, there are often some unimportant daily message content. Therefore, when data is processed, tweets are also filtered through classification models to filter out unimportant daily tweets. We use about 200,000 HuffPost news headline data sets from 2012 to 2018 [52]. This data set contains 40 different categories of news headline categories, plus the unimportant daily tweet content from our collected tweets and mark it as the 41st category ‘OTHERS’. The text data are classified into 41 categories in total. The classification data is divided into three parts, 60% of the data is used to train the model, 20% of the data is used to verify whether the model has learned useful information, and 20% of the data is used to test the accuracy of the model. Since there are about 40,000 to 500,000 non-retweeted tweets in the daily collected Appl. Sci. 2021, 11, 10596 11 of 29 tweets, which contain a large number of daily trivial tweets, such as happy birthday, and holiday wishes, in this work the messages are filtered to avoid too much unnecessary data and reduce the calculation time. In order to alleviate this problem, we decided to use the characteristics of the BERT model as a classier to perform the function of tweet filtering. Subsequently, the filtered data is clustered through an unsupervised learning algorithm, and the content of the tweet is retained by the density of the content of the tweet, so as to ensure that the tweet is a hot topic on the day. As a result, in this system, the tweets are aggregated into a daily form and screened and filtered the content of tweets that were not reposted in English, and filtered the tweets through the BERT classification method, using the HuffPost data set plus the tweet content from the tweets. The nonessential daily tweets are filtered out and divided into 41 categories in total. The filtered tweets are processed through the SBERT (Sentence-BERT) model to obtain sentence vectors, and the Euclidean distance between the vectors is used to calculate the similarity between sentences. Here we use the DBSCAN algorithm to find more intensive discussion content, and finally form a daily hot topic. This study will learn text summaries by fine-tuning the pre-trained model in the Xsum data sets. In the experiment, the T5 pre-trained model is firstly used for text summarization, as shown in Figure 2. Subsequently, we compare the T5 model with the BERT pre-trained model for summarization performance comparison. Figure 2. Developed text summarization method using a hybrid pre-trained model system. 4.4. Evaluation Method for Text Summarization This research work mainly uses Rouge (Recall-Oriented Understudy for Gisting Evalu- ation) [53] as the evaluation method of text summarization. It uses automatically generated summaries and a set of reference summaries to do calculations to measure the similarity between automatically generated summaries and reference summaries. Rouge-N is used to calculate the number of n-grams shared by the generated summary and the reference summary, as shown in Formula (4). Count (gram ) å å match S2fRe f eremceSummariesg gram 2S n Rouge N = (4) Count(gram ) å å S2fRe f eremceSummariesg gram 2S Rouge-L is used to match the longest common subsequence (Longest Common Subse- quence, LCS) between two texts. 5. Experiment and Case Study We designed two sets of experiments to evaluate (i) the performance of the text summarization method based on Rouge evaluation method, and (ii) the effectiveness of the real-time summary method developed for intelligence collection and event monitoring, through the implementation of a case study in the real world. Appl. Sci. 2021, 11, 10596 12 of 29 5.1. Experiment and Performance Comparison We first performed experiments on the Xsum dataset on the BERT and T5 pre-trained models. After the evaluation of the text summaries, it can be found that the performance and output length of different pre-training models used in Tables 2–4 are different. Com- pared with BERT, T5 is more excellent in the performance of text summarization. The training time spent is much less than that of BERT, so the T5-base pre-trained model and the output length of 150 parameters are finally used in the text summary tool. Table 5 shows the result with the complete training data and most suitable parameters in the experiment. In order to save the training cost of the pre-trained model, the training data is compressed to find the most suitable parameters. Table 6 illustrates a comparison with other models. Table 2. Performance evaluation (output length: 80). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 28.17 8.08 21.76 T5-base 33.01 11.46 25.76 BERT 23.50 6.51 17.62 Table 3. Performance evaluation (output length: 100). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 28.21 8.37 21.66 T5-base 33.64 11.80 26.34 BERT 23.42 6.56 17.43 Table 4. Performance evaluation (output length: 150). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 28.97 8.26 22.07 T5-base 33.70 12.06 26.47 BERT 23.66 6.82 17.75 Table 5. Performance evaluation (using complete training data). Model/Score Rouge-1 Rouge-2 Rouge-L T5-small 32.64 11.37 25.66 T5-base 38.13 16.39 30.88 BERT 29.58 11.14 22.22 Table 6. Performance evaluation (model comparison). Model Dataset Rouge Score T5-base Xsum 38.13 UniLm CNN/DailyMail 43.08 BERT CNN/DailyMail 44.16 5.2. Xsum and Tweet Data Sets for Fine-Turning and Downstream Task In this study, we use the Xsum data set and a pre-trained language model, the T5 model, to train the summary model. Since the downstream task is to generate a summary of the tweet data set, the features required for the summary generated in the proposed system are mainly learned by fine-tuning the pre-trained T5 model in the Xsum data set. The Xsum dataset contains 226,711 BBC news articles and their summaries. Articles cover news, politics, sports, weather, business, technology, science, health, family, education, entertainment, art and other fields. The coverage and diversity of Xsum are very suitable for our downstream task, which is to summarize a large number of tweets. Therefore, Appl. Sci. 2021, 11, 10596 13 of 29 after the fine-tuning process, the well-trained model is used in massive tweet data sets for summarization. 5.3. Detected Hot Topics with Twitter Data In the hot topic detection system, a large amount of social media content is integrated into about hundreds of hot topics every day. Taking the Twitter data in 2020 as an example, there are about 36,000 hot topics in total, as shown in Figure 3. Based on the method of text summarization on popular topics in social media, the technique is applied to the task of text summarization in various fields on social media and specific data analysis. Figure 3. Daily detected event hot topics in 2020. Appl. Sci. 2021, 11, 10596 14 of 29 5.4. Case Study:Tweet Summary of the “COVID-19 Pandemic” Event In 2020, COVID-19 is undoubtedly one of the hottest topics. The COVID-19 pandemic has exposed many serious errors in global public policy preparation and response. Al- though it has been affected by other epidemics and pandemics for more than a decade, many countries have failed to incorporate lessons learned into their pandemic preparedness and response plans. One of the main factors is that we lack effective real-time event-related intelligence collection systems which can provide a summary of updates related to the COVID-19 pandemic in all countries. Hence, we cannot learn from the successful experi- ence of other countries in fighting the epidemic in time. Since the COVID-19 outbreak, very few studies that respond to practical needs, such as through rapid advice guidelines or application modules, have been conducted. However, many professionals, such as health workers, are in urgent need of effective ways to help people improve their well-being during this pandemic. According to the US Centers for Disease Control and Prevention (CDC), information on the outbreak of the coronavirus disease may increase the levels of distress among specific population—especially in those who were already struggling with mental health issues before the pandemic. They may be likely to experience higher levels of worry and stress regarding issues surrounding COVID-19. This trepidation may worsen over time and cause a long-lasting mental disturbance [54]. We wish that the proposed system can help the professionals detect and track COVID-19 events at an early stage as well as monitoring the emergency events within specific groups. The view is taken, therefore, in this case study, we combine our developed Twitter- based real-time event detection algorithm with pre-trained language models for summariz- ing COVID-19 events and sub-events. In Figure 4, it can be found that there are many detected hot-event topics around COVID-19 in 2020. In order to satisfy users’ information needs, we pick up all of the real-time event topics with a comprehensive spatio-temporal viewpoint. By analyzing the events detected by our system, the spatial and temporal impacts of the emerging events can be estimated. Therefore, we first build a representation of the spatial distribution by mapping the summary of events related to COVID-19 in the world in 2020, as shown in Figure 5. After that, in terms of temporal aspects, we constructed a storyline retrospection mechanism that demonstrates the story telling capability for users to understand the context of event development in a timeline, as illustrated in Figures 6 and A1–A3. Figure 4. Detected event topics related to COVID-19 in 2020. Appl. Sci. 2021, 11, 10596 15 of 29 Figure 5. Representation of a map of generated summaries related to COVID-19 events in 2020. In Figure 5, we present hot topics in the form of a map to illustrate global events to understand the spatial impact of the COVID-19 pandemic on the world. Furthermore, in this work, we developed a storyline retrospection mechanism that demonstrates the story telling capability for users to understand the context of event development in an efficient way. It can be used to identify events and their relations under an event topic and then composes a storyline which gives users the sketch of event evolution in the topic, by computing daily related tweets and summaries of formulated event topics. The main storyline construction removes irrelevant events and presents a main theme that exhibits the sequence of events in a hot topic. In Figures 6 and A1–A3, we can see the storyline of COVID-19 in 2020, from the beginning of the outbreak to the emergence of the final vaccine. Figure 7 shows the example resulting summaries of COVID-19 event related tweets in 2020, starting with a new virus related to the SARS outbreak that may cause Chinese health officials to discuss on social media Pneumonia. From the time trend shown in Figures A1–A3, when the COVID-19 pandemic first broke out, social media were mainly discussing where the pandemic occurred and how to properly protect themselves from the pandemic. After a period of time, the social media began to gradually appear on the topic of what drugs may be able to treat the symptoms of COVID-19 or the topic of vaccine-related research and development. As time progressed, social media began to see related topics such as whether the vaccine was safe. Appl. Sci. 2021, 11, 10596 16 of 29 Figure 6. Generated storyline related to COVID-19 from January to March 2020. Appl. Sci. 2021, 11, 10596 17 of 29 Figure 7. A sample resulting summary of formulated event topics (9 January 2020). 5.5. Human Evaluation for Quality Assessment of Generated Summaries In order to verify the quality of the text summary results, we conducted a survey on the example resulting summaries (in Figures 7 and A4–A11) with human judges for quality assessment of generated summaries. In this work, we invited 39 participants from University graduate students, including 19 males and 20 females. Their age ranges from 20 to 30 years old. Participants were asked to rate the summaries on a 1–5 scale, with higher being better. The assessed summaries were selected using random sampling. Thus, we randomly selected 10 sample summaries from our summarization system for evaluation. In this work, the summaries are evaluated from three aspects. The first aspect is to assess the consistency of the topic between the tweets and the summary. The second aspect is to evaluate the closeness to the “facts” of the summary. The third aspect is to assess the degree of grammatical correctness of the summary. The summaries are reviewed on a scale of 1 to 5 points, and the evaluation results are shown in Table 7. Options for grading summaries are as follows: 1: Bad 2: Below Average 3: Average 4: Good 5: Excellent. In Table 7, we find that the mean values (i.e., 3.92, 3.89, 3.92, respectively) are close to 4 points, which show that the experimental result is acceptable and very close to the performance of human. In addition, we used Cronbach’s alpha test to check whether the multi-question Likert scale survey is reliable and acquires results (Cronbach’s alpha coefficients are 0.725, 0.812, 0.846). It shows that Cronbach’s alpha coefficient is greater than 0.7, and the resulting coefficient indicates that the results are reliable. Table 7. Quality assessment of generated summaries. Consistency of the Closeness to the Grammatical Topic Facts Correctness No. of Participant 39 39 39 Minimum 2 2 1 Maximum 5 5 5 Mean 3.92 3.89 3.92 Median 4 4 4 Std. Deviation 0.824 0.819 0.928 Std. Error 0.132 0.131 0.148 Cronbach’s alpha 0.725 0.812 0.846 Appl. Sci. 2021, 11, 10596 18 of 29 6. Discussion In this work we use natural language processing techniques to deal with social media data, extract event topics from daily social messages, and summarize them to quickly grasp various event intelligence information. In the experiment, we adopted a pre-trained model (i.e., T5) method to quickly fine-tune the model to learn text summaries through the Xsum data set, which was achieved a score of 38.13 in the evaluation of Rouge-1. After that, the trained model is actually applied to real social media data. We started to use the trained model to summarize the incoming Twitter data set for experimentation. The model summarizes the collected hot topics of social media into a short sentence. The resulting summary can help users quickly understand topic trends on social media, most of which are in line with the development of real-world events. Furthermore, in order to verify the quality of the text summary results, we conducted a survey on the resulting summaries with human judges for quality assessment of generated summaries. The experimental result is discussed in detail as follows: (1) The T5 model is trained in an end-to-end and text-to-text manner, with text as input and summary text as output. In this work, we demonstrated that T5 has achieved state-of-the-art results in many NLP benchmarks and can be fine-tuned to perform automatic summarization tasks well. (2) As mentioned earlier, due to the real-time nature of social media, an automatic event summary system using Twitter data was developed in this work to provide people with an updated summary of the event while the situation is still evolving. However, it is worth mentioning that the “real-time” here means that the summary system can expand or update the summary at any point in time, rather than real-time computing. (3) Going further, “real-time event summarization” is different from “real-time event detection” methods. The summarization system we developed can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. Compared to real-time event detection methods, detected events on Twitter are often short-term. The summary generated by our system can represent a simplified version of the event tweet message of interest, starting at any point in time of the ongoing event. (4) Generally speaking, abstractive summarization models are prone to produce gram- matically correct incoherent sentences. In the work of summarizing text, it is easy for human researchers to make language errors. The advantages of using abstractive summarization models enable us to alleviate some of the grammatical problems of real people. Despite the results of the system evaluation show that the summaries we got are not perfect in terms of grammatical fluency, the results are acceptable and close to the performance of human researchers in the real world. (5) Although this developed model can quickly summarize the text, it still has certain limitations in generality. The social media data used in this study only uses English data on Twitter to generate summaries. Although English is the common language of most people in the world, it must be only through English to understand the various events in the world, there are still some prejudices. In the future, we consider using multilingual models and multilingual data sets (made by different native speakers) to train the model so that we can more accurately grasp various world events from different perspectives. (6) In this work, we implemented a method of generating storylines using the automatic summarization technique we developed. The generated storyline is composed of sub-events or key events related to a given major event. The sub-events show the status of the main event in progress. Automated storyline generation has been a research issue since nearly the inception of artificial intelligence. In particular, the storylines of events relate to identify entities and summarize events that leads to the event of interest. However, the proposed system model has not yet performed the full function of the “automated storyline generation” paradigm for understanding event development. This is because in this work, we have not provided methods for Appl. Sci. 2021, 11, 10596 19 of 29 analyzing the evolution of specific event entities and exploring the entities and their relationships in event stories. We will leave the work of constructing event storyline for future work to complete. 7. Conclusions The uncertainty of emergencies and natural disasters may affect companies and everyone’s lives. During such high-impact events, people may want to stay up-to-date as the event develops because they want to better understand the situation and possible trend of the event. In this case, it is almost impossible for anyone to browse through all of the news articles to understand the event stories required for decision-making. Automated summarization systems that access a large number of information sources in real time can help by providing event summaries, which are updated whenever important sub-events are detected. The update of the event will help those interested persons to understand the situation of the current event. In this work, we developed a hybrid method to combine a real-time event summariza- tion system and a self-adaptive Twitter-based event detection method, using pre-trained language models on the tweet data set for online machine learning. Hence, each user can obtain real-time information about specific incidents (or events) and the agility to make appropriate decisions quickly. The goal of this work can be divided into the following two parts: (1). In a high-impact event, people may want to learn the latest information as the event develops. It is necessary to develop an effective real-time event summary system to obtain sufficient real-time intelligence information for effective decision-making; (2). In order to acquire extensive coverage of event details, using pre-trained language models can help us expand more knowledge and trending topics about how event stories develop. Our research is intended as a starting point for exploring the benefits of this potential for event awareness. The novelty and contributions of the work are summarized as follows: We implemented an end-to-end and text-to-text summarization system model, in which text streams as input and summary text as output. Our system model combined a real-time event summarization system and a self- adaptive Twitter-based event detection method. The developed summarization system can summarize the continuously accumulated hot topic tweets into a short sentence, instead of just performing real-time event detection. The developed system used pre-trained language models to help users obtain more knowledge and trending topics about how event stories develop. The system can support users to obtain sufficient real-time intelligence information in specific events to make effective decisions. For future work, it will be interesting to further incorporate the sentiment of the tweet into the summary of related events to predict the world market trends affected by the event. On the other hand, as mentioned earlier, building a summary system in the “automatic storyline generation” paradigm is very useful for understanding the details of event development. In this way, it is possible to explore the evolution of a specific event entity in the event story and the relationship between the entity and its event. In addition, by taking advantage of the potential of pre-trained language models, it will be very useful to gather intelligence on multilingual event summaries from multilingual social media messages. Finally, the use of automatic video summarization methods to extend the developed real-time text summarization methods to generate event summaries is also a promising application. Appl. Sci. 2021, 11, 10596 20 of 29 Author Contributions: Conceptualization, C.-H.L. and Y.-L.C.; methodology, C.-H.L.; software, Y.- L.C.; validation, H.-C.Y., Y.J.C. and Y.-L.C.; formal analysis, H.-C.Y.; investigation, C.-H.L.; resources, Y.J.C.; data curation, H.-C.Y.; writing—original draft preparation, C.-H.L.; writing—review and editing, C.-H.L.; visualization, Y.-L.C.; supervision, C.-H.L.; project administration, C.-H.L.; funding acquisition, C.-H.L. All authors have read and agreed to the published version of the manuscript. Funding: This research received no external funding. Institutional Review Board Statement: Not applicable. Informed Consent Statement: Not applicable. Data Availability Statement: The study did not report any data. Conflicts of Interest: The authors declare no conflict of interest. Appendix A. Examples of Generated Storyline Related to COVID-19 Pandemic Figure A1. Generated storyline related to COVID-19 from April to June 2020. Appl. Sci. 2021, 11, 10596 21 of 29 Figure A2. Generated storyline related to COVID-19 from July to September 2020. Appl. Sci. 2021, 11, 10596 22 of 29 Figure A3. Generated storyline related to COVID-19 from October to December 2020. Appl. Sci. 2021, 11, 10596 23 of 29 Appendix B. Example Model Outputs: Tweet Summaries of the “COVID-19 Pandemic” Sub-Events Figure A4 shows that on 24 January 2020, the World Health Organization issued a global warning about the new coronavirus that was discussed in social media. Figure A4. A sample resulting summary of formulated event topics (24 January 2020). Figure A5 shows that on 12 February 2020, the World Health Organization officially named the new coronavirus COVID and was discussed in social media. Figure A5. A sample resulting summary of formulated event topics (12 February 2020). Appl. Sci. 2021, 11, 10596 24 of 29 Figure A6 illustrates the topics that will be postponed for the 2020 Olympic Games discussed on 30 March 2020. Figure A6. A sample resulting summary of formulated event topics (30 March 2020). Figure A7 shows that the possible variants of the new coronavirus were discussed in social media on 6 May 2020, and they may become more toxic and more contagious than the original. Figure A7. A sample resulting summary of formulated event topics (6 May 2020). Appl. Sci. 2021, 11, 10596 25 of 29 Figure A8 shows that the Dexamethasone drug was discussed in social media on 17 June 2020 to help save the lives of patients with the new coronavirus. Figure A8. A sample resulting summary of formulated event topics (17 June 2020). Figure A9 shows that on 15 July 2020, in social media, the Center for Disease and Control of the United States called on people to wear masks to effectively prevent the new coronavirus. Figure A9. A sample resulting summary of formulated event topics (15 July 2020). Appl. Sci. 2021, 11, 10596 26 of 29 Figure A10 shows that it was discussed in social media on 16 November 2020. American drug maker Moderna stated that its COVID vaccine can effectively prevent the coronavirus. Figure A10. A sample resulting summary of formulated event topics (16 November 2020). Figure A11 shows that Pfizer’s COVID vaccine was discussed in social media on 9 December 2020, in compliance with the safety and effectiveness guidelines for emergency use. Figure A11. A sample resulting summary of formulated event topics (9 December 2020). Appl. Sci. 2021, 11, 10596 27 of 29 References 1. Rudrapal, D.; Das, A.; Bhattacharya, B. A survey on automatic Twitter event summarization. J. Inf. Process. Syst. 2018, 14, 79–100. 2. Chakrabarti, D.; Punera, K. Event summarization using tweets. In Proceedings of the International AAAI Conference on Web and Social Media, Barcelona, Spain, 17–21 July 2011; Volume 5. 3. Chua, F.C.T.; Asur, S. Automatic summarization of events from social media. In Proceedings of the Seventh International AAAI Conference on Weblogs and Social Media, Cambridge, MA, USA, 8–11 July 2013. 4. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv 2019, arXiv:1910.10683. 5. Lee, C.-H. Mining spatio-temporal information on microblogging streams using a density-based online clustering method. Expert Syst. Appl. 2012, 39, 9623–9641. [CrossRef] 6. Lee, C.-H.; Chien, T.-F. Leveraging microblogging big data with a modified density-based clustering approach for event awareness and topic ranking. J. Inf. Sci. 2013, 39, 523–543. [CrossRef] 7. Sharifi, B.; Hutton, M.A.; Kalita, J. Summarizing microblogs automatically. In Proceedings of the 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Los Angeles, CA, USA, 2–4 June 2010; pp. 685–688. 8. Inouye, D.; Kalita, J.K. Comparing Twitter Summarization Algorithms for Multiple Post Summaries. In Proceedings of the 2011 IEEE Third Int’l Conference on Privacy, Security, Risk and Trust and 2011 IEEE Third International Conference on Social Computing, Boston, MA, USA, 9–11 October 2011; pp. 298–306. 9. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. 10. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. 11. Rachabathuni, P.K. A survey on abstractive summarization techniques. In Proceedings of the 2017 International Conference on Inventive Computing and Informatics (ICICI), Coimbatore, India, 23–24 November 2017; pp. 762–765. 12. Gambhir, M.; Gupta, V. Recent automatic text summarization techniques: A survey. Artif. Intell. Rev. 2017, 47, 1–66. [CrossRef] 13. Neto, J.L.; Freitas, A.A.; Kaestner, C.A.A. Automatic Text Summarization Using a Machine Learning Approach. In Proceedings of the 16th Brazilian Symposium on Artificial Intelligence, Porto de Galinhas/Recife, Brazil, 11–14 November 2002; pp. 205–215. 14. Tas, O.; Kiyani, F. A survey automatic text summarization. J. PressAcademia Procedia 2017, 5, 205–213. [CrossRef] 15. Radev, D.R.; Allison, T.; Blair-Goldensohn, S.; Blitzer, J.; Celebi, A.; Dimitrov, S.; Drabeck, E.; Hakim, A.; Lam, W.; Liu, D.; et al. MEAD-a platform for multidocument multilingual text summarization. In Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC’04), Lisbon, Portugal, 26–28 May 2004; pp. 699–702. 16. Nenkova, A.; Vanderwende, L.; McKeown, K. A compositional context sensitive multi-document summarizer: Exploring the factors that influence summarization. In Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle, WA, USA, 6–11 August 2006; pp. 573–580. 17. Filatova, E.; Hatzivassiloglou, V. Event-based extractive summarization. In Proceedings of the ACL-04 Workshop, Association for Computational Linguistics, Barcelona, Spain, 21–26 July 2004; pp. 104–111. 18. Moratanch, N.; Chitrakala, S. A survey on extractive text summarization. In Proceedings of the 2017 International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 10–11 January 2017; pp. 1–6. 19. Mutlu, B.; Sezer, E.A.; Akcayol, M.A. Multi-document extractive text summarization: A comparative assessment on features. J. Knowl.-Based Syst. 2019, 183, 104848. [CrossRef] 20. Kupiec, J.; Pedersen, J.; Chen, F. A trainable document summarizer. In Proceedings of the 18th Annual International ACM SIGIR conference on Research and Development in Information Retrieval–SIGIR ’95, Seattle, WA, USA, 9–13 July 1995; pp. 68–73. 21. Conroy, J.M.; O’Leary, D.P. Text summarization via hidden Markov models. In Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval—SIGIR ’01, New Orleans, LA, USA, 9–13 September 2001; pp. 406–407. 22. Woodsend, K.; Lapata, M. Automatic generation of story highlights. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 11–16 July 2010; pp. 565–574. 23. Isonuma, M.; Fujino, T.; Mori, J.; Matsuo, Y.; Sakata, I. Extractive Summarization Using Multi-Task Learning with Document Classification. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2101–2110. 24. Nallapati, R.; Zhai, F.; Zhou, B. Summarunner: A recurrent neural network based sequence model for extractive summarization of documents. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 25. Gunawan, D.; Pasaribu, A.; Rahmat, R.F.; Budiarto, R. Automatic Text Summarization for Indonesian Language Using TextTeaser. In IOP Conference Series: Materials Science and Engineering; IOP Publishing Ltd.: Semarang, Indonesia, 2016; Volume 190, p. 012048. 26. Narayan, S.; Cohen, S.B.; Lapata, M. Ranking Sentences for Extractive Summarization with Reinforcement Learning. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; pp. 1747–1759. Appl. Sci. 2021, 11, 10596 28 of 29 27. Dong, Y.; Shen, Y.; Crawford, E.; Van Hoof, H.; Cheung, J.C.K. BanditSum: Extractive Summarization as a Contextual Bandit. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 3739–3748. 28. Rush, A.M.; Chopra, S.; Weston, J. A Neural Attention Model for Abstractive Sentence Summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 379–389. 29. Chopra, S.; Auli, M.; Rush, A.M. Abstractive Sentence Summarization with Attentive Recurrent Neural Networks. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 93–98. 30. Song, S.; Huang, H.; Ruan, T. Abstractive text summarization using LSTM-CNN based deep learning. J. Multimed. Tools Appl. 2018, 78, 857–875. [CrossRef] 31. Paulus, R.; Xiong, C.; Socher, R. A deep reinforced model for abstractive summarization. arXiv 2017, arXiv:1705.04304. 32. Chen, Y.-C.; Bansal, M. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 675–686. 33. Masum, A.K.M.; Abujar, S.; Talukder, A.I.; Rabby, A.S.A.; Hossain, S.A. Abstractive method of text summarization with sequence to sequence RNNs. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; pp. 1–5. 34. Cai, T.; Shen, M.; Peng, H.; Jiang, L.; Dai, Q. Improving Transformer with Sequential Context Representations for Abstractive Text Summarization. In Proceedings of the CCF International Conference on Natural Language Processing and Chinese Computing, Dunhuang, China, 9–14 October 2019; pp. 512–524. 35. Qiu, X.; Sun, T.; Xu, Y.; Shao, Y.; Dai, N.; Huang, X. Pre-trained models for natural language processing: A survey. J. Sci. China Technol. Sci. 2020, 63, 1872–1897. [CrossRef] 36. Liu, Y.; Lapata, M. Text Summarization with Pretrained Encoders. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 3721–3731. 37. Kieuvongngam, V.; Tan, B.; Niu, Y. Automatic text summarization of Covid-19 medical research articles using bert and gpt-2. arXiv 2020, arXiv:2006.01997. 38. Khandelwal, U.; Clark, K.; Jurafsky, D.; Kaiser, L. Sample efficient text summarization using a single pre-trained transformer. arXiv 2019, arXiv:1905.08836. 39. Zhang, X.; Wei, F.; Zhou, M. HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 5059–5069. 40. Farahani, M.; Gharachorloo, M.; Manthouri, M. Leveraging ParsBERT and Pretrained mT5 for Persian Abstractive Text Summa- rization. In Proceedings of the 2021 26th International Computer Conference, Computer Society of Iran (CSICC), Tehran, Iran, 3–4 March 2021; pp. 1–6. 41. Ma, T.; Pan, Q.; Rong, H.; Qian, Y.; Tian, Y.; Al-Nabhan, N. T-BERTSum: Topic-Aware Text Summarization Based on BERT. J. IEEE Trans. Comput. Social Syst. 2021, 1–12. [CrossRef] 42. Kerui, Z.; Haichao, H.; Yuxia, L.; Zhang Kerui Lazada Group, China; Hu Haichao SF Technology, Shenzhen China; Liu Yuxia SF Technology, Shenzhen China. Automatic Text Summarization on Social Media. In Proceedings of the 2020 4th International Symposium on Computer Science and Intelligent Control, Newcastle upon Tyne, UK, 17–19 November 2020; pp. 1–5. 43. Garg, A.; Adusumilli, S.; Yenneti, S.; Badal, T.; Garg, D.; Pandey, V.; Nigam, A.; Gupta, Y.K.; Mittal, G.; Agarwal, R. NEWS Article Summarization with Pretrained Transformer. In Communications in Computer and Information Science, Proceedings of the Advanced Computing, IACC 2020, Panaji, Goa, India, 5–6 December 2020; Springer: Singapore, 2020; pp. 203–211. [CrossRef] 44. Daiya, D. Combining Temporal Event Relations and Pre-Trained Language Models for Text Summarization. In Proceedings of the 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 14–17 December 2020; pp. 641–646. 45. Fecht, P.; Blank, S.; Zorn, H.P. Sequential Transfer Learning in NLP for German Text Summarization. In Proceedings of the 4th edition of the Swiss Text Analytics Conference, Winterthur, Switzerland, 18–19 June 2019. 46. Narayan, S.; Cohen, S.B.; Lapata, M. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 1797–1807. 47. Lee, C.-H.; Wu, C.-H.; Chien, T.-F. BursT: A Dynamic Term Weighting Scheme for Mining Microblogging Messages. J. Int. Symp. Neural Netw. 2011, 6677, 548–557. [CrossRef] 48. Lee, C.-H.; Wu, C.-H. A Self-adaptive Clustering Scheme with a Time-Decay Function for Microblogging Text Mining. J. Inf. Technol. 2011, 185, 62–71. [CrossRef] 49. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Proceedings of the 27th Interna-tional Conference on Neural Information Processing Systems, Montréal, QC, Canada, 6–12 December 2014; pp. 3104–3112. Appl. Sci. 2021, 11, 10596 29 of 29 50. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; Volume 1, pp. 2227–2237. 51. Howard, J.; Ruder, S. Universal Language Model Fine-tuning for Text Classification. arXiv 2018, arXiv:1801.06146. 52. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv 2019, arXiv:1908.10084. 53. Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Workshop on Text Summarization Branches Out (WAS 2004), Barcelona, Spain, 21–26 July 2004; pp. 74–81. 54. Shigemura, J.; Ursano, R.J.; Morganstein, J.C.; Kurosawa, M.; Benedek, D.M. Public responses to the novel 2019 coronavirus (2019-nCoV) in Japan: Mental health consequences and target populations. J. Psychiatry Clin. Neurosci. 2020, 74, 281–282. [CrossRef] [PubMed]

Journal

Applied SciencesMultidisciplinary Digital Publishing Institute

Published: Nov 11, 2021

Keywords: text summarization; event detection; machine learning; natural language processing; Twitter

There are no references for this article.