Topic Modelling: Going Beyond Token Outputs

Topic modelling is a text mining technique for identifying salient themes from a number of documents. The output is commonly a set of topics consisting of isolated tokens that often co-occur in such documents. Manual effort is often associated with interpreting a topic's description from such tokens. However, from a human's perspective, such outputs may not adequately provide enough information to infer the meaning of the topics; thus, their interpretability is often inaccurately understood. Although several studies have attempted to automatically extend topic descriptions as a means of enhancing the interpretation of topic models, they rely on external language sources that may become unavailable, must be kept up-to-date to generate relevant results, and present privacy issues when training on or processing data. This paper presents a novel approach towards extending the output of traditional topic modelling methods beyond a list of isolated tokens. This approach removes the dependence on external sources by using the textual data itself by extracting high-scoring keywords and mapping them to the topic model's token outputs. To measure the interpretability of the proposed outputs against those of the traditional topic modelling approach, independent annotators manually scored each output based on their quality and usefulness, as well as the efficiency of the annotation task. The proposed approach demonstrated higher quality and usefulness, as well as higher efficiency in the annotation task, in comparison to the outputs of a traditional topic modelling method, demonstrating an increase in their interpretability.


Introduction
The increase of textual user-generated content on online platforms and social networks has become a key source for a wealth of information.Such platforms are often used to share information such as news, brands, political discussions, and more [1].Nevertheless, the large volume of text data that is broadly available on the web increases the challenge of identifying the most relevant information in real-time.Text mining is a promising solution to the problem of information overload associated with summarising and understanding vast amounts of unstructured text originating from diverse sources.More specifically, topic modelling aims to identify and extract salient concepts or themes, also known as "topics", distributed across a collection of documents [2].
Many studies have focused on extracting topics expressed in several domains, such as online social networks (e.g.[3], [4]), particularly to identify influential individuals on a social media platform (e.g.[5]) and detecting signs of depression in related language on Twitter (e.g.[6]), clinical applications such as triaging patients based on referral letters (e.g.[7]), extracting scientific topics in academic journals (e.g.[8]), and most recently, understanding the public's opinion of governments during the COVID-19 pandemic (e.g.[9]).
The output from applying topic modelling to a collection of documents is commonly presented as a set of the top most co-occurring terms appearing in each topic [10], [11].Manual effort is often associated with interpreting a topic's description from a set of isolated tokens, where a topic is often given a title or a name to reflect the understanding of its underlying meaning [12].For example, "car, power, light, drive, engine, turn" may infer topics surrounding Vehicles, and "game, team, play, win, run, score" may infer Sports.However, one of the key concerns with topic models lies with the subjectivity surrounding this task, as well as how well human readers can understand the topics, otherwise referred to as topic interpretability [12].More often than not, topic modelling outputs may not adequately describe the meaning of the topic itself.Problems include generic topic descriptions with too many words [13], topics with disparate or poorly connected words [14], misaligned topics [15], and multiple nearly identical topics [13].Subsequently, this misinterpretation leads to ineffective information retrieval and decision making.
In this case, several studies have attempted to automatically extend topic descriptions, or select the best label for a given topic, as a means of enhancing the interpretation of topic models and reducing the human in the loop.Such approaches often rely on external textual sources (e.g.Wikipedia, WordNet) or sophisticated large language models (e.g.ChatGPT).In recent advancements, the development and maintenance of machine learning models and resources in privacy-aware settings have seen significant improvements, making the utilisation of external resources more feasible and secure than ever.However, despite these advancements, relying solely on such resources can still pose inherent risks and limitations.Firstly, the availability and stability of these resources remain a concern.Although less frequent, outages or restrictions in access to these external resources can disrupt the continuity and reliability of research relying on them.Moreover, the relevance and currency of information from these sources are not always guaranteed.In rapidly evolving fields, external resources might not be updated promptly, leading to the generation of outputs based on outdated information.Secondly, while privacy-aware machine learning models have made strides, they are not entirely foolproof.There are ongoing concerns regarding data breaches and the potential for these models to inadvertently memorise and reveal sensitive data.This is particularly crucial in contexts dealing with confidential or proprietary information, where the inadvertent leakage of data through these models could have serious ramifications.
Given these considerations, this paper proposes an approach which aims to strike a balance between leveraging the advancements in privacy-aware machine learning and mitigating the risks associated with dependency on external resources.By developing a methodology that primarily relies on internal data processing and analysis, this method aims to maintain a high degree of control over data relevance and privacy, ensuring that the model remains robust and reliable.More specifically, this paper proposes: • A novel approach towards extending the output of traditional topic modelling methods beyond token outputs.
The initial experiments presented in this paper are tested using Latent Dirichlet Allocation (LDA), a traditional approach for distributing text segments into topics.However, given the positive findings of the experiments, and to demonstrate the generalisation of the proposed method, the approach is also applied when two of the latest state-of-the-art topic modelling methods, BERTopic and Top2Vec, are used.
• An evaluation of how such extended outputs improve the interpretability of topic descriptions from a human perspective.
The study was designed as shown in Figure 1: 1) pre-process text responses using traditional Natural Language Processing (NLP) techniques, 2) apply the topic modelling algorithm, 3) for each topic in 2, extend the token list output by applying keyword extraction to the original text segments, 4) map token outputs from 2 to the extended outputs in 3, and 5) evaluate the ease of interpreting outputs from 2 and 4 by human readers.
The remainder of this paper is structured as follows: Section 2 presents the related work, Section 3 discusses the selection of text corpora used to support the experiments herein and the techniques used to prepare the data for such experiments, Section 4 discusses topic modelling and how it was applied to the datasets, Section 5 discusses the proposed methodology of using keyword extraction to extended topic modelling outputs, Section 6 evaluates the ease of interpreting such outputs by human readers, Section 7 discusses the generalisation of the proposed method by using other topic modelling methods, Section 8 discusses the generalisation of the proposed method when using new data, and finally, Section 10 concludes the paper.

Related Work
Several current works that have adopted topic modelling often rely on the manual approach of interpreting and determining the topic's main descriptive label based on the token results generated by the topic modelling approach (e.g.[16], [17]).However, previous work has shown that using top terms are not enough for interpreting the coherent meaning of a topic following topic modelling [18] [19].Such studies include Lee et al. [20], who qualitatively evaluated how readers understand, assess, and refine topics.They demonstrate a disconnect between how readers perceive topic model outputs and the operations that current interactive systems support.In this case, several studies have attempted to automatically extend topic descriptions as a means of enhancing the interpretation of topic models.Recent works have explored the use of external sources, such as Wikipedia, WordNet, or other ontologies for supporting the automatic labelling of topics by deriving candidate labels using lexical-based (e.g.[21], [22], [18], [23], [24]) or graph-based algorithms applied on these sources (e.g.[25], [26]).For example, recently, both Allahyari et al. [27] and Kinariwala and Deshmukh [28] propose an ontological approach for generating topic labels.Both approaches use the semantic relatedness of LDA outputs and their ontological categories to determine topic labels.However, limitations of this approach include the reliance of an ontology of terms, which in the case of [28], is only 500 words which are categorised and limited to only four main domains (crime, environment, politics, and sports).Others, such as Bhatia et al. [23], have used word embeddings to map topic modelling outputs to Wikipedia article titles using cosine similarity and relative ranking measures.Whilst others, such as Bechara et al. [29], propose a transfer learning approach for topic labelling which uses domain-specific codebooks to automatically label topics.
Although the aforementioned works propose methods of automatically assigning more informative and more generic labels based on topic models outputs, they lack emphasis on whether outputs from such approaches are interpretable and meaningful to human readers.To address this, other studies have evaluated their outputs by asking human annotators to label or score them based on how semantically related they were to a topic, how meaningful they are, and their usefulness and coherence.Wan and Wang [30] propose an algorithm to extract text summaries that are much longer than keywords or phrases for describing topics.Aletras et al. [31] compare three different topic representations in a document retrieval task and report that readers find phrased labels easier to interpret in comparison to a list of terms.Lau et al. [21] also use phrases as topic labels, and propose a supervised approach for ranking candidate labels.In their work, candidate labels include the top five topic terms and noun chunks extracted from Wikipedia articles.Kou et al. [32] propose another method of semantically mapping topics to candidate labels using word vectors and letter trigram vectors.Mei et al. [18] propose to use n-grams for topic labelling and approach the problem of labelling as an optimisation problem which involves minimising the Kullback-Leibler divergence between word distributions and maximising the mutual information between a label and a topic model.Chang and Boyd-Graber [33] and Morstatter and Liu [12] use crowdsourcing to measure topic interpretability using a numerical score.They validate topic models using word intrusion, an approach which requires participants to study the top words within a topic and to identify which words that do not belong.In addition, they also validate topics models using topic intrusion, an approach similar task to word intrusion, where participants study the topic probabilities for a document and evaluate how relevant they are to their understanding of the text.
The work cited above focuses on extending topic modelling outputs, and more importantly, evaluating how interpretable such outputs are to human readers.However, such approaches use external textual sources that may be at risk of becoming unavailable.The size of such resources may also be relatively large, and therefore, generating results from them in real-time may be ineffective and computationally intense.There has yet to be an investigation into a lightweight approach that uses the textual data itself to extend topic modelling outputs for systems that require the almost immediate presentation of interpretable topics to humans.

Text Corpora and Data Preparation
To conduct the experiments described herein, the data was provided by our industrial research collaborator, a software company that focuses on the concept of crowdsourcing solutions to strategic business challenges.The eight datasets consist of short informal responses to a range of different questions, propositions, and requests for ideas.For example, dataset ID = 1 asked its participants, "Who does the channel reach, what is the purpose of the channel, and what information does the channel share?", whereas dataset ID = 4 asked, "What is your supply chain going to be, who are your most important customers, and how will you generate income?".Table 1 reports the distribution of responses across each dataset, with Table 2 reporting some examples of the text responses.By utilising these datasets, each representing unique variations and characteristics, the generality of the proposed approach of expanding topic modelling outputs can be tested, enhancing its reliability and real-world applicability.The data preparation and analysis in this study was conducted using Python (version 3.7.2).For text pre-processing, the traditional NLP techniques were applied, including: • Converting text to lowercase.
• Removing stop words as part of the NLTK package.
• Stemming tokens using the Porter Stemmer as part of the NLTK package.
Table 2 shows examples of the original text responses against how they are represented following pre-processing.

Topic Modelling
There are various methods by which texts can be distributed into topics, with some of the most traditional methods being probabilistic Latent Semantic Analysis (pLSA) [34], Latent Semantic Analysis (LSA) [35], and LDA [36].More recently, however, with the recent advancement in the NLP field, newly developed algorithms, such as BERTopic [37] and Top2Vec [38], are continuing to attract attention.
Such methods vary in complexity, the representation of text segments taken as the model's input (e.g.bag-of-words, word embeddings), their computational speed, and their performance.For example, the LDA algorithm proposes a fixed number of topics in a collection of documents and assumes that each document reflects a combination of those topics.When a collection of documents are analysed under these assumptions, probabilistic inference algorithms reveal an embedded thematic structure, allowing for large collections of documents to be quickly summarised, explored, and searched [39].In general, the LDA algorithm calculates the probability that a word within a document will be included in each topic.A topic may be described by extracting words with the highest probabilities which correspond to such topic.That is, LDA analysis finds the latent topic corresponding to the words in any given document [2].A document is determined to address a topic by calculating a probability distribution over a range of topics for each document [3] and selects the topic with the highest probability as the main topic description.
However, BERTopic and Top2Vec use word embeddings.That is, the vectorisation of text data makes it possible to locate semantically similar words, sentences, or documents within spatial proximity.As word vectors that emerge closest to the document vectors are considered as being the best description of topic of the document, the number of documents that can be grouped together represents the number of topics [40].Whereas BERTopic uses Hierarchical Dirichlet Process (HDP) to cluster vectors into topics, Top2Vec employs a combination of document clustering and word embedding techniques.It uses the Doc2Vec algorithm to generate document embeddings [41].Unlike traditional topic modelling methods, Top2Vec does not require specifying the number of topics in advance.Instead, it identifies topic clusters based on the density of document embeddings and extracts representative keywords and documents for each topic.
Before applying the proposed approach to state-of-the-art models, initial experimentation was conducted using the traditional topic modelling method, LDA.By validating the method with LDA, which is a well-established and widely used technique, it is possible to verify that the approach functions as intended and produced satisfactory results.
For each pre-processed dataset described in Section 3, the topic modelling approach was applied using the 'Latent Dirichlet Allocation' [42] class available as part of the Scikit-learn package.A key hyperparameters of the LDA algorithm is the number of topics [7].To calculate the optimal number of topics across a collection of documents, a good indication is the number of topics with which the model best predicts the data.For topic models such as LDA, a common indicator to measure the optimal number of topics is perplexity, a measure of how well a probabilistic model predicts a sample [36].Yet, recent studies have shown that predictive likelihood (in this case, perplexity) and human judgement are often not correlated [33].In this case, and to retrieve a comparable number of topics, for each dataset, text responses were distributed as one of ten topics.To maintain the comparability across each dataset, the default values of the remaining hyperparameters of the LDA algorithm were set [42].Once topics were identified as a probability distribution of words, each text sample is denoted as a probability distribution over topics.The topic with highest probability was therefore applied to each text segment.In addition, for each topic, the top ten tokens with the highest probabilities were selected as the description of the topic.
Dataset ID = 5 is used as an ongoing example to illustrate the results of the traditional and extended topic modelling outputs discussed in this paper.However, the results are applicable across all datasets.Table 3 reports the token list output produced by the LDA approach.Evidently, based on the content of the textual responses, several words are extracted from topics that may infer its description.For example, blood donation service may be inferred as the description for topic 1, staff mental health may infer topic 4, and email signatures may infer topic 8. Other topic descriptions highlight the possible challenges surrounding the inferences of topic descriptions to human readers as they include inconsistent tokens that may be ambiguous without context.Such examples include topic 3, which includes the words piano, kitchen, bond, etc.

Extending Topic Modelling Outputs
To reiterate, this paper proposes a new method towards extending the outputs of traditional topic modelling methods.
The initial experiments presented herein are tested using the LDA method described in Section 4. The proposed approach includes the following steps: • Mapping the original text responses in each dataset (Table 2) to the most dominant topic expressed in such texts by selecting the topic with the highest relevance score produced by the topic model, in this case, LDA.• For each topic, apply a keyword extraction approach to the original text responses.
• For each topic, map the aforementioned extracted keywords to LDA's output (Section 4) by selecting the keywords with the highest number of intersecting tokens with those produced by the LDA model.Subsequently, of those further refined keywords phrases, the top-scoring ones were assigned as the topic's main description.
Keywords, which are often defined as a sequence of one or more words, provides a short description of the content of a text document [43].Such keywords may be useful entries for building an automatic indexing system for a document collection or can be used to classify text.In the context herein, keywords may serve as a concise label for a given topic.
A popular keyword extraction algorithm is Rapid Automatic Keyword Extraction (RAKE) [43], an unsupervised, domain and language-independent tool for extracting keywords from documents.RAKE identifies stop words and phrase delimiters to split the document into candidate keywords, which are sequences of content words as they occur in the text [44].Firstly, RAKE tokenises the document text by specific word delimiters.Using phrase delimiters and the positions of stop words, tokenised texts are then split into sequences of continuous words.Words within a sequence are assigned the same position in the text and together are considered a candidate keyword [45].Once each candidate keyword is identified, a score is calculated for each and is defined as the sum of its member word scores.Such scores are calculated using one of three methods: the degree of a word in the matrix (i.e. the sum of the number of co-occurrences the word has with any other content word in the text), the word frequency (i.e. the number of times the word appears in the text), or as the degree of the word divided by its frequency.To find specific keywords, RAKE searches for pairs of keywords that adjoin one another at least twice in the same document and in the same order.A new candidate keyword is then created using a combination of the extracted keywords.The score for the new keyword is the sum of its member keyword scores.As RAKE splits candidate keywords by stop words, interior stop words are not present in the extracted keywords.
After applying topic modelling (Section 4), the original text responses in each dataset (see Table 2) were mapped to the most dominant topic expressed in such texts by choosing the topic with the highest relevance score.For example, for dataset = 5, topic 8 was assigned to the text response "Implementation of electronic signatures on contracts and offer letters.This will support with process improvement and digitisation" as a consequence of achieving the highest relevant score of 0.94, whereas topic 4 was assigned to "Mental health is a very important aspect of staff wellbeing, and it affects a lot of people" as a consequence of achieving the highest relevant score of 0.94.For each topic, RAKE was applied to the original text responses and the keywords, as well as their scores, were extracted.
The keywords extracted by RAKE were pre-processed following the techniques discussed in Section 3 and mapped to those extracted as part of the LDA's output.To facilitate this mapping, after applying direct string matching, the keywords with the highest number of intersecting tokens with those produced by the LDA model were extracted.Subsequently, of those further refined keywords phrases, the top-scoring ones were assigned as the topic's main description.Topics with more than one top-scoring keyword were concatenated and separated by the '/' delimiter.
For example, as a consequence of including the highest number of intersecting tokens, i.e. aid, train, mental and health (see Table 3), as well as achieving the highest score produced by the RAKE algorithm (see )), extended outputs may lack minor subject-verb agreement and coherence.However, in comparison to the token list output produced by the LDA topic modelling approach (Table 3), the resulting extended method presents more cohesive and contextualised topic descriptions.Inspired by several studies discussed in Section 2 (e.g.[32], [30], [33]), to gain an insight into whether the proposed method of extending topic modelling outputs increases the interpretability of topic descriptions from a human's perspective, independent annotators were asked to label both topic modelling outputs according to their: • Quality -how easy is it to extract meaning from the text?
• Usefulness -how relevant or helpful is the text in providing information about a topic?
• Efficiency -how efficient is the annotation task?
To facilitate the annotation task, a bespoke web-based annotation platform was implemented.This reduced any installation overhead and widened the reach of annotators as it was accessible using a web browser regardless of the device type (i.e.smartphone, laptops, PC, etc.).Annotators were presented with instructions explaining the task's requirements and then with the platform's interface (see Figure 3.The first pane contained a randomly selected topic modelling output to be annotated, as well as the remaining number of annotations left to complete.For the LDA outputs, to increase interpretability, annotators were presented with non-stemmed tokens.The subsequent panes contained the annotation choices for the aforementioned metrics.For each metric, annotators were required to label each output on a 5-point Likert scale, where a score of 0 signified poor quality, no usefulness, and not an efficient task, and a score of 4 signified high quality descriptions containing clear and effective text that is easy to extract meaning from, are useful in describing a topic, and they are performing the annotation task in the best possible manner with the least amount of effort.
The crowdsourcing of labelling natural language often use a limited number of annotators with the expectation that they are perceived to be experts [46].However, the task of annotating text is considered as being highly subjective and varies with the annotator's age, gender, experience, cultural location, and individual psychological differences [47].For instance, Snow et al. [48] investigate collecting annotations from a wide audience of non-expert annotators over the Web.They show high agreement between the 10 annotations provided by non-experts and those provided by experts.
In this case, to crowdsource annotations in this study and to reach a diverse range of annotators, Twitter was leveraged as a distribution channel.The study was able to attract a broad audience interested in contributing to the annotation process.Annotators were not selected selectively; instead, participation was open to anyone who expressed interest in contributing their insights.To maintain anonymity and avoid duplication of annotations, annotators were distinguished solely by their IP addresses.No personal information was collected from annotators, as explained in the privacy policy, ensuring the confidentiality of their identities.All annotation results were securely stored in a relational database.
A total of 1,600 annotations were collected for all 80 outputs described in Section 4 and 80 extended outputs described in Section 5 with 10 annotations per output.A total of 63 annotators participated in the study.In terms of task efficiency, similar results can be observed.For the extended outputs, a large distribution of 3 and 4 on the Likert scale were used by annotators, denoting that the task of interpreting the extended topic descriptions did not require a significant amount of effort to complete.On the other hand, the LDA outputs report to have a high distribution of 0 on the Likert scale, indicating that annotators required much more effort in defining whether the quality and usefulness of the topic descriptions.
We hypothesise that when an output is unambiguous and helpful in providing information about a topic, then the likelihood of independent annotators selecting higher scores for the metrics increases.This leads to a higher interannotator agreement which indicates that a topic output is more interpretable.Krippendorff's alpha coefficient [49] was used to measure the inter-annotator agreement.As a generalisation of known reliability indices, it was used as it applies to: (1) any number of annotators, not just two, (2) any number of categories, and (3) corrects for chance expected agreement.Krippendorff's alpha coefficient of 1 represents full agreement, 0 represents no agreement beyond chance, and -1 represents disagreement.
Figures 10-12 report the inter-annotator agreement for the quality and usefulness of both the topic modelling outputs across each dataset, as well as the efficiency of the annotation task.For quality, the extended outputs reported the highest agreement of α = 0.522 (dataset ID = 4) and the lowest agreement of α = 0.216 (dataset ID = 8).Likewise, for usefulness, the extended outputs reported the highest agreement of α = 0.575 (dataset ID = 1) and the lowest agreement of α = 0.220 (dataset ID = 5 and 6).In terms of the LDA outputs, for quality, the highest agreement reported was α = 0.075 (dataset ID = 3) and the lowest agreement was α = -0.036(dataset ID = 7).Likewise, for usefulness, the LDA outputs reported the highest agreement of α = 0.090 (dataset ID = 3) and the lowest agreements of α = -0.040(dataset ID = 1).For extended outputs, the relatively high agreement for both quality and usefulness may be explained by the fact that they received a higher distribution of higher scoring annotations.For instance, referring back to the ongoing example of dataset ID = 5 (see Table 2), the extended output Group Email Address received an unanimous agreement of a score of 4 across both quality and usefulness.Whereas due to the ambiguity of the topic description given by the LDA output ("address, request, item, power, group, receive, start, environment, real, differ"), it received the following distribution of annotations: quality (0 = 8 annotations, 1 = 1 annotation, 2 = 1 annotation) and usefulness (0 = 8 annotations, 1 = 2 annotations).
In terms of task efficiency, similar agreements can be found.It was reported, for extended outputs, the highest agreement of α = 0.709 (dataset ID = 1) and the lowest agreement of α = 0.126 (dataset ID = 3).For the LDA outputs, the highest agreement of α = 0.094 (dataset ID = 4) and the lowest agreement of α = -0.012(dataset ID = 1) were reported.Overall, the results demonstrate that the extended outputs are reliably more interpretable to human readers compared to LDA outputs, and the task of annotating their quality and usefulness is particularly less challenging for annotators when observing the extended outputs in comparison to LDA outputs.

Generalisation of the Proposed Method using other Topic Modelling Approaches
The initial experiments conducted with LDA provided a solid foundation, demonstrating the effectiveness of the approach in generating intepretable topic descriptors.Building upon this, to demonstrate the generalisability and robustness of the proposed approach, the methodology is extended to include BERTopic and Top2Vec, two advanced state-of-the-art topic modelling techniques.
Figures 13 and 14 report the extended outputs across each of the topics extracted for all datasets when BERTopic and Top2Vec were used as the underlying topic modelling method, respectively.Both methods were applied without any constraint on the number of topics to be generated.
The implementation of BERTopic and Top2Vec present a broader spectrum of insights.BERTopic, with its deep learning foundation, offered a more nuanced understanding of the text data, generating topic descriptions that were not only coherent but also contextually rich.This method was particularly effective in capturing the subtleties and complexities within the datasets, which LDA might have overlooked due to its probabilistic nature.For instance, in Dataset ID = 4, BERTopic provided a multi-dimensional view of topics, reflecting a deeper layer of thematic understanding.
Figure 13: Extended topic outputs for all datasets using BERTopic as the topic model Top2Vec, on the other hand, presented a unique perspective by generating an extensive array of topics, exceeding 100 in Dataset ID = 4.This method's ability to produce such a wide range of topics highlights its utility in exploring large and diverse datasets, offering a comprehensive landscape of themes and ideas present in the data.However, it's important to note the presence of single-word topic descriptions in some cases.While these may seem less informative at first glance, they play a crucial role in offering a focus on specific themes, as seen with terms like Insurance and Docbot.
The comparative analysis of topic descriptions across LDA, BERTopic, and Top2Vec revealed interesting patterns.The consistency in topic descriptions across different methods, as shown in Dataset ID = 8 with descriptions like Daily 5 Minute Stretch Exercise, underscores the reliability of the approach.This consistency is also an affirmation of the underlying thematic structures within the datasets.Furthermore, the alignment of topic descriptions across various models indicates a convergence of thematic interpretation, despite the inherent differences in the methodologies the topic models.This congruence suggests that this approach is capable of capturing the core essence of topics regardless of the underlying topic modelling technique used.
8 Generalisation of the Proposed Method using New Data The initial experiments conducted herein were with a private dataset provided by our industrial research collaborator.To demonstrate the generalisability of the proposed approach of extending topic model descriptors, the method is applied when using an unseen dataset.
The 20 Newsgroups dataset [50] is commonly used for text mining applications.It is a collection of approximately 20,000 news group documents, distributed (nearly) evenly across 20 different news groups.The documents within the 20 Newsgroups dataset exhibit a range of lengths and cover a spectrum of topics, from religion to politics to automotive and technological discussions, providing a comprehensive basis for testing the robustness of the proposed approach in a more heterogeneous textual environment, enhancing the external validity of this research.Table 5 shows how the 20 news groups are partitioned according to subject matter.
Figures 15-17 report an excerpt of categories from the 20 Newsgroup dataset, where the extended outputs across each of the topics extracted when the LDA, BERTopic, and Top2Vec methods were used as the underlying topic modelling method, respectively.For LDA, the model was requested to distribute texts from each category into one of 10 topics.The BERTopic and Top2Vec were applied without any constraint on the number of topics to be generated.Upon reapplication of the proposed method to the 20 Newsgroups dataset, the extended descriptors illustrate coherent and distinctive topics that align with known categories within the dataset.For instance, when using all three models, it is reported that topics related to 'talk.politics.mideast'are accurately associated with region-specific political discussions (e.g.Jewish National Liberation Movement, Iraqi Death Toll Numbers), while categories like 'soc.religion.christian'and 'rec.autos'have keywords strongly related to religious discourse (e.g.Lord Jesus Christ, Jehovah Thy Redeemer) and automotive subjects (e.g.Changing Brake Fluid, BMW Motorcycles), respectively.These outputs also demonstrate a nuanced understanding of the dataset, capturing the finer subtleties and variations within the broader topics.For example, they distinguish between different facets of religious discussion, separating general Christian talk from more specific debates around biblical interpretation.Similarly, in the automotive category, a separation between general automotive discussions and more technical conversations about car maintenance and mechanical issues may be observed.The proposed approach utilises a safeguard against the scenario where a text document may not contain a keyword that accurately encapsulates its topic.As keyword extraction is employed on a collective set of texts associated with a topic rather than on individual documents, the likelihood of not finding a representative keyword is significantly reduced.The aggregation of texts for each topic increases the probability that relevant keywords will emerge from the collective context, rather than relying on a singular document to provide them.Moreover, the intersectionality between tokens generated by topic modelling methods and extracted keywords further refines the relevance, as only those keywords sharing common tokens with the topic modelling output are considered.This filtration ensures that the final topic descriptions are not only grounded in the statistically derived patterns of the topic model, but are also contextualised by the actual language usage within the texts.However, despite the robustness of this method and its effectiveness in providing topic descriptions for all topics, there remains a possibility that the keywords may not always perfectly match the topic modelling output.In instances where a clear and representative keyword is not identified, or if the extracted keywords do not align well with the modeled topics, such topics can be labeled as 'miscellaneous'.This categorisation serves as an acknowledgment of the method's limitations and provides an opportunity for further exploration and analysis, possibly involving manual review or alternative methods, to better understand and describe these less straightforward topics.
In addition, while the proposed method of expanding topic descriptors in this paper has shown to be effective when using different underlying topic modelling approaches, as well as when it is applied on unseen data, the limitation of this study is the comparison of the proposed approach against relevant baselines discussed in Section 2. As noted in Section 2, other proposed methods are not generalised and are instead tailored to specific datasets and resources, supporting topic extraction from particular genres of text, such as legal documents.These methods also often aim to provide generalised terms for topics based on the specific nature and source of the texts, rather than offering a broad application across various domains.This specialisation limits the comparability of these methods with the more generalised approach proposed in this study.Thus, while a comparison with these methods would be informative, the distinct methodologies and domain-specific applications make direct comparisons complex and perhaps not entirely equitable.

Conclusion
This paper presents an approach towards extending the output of traditional topic modelling methods beyond a list of isolated tokens.The proposed approach removes the dependence on external sources by using the textual data itself by extracting high-scoring keywords and mapping them to the topic modelling method's token outputs.This approach, in turn, increases the interpretability of topic descriptions for human readers by allowing more context and information to be identified, which is an important factor to consider in use cases such as decision making and information retrieval.
To support the experiments presented herein, eight datasets containing short textual responses were used.Once each dataset was pre-processed, the LDA algorithm was applied and set to distribute each text response across one of ten topics.For each topic, the keyword extraction algorithm, RAKE, was applied to the original text responses to extract important keywords and phrases.The token outputs generated by the LDA were subsequently mapped to outputs extracted by RAKE.The keywords with the highest number of intersecting tokens with those produced by the LDA model, as well as those that achieved the highest score from RAKE, were assigned as the topic's main description.
Under the hypothesis that the extended topic modelling outputs are more interpretable than the LDA algorithm's output, such results were expected to be associated with high inter-annotator agreement across interpretability scores.Krippendorff's alpha coefficient was used to measure inter-annotator agreement according to which the extended outputs achieved higher agreement for quality (α = 0.522) and usefulness (α = 0.575).Whereas the LDA outputs achieved lower agreement for quality (α = 0.075) and usefulness (α = 0.090).In terms of the efficiency of the annotation task, the extended outputs achieved higher agreement (α = 0.709) in comparison to the LDA outputs (α = -0.012).Subsequently, the analysis highlights that proposed approach increases the interpretability of the topic model results from a human perspective.To demonstrate and reinforce the generalisation of the proposed method, the approach was further applied using two of the latest state-of-the-art topic modelling methods, BERTopic and Top2Vec, using a heterogeneous unseen dataset.

Future Work
While the current method focuses on making topic descriptors more interpretable to human readers, it is crucial to determine whether these descriptors accurately reflect the underlying content of the topics.To assess this, as part of future work, a combination of qualitative and quantitative methods can be employed.Qualitatively, user studies can be conducted where domain experts and potential users evaluate the relevance and clarity of the topic descriptors.Quantitatively, statistical methods such as coherence measures, which assess the degree of semantic similarity within the topics, can be applied to evaluate the precision of the descriptors.Furthermore, computational experiments comparing the descriptors with unaltered topic outputs could provide quantitative insights into their accuracy and utility.Conducting this further work may offer valuable insights into optimising or fine-tuning the current methodology.This comparative analysis could highlight specific areas where the current approach may be adapted or improved to better suit different text genres or data sources.Additionally, such a study might also pave the way for developing a new, more robust method for expanding topic modeling descriptors.This potentially new method could incorporate the strengths of both generalised and specialised approaches, offering a more versatile solution that can be effectively applied across all domains and text types.By exploring these possibilities, we aim not only to refine our current approach but also to contribute to the broader field of topic modeling, enhancing its applicability and effectiveness in various contexts.

Figure 1 :
Figure 1: An overview of the study design

Figure 2 :
Figure 2: Extended topic outputs for all datasets

Figure 5 :Figure 6 :
Figure 5: Distribution of annotations across the usefulness of extended outputs

Figure 7 :Figure 8 :
Figure 7: Distribution of annotations across the quality of LDA outputs

Figure 9 :
Figure 9: Distribution of annotations across the efficiency of LDA outputs

Figure 14 :
Figure 14: Extended topic outputs for all datasets using Top2Vec as the topic model

Figure 15 :Figure 16 :Figure 17 :
Figure 15: An excerpt of the extended topic outputs for the 20 Newsgroups dataset using LDA as the topic model

Table 1 :
Distribution of responses across datasets

Table 4
), topic 4 in dataset ID = 5 was assigned Aid Training Courses/Mental Health Issues as the topic's descriptor.The remaining keywords may be useful in describing other topics discussed in the text responses (e.g.Mental Health Illness, Aid Training and Training Providers).

Table 4 :
Top scoring keyword phrases for topic 4, dataset ID = 5 Figure 2 reports the extended outputs across each of the ten topic extracted for all datasets.In very few examples (e.g.Free Blood Pressure Testing Carries (Dataset ID = 6) and Dedicated File Preparation Officers Located (Dataset ID =