You are currently viewing a new version of our website. To view the old version click .
Algorithms
  • Article
  • Open Access

28 November 2023

An Efficient Optimized DenseNet Model for Aspect-Based Multi-Label Classification

,
,
,
and
1
Department of Creative Technologies, Air University Islamabad, Islamabad 44000, Pakistan
2
Department of Computing, Riphah International University, Faisalabad 38000, Pakistan
3
School of Digital Science, Universiti Brunei Darussalam, Jalan Tungku Link, Gadong BE1410, Brunei
4
Department of Information and Communication Technology, University of Agder (UiA), N-4898 Grimstad, Norway
This article belongs to the Special Issue Machine Learning in Big Data Modeling

Abstract

Sentiment analysis holds great importance within the domain of natural language processing as it examines both the expressed and underlying emotions conveyed through review content. Furthermore, researchers have discovered that relying solely on the overall sentiment derived from the textual content is inadequate. Consequently, sentiment analysis was developed to extract nuanced expressions from textual information. One of the challenges in this field is effectively extracting emotional elements using multi-label data that covers various aspects. This article presents a novel approach called the Ensemble of DenseNet based on Aquila Optimizer (EDAO). EDAO is specifically designed to enhance the precision and diversity of multi-label learners. Unlike traditional multi-label methods, EDAO strongly emphasizes improving model diversity and accuracy in multi-label scenarios. To evaluate the effectiveness of our approach, we conducted experiments on seven distinct datasets, including emotions, hotels, movies, proteins, automobiles, medical, news, and birds. Our initial strategy involves establishing a preprocessing mechanism to obtain precise and refined data. Subsequently, we used the Vader tool with Bag of Words (BoW) for feature extraction. In the third stage, we created word associations using the word2vec method. The improved data were also used to train and test the DenseNet model, which was fine-tuned using the Aquila Optimizer (AO). On the news, emotion, auto, bird, movie, hotel, protein, and medical datasets, utilizing the aspect-based multi-labeling technique, we achieved accuracy rates of 95%, 97%, and 96%, respectively, with DenseNet-AO. Our proposed model demonstrates that EDAO outperforms other standard methods across various multi-label datasets with different dimensions. The implemented strategy has been rigorously validated through experimental results, showcasing its effectiveness compared to existing benchmark approaches.

1. Introduction

Sentiment analysis, also known as opinion mining, has become an increasingly important topic of discussion in various fields. It is crucial in evaluating customer feedback, understanding public opinions, and tracking real-world activities []. Analyzing sentiment in textual data allows us to extract valuable insights and to make informed decisions based on the emotions expressed. Several existing sentiment analysis techniques exist, including machine learning-based strategies and lexicon-based approaches [,,]. Machine Learning (ML) methods leverage algorithms and models to automatically classify sentiment, while lexicon-based techniques rely on predefined sentiment dictionaries. These methodologies have been widely used in sentiment classification tasks, offering different advantages and trade-offs depending on the application. In sentiment analysis, classification typically involves binary methods for sentiment polarity (positive or negative) or multi-class methods for categorizing emotions into multiple classes. This distinction sets sentiment analysis apart from other techniques, as it requires specialized approaches to handle the complexity of emotions expressed in a text. This paper proposes a multi-label labeling strategy to categorize emotions, encompassing a diverse range of emotional aspects of the human mind.
The impact of sentiment analysis is evident in various domains. For instance, in one study [], a multi-task, multi-label classification approach was proposed to integrate attitude categorization using subject tweets and sentiment analysis. Although the sentiment analysis task was framed as a multi-label problem, where multiple emotions could be associated with a single text, the study utilized data that allowed for a wide range of emotional reactions. Similarly, another article [] classified emotions by categorizing news articles using labels and phrases obtained from a newspaper repository. The potential application of aspect-based multi-label techniques to classify textual data according to mood remains an open question.
An exponential growth in potential label combinations presents a major hurdle in multi-label sentiment analysis. The objective of conventional multi-label learner (MLL) approaches is to enhance the precision and effectiveness of each multi-label learner through the utilization of label associations []. These methods can be categorized based on their approach to program development. Some techniques focus on testing various MLLs, assuming that the MLL can predict all labels and incorporate label correlations into the learning process [,]. Other methods, such as EPS and RAKEL, rely on ensembles of single learners [,]. This proposed work employs an ensemble learning approach, creating a group of learners capable of predicting a single label. These foundational learners are combined to create a multi-label classifier capable of predicting among all labels. The starting point for these learners is often the utilization of label associations with a single label. Figure 1 provides a visualization of the multi-labeling process.
Figure 1. Virtualization of multi-labeling [].
Sentiment analysis, a rapidly evolving field, has gained significant attention due to its relevance in various domains. Traditional approaches in MLL techniques have focused on developing individual MLLs, which may limit the generalization ability of a single learner []. To overcome this limitation, ensemble learning methods have surfaced, intending to enhance the overall capability of the learning system. This approach mitigates the likelihood of overfitting and fosters cross-learner creation within a single-label environment []. Ensemble learning involves building a cluster of basic MLLs that can collectively predict all connected labels, thereby significantly enhancing the generalization potential of the training approach []. This group of learners is referred to as an MLL ensemble []. The primary goal of multi-label-based ensemble learning is to develop a network of learners that is both accurate and diverse []. It is crucial because the base learners’ collective abstraction capability and diversity are closely related to the ensemble’s generalization flaw. Unlike conventional MLL ensemble methods that combine multiple base pair trainers into a single MLL, multi-label-driven ensemble methods utilize MLL as the foundation of the learning process.
Despite their significance, multi-label supervised methods have received little attention in this context. Creating a collection of precise and diverse base solutions suitable for multi-label situations presents several challenges. One of the key research obstacles in this field involves the conventional ensemble learning methods, which mainly address challenges related to single-label learning. However, in multi-label scenarios, the most difficult task lies in assessing the reliability of the multi-label base learners, as it demands thorough comprehension of the interrelationships between different types of labels. Differentiating between diverse label learning within a multi-label context poses challenges since the behavior of the base learners depends on the entire set of labels rather than a single term [].
This paper presents a unique methodology that leverages an evolving ensembler [] and employs multi-label categorization on textual data, thereby attaining consistent enhancements in precision. The proposed method addresses the challenges previously encountered in multi-labeling text data. By incorporating a genetic algorithm with a support vector machine, improvements in various performance measures, including precision, have been observed. The suggested research demonstrates that merging DenseNet with the Aquila Optimizer (AO), i.e., the proposed method, yields promising results by significantly increasing quality performance while maintaining reasonable cost []. To further validate the study, comparisons are made between the newly developed model and five other classifiers, including stable discriminant analysis, decision tree, Convolutional Neural Network (CNN), maximum entropy, random forest, and generalized linear model. These comparisons confirm the superiority of the suggested methodology over all benchmark techniques []. The contributions are listed below in alphabetical order.
  • Introducing a novel approach that utilizes aspect-based multi-label classification for accurately classifying emotions in textual data, enabling more granular sentiment analysis.
  • Comparative Analysis of Sentiment Classification Techniques: Conducting an extensive analysis to compare the effectiveness of different sentiment classification techniques in categorizing emotions, providing insights into their suitability for multi-label sentiment analysis tasks.
  • Evaluation of State-of-the-Art Algorithms: Assessing the performance of five advanced multi-label classification algorithms on diverse emotion-based textual datasets, offering a comprehensive understanding of their strengths and weaknesses in handling multi-label sentiment analysis.
  • Introduction of Ensembler: Presenting the approach called Ensemble of DenseNet based on AO (EDAO) technique, which enhances the accuracy and variation of multi-label sentiment analysis models by integrating accuracy and diversity, offering a new perspective on multi-label instruction approaches.
  • Development of Comprehensive Workflow: Establishing a comprehensive workflow that incorporates preprocessing, feature extraction, and model tuning techniques, elevating the performance of sentiment analysis models and ensuring the use of refined and precise data.
  • Experimental Validation and Performance Comparison: Extensive experiments are being conducted to validate the effectiveness of the proposed EDAO approach. These experiments involve comparing the approach with existing benchmark methods and showcasing its superior ability to capture sentiment variations and handle complex multi-label datasets.
These contributions significantly advance the field of multi-label sentiment analysis by introducing innovative approaches, conducting thorough evaluations, and providing valuable insights into sentiment classification model performance.
The following is a breakdown of the manuscript’s structure. Section 2 contains relevant work on MLL by other scholars in the area. The specifics of our conceptual framework for aspect-based multi-labeling, which utilizes a DenseNet-based ensembler, are described in Section 3. Section 4 describes the general framework. Section 4 of the study thoroughly presents the experimental results collected. Section 5 concludes with a summary of the findings and recommendations for further research.

3. Proposed Methodology

In the proposed work, the following stages are conducted: data collection, preprocessing, aspect/feature extraction, representation of word2vec, and implementation of the swarm ensembler, as depicted in Figure 2.
Figure 2. Proposed System Model for multi-labeling.
Data gathering begins with site scraping and online repository access, which yielded the dataset utilized in this paper. Another crucial step is the analysis of the data. It is imperative to refine the acquired data thoroughly. Secondly, the raw, unclean data undergo a preprocessing stage, during which it is refined and formatted in preparation for subsequent accurate processing. Thirdly, we extract and compare the features (aspects) of the enhanced data. We were working with aspect-based labels and multidimensional data, which can be challenging and prone to various interpretations. Fourthly, we employed the word2vec tool to generate data vectors, serving as a tool for data representation. Before the multi-labeling application, the data undergo training utilizing the suggested method known as EDAO. Subsequently, an analysis is conducted. The following sections of this article provide a detailed elucidation.

3.1. Dataset Collection

The evaluations are gathered using a web data extraction approach, and specifically collected data are taken through the Kaggle website. To achieve nominal scaling, we propose utilizing a novel data format and evaluating correlations between labels. To assess EDAO’s effectiveness, we carefully selected eight substantial datasets for text labeling from the Kaggle collection [,,]. To determine the diversity of the proposed SEn, benchmark datasets from various application domains are being utilized to evaluate the proposed EDAO. According to these datasets, texts are categorized into several categories, including emotions, movies, medical, and hotels.
Table 3 provides an overview of the characteristics of multi-label datasets, each containing 1000 instances. These datasets are designed to represent various protein types and include six labels, each corresponding to different protein classes.
Table 3. Characteristics of datasets.
To clarify the relevance of the dataset section, it is important to note that these datasets were carefully chosen due to their direct applicability to sentiment analysis and their representation of diverse subject areas. Including datasets spanning different domains is instrumental in evaluating the generalizability and robustness of the proposed EDAO algorithm. For instance, the hotel dataset focuses on textual data from the hospitality industry, allowing the algorithm to comprehend sentiments expressed in hotel reviews. Conversely, the medical dataset deals with biological texts, facilitating sentiment analysis within medical literature and healthcare-related content. The movie dataset comprises textual data from the film industry, enabling the algorithm to gauge sentiments expressed in movie reviews and discussions. On the other hand, while also textual, the protein dataset delves into protein classification, presenting unique sentiment analysis challenges. Similarly, the automobile dataset provides a collection of text related to automobiles, facilitating sentiment analysis in the context of car reviews or vehicle-related discussions. As previously mentioned, the emotion dataset serves as a benchmark dataset, covering a wide spectrum of emotions and allowing the algorithm to detect and analyze sentiments across various emotional states. Incorporating datasets like the bird and news ones further broadens the evaluation scope, encompassing additional domains where sentiment analysis is valuable. These datasets empower the algorithm to comprehend sentiments expressed in texts related to ornithology or news articles. By encompassing datasets from diverse application domains, the evaluation of the proposed EDAO algorithm becomes more comprehensive and meaningful. This approach ensures that the algorithm’s performance is scrutinized across a wide range of textual data, leading to a more reliable and robust assessment of its effectiveness in sentiment analysis.

3.2. Preprocessing of Data

Data preprocessing is a method for converting data collected into a convenient format. The raw data are converted into structural data at this step. There may be blank lines and blank columns for data obtained from scrapping. The data are then examined using the Panda library, which is only utilized to save pertinent data.
Comprehensive preprocessing is essential for multi-labeling as it directly impacts the success rate of a project. The presence of missing attributes or anomalies, outliers, null values, or incorrect information in the data renders it impure. If any of the variables mentioned above occur, the trustworthiness of the findings would be undermined. In this work, we explore the complex problem of recognizing and classifying different features or dimensions in textual data. This particular type of multi-labeling goes beyond a general categorization strategy and attempts to give several labels to various textual elements. We use a number of methods to do this, concentrating especially on tokenization and lemmatization.
Tokenization: To implement aspect-based multi-labeling, detailed comprehension of the text is necessary, broken down into smaller components called tokens. Tokens encompass not only single words but also crucial phrases or components that aid in pinpointing particular details inside the text. Tokenization is the first stage of deconstructing the text, making it easier to analyze each component in more detail.
Lemmatizing: It is essential for standardizing word representation since aspect-based multi-labeling is subtle. We establish a consistent framework for comprehending various kinds of expression connected to each feature by breaking words down into their basic or lemma forms. This guarantees that differences in word forms do not impede the proper labeling of particular characteristics.

3.3. Feature Extraction

Words are converted into functional vectors at the network’s first level to provide semanticized and structural word data. Vocabulary Word consists of Vwords. The lexicon of a character in Vcharr is reflected. Every sentence containing n words supplies the vectorization of wn (each word) (w1, w2, w3, … wn).
V n = r w o r d , r w c h a r
The variable rwchar is for the character level, whereas the variable rword is for the implantation of the word level. The acquisition of semantic and syntactic information typically involves words, while character-level insertion aims to capture morphological and structural details.

3.4. Word2Vec Representation

Word2vec, a potent NLP technique, utilizes vector functions to decode the complexities of language. With a rich text corpus, it employs NNs to gain profound linguistic insights. It autonomously extracts location characteristics even from brief text snippets. This versatile tool excels with short text or single words. Leveraging extensive text data and word2vec’s capabilities enables efficient word generation and seamless navigation of vast datasets. Deep learning primarily focuses on words and their meanings, making word2vec a valuable asset for tackling word meaning-related challenges. Word documents are transformed into vectors for training. Within word2vec’s intricate framework lie three million words intricately connected. Exploring relationships between terms becomes effortless in this expansive linguistic landscape.

3.5. Swarm-Based Ensembler

The swarm-based ensembler that is currently available in the literature is discussed in this section.

3.5.1. DenseNet Classification

Through this research, we aim to develop a classification system for categorizing aspects based on multi-labeling from a dataset. A robust CNN classifier is meticulously constructed to classify the labels within a dataset and to extract meaningful features from the data. This endeavor necessitates the utilization of two formidable deep-learning libraries: Torchvision and PyTorch. Torchvision, a pre-trained data learning model, empowers the process by endowing maximum control against overfitting while simultaneously enhancing the optimization of outcomes right from the outset. A five-layer dense block is defined as a consequence of the computation carried out by the network using the DenseNet block architecture illustrated in Figure 3 and a growth rate of L = 5. It is vital to recognize that the number 121 in DenseNet-121 denotes that the NN has 121 layers. A typical DenseNet-121 composition consists of several different layers that are combined. A total of five convolution and pooling layers are used, as well as three transition layers (6, 12, 24, and 48), two DenseBlocks (11 and 33) convolutions, and one classification layer (16). DenseNet supports feature reuse, reduces the number of parameters, and enhances the model’s performance for multi-label classification and classification accuracy.
Figure 3. DenseNet model.

3.5.2. DenseNet Mechanism

The distinguishing essence of DenseNet lies in its foundational concept of feed-forward networks, bestowing upon it a prominent edge over other network architectures. DenseNet offers a plethora of enthralling advantages. First and foremost, DenseNet helps to mitigate the vanishing-gradient issue. Furthermore, DenseNet exhibits its prowess by intensifying the propagation of features, fostering their regeneration, and effectively reducing the parameter count. It is achieved through the unique mechanism of a dense layer, which elegantly consolidates the outputs from preceding layers by concatenating them along the depth dimension. Figure 4 depicts the function of a DenseNet process, which combines transition layers and dense blocks to categorize an input review depending on its content. When a text is supplied as input to the DenseNet, the text is processed through many dense blocks. Each layer’s feature maps remain the same from layer to layer except for the number of filters, which change from layer to layer within a single dense block. Upon traversing a dense block, data moving within DenseNet progress to the next stage, the transition layer.
Figure 4. Flow of DenseNet.
Within the architectural framework of DenseNet, the transition layer assumes the crucial responsibility of executing two fundamental operations: convolution and pooling. These operations are diligently performed in the downsampling procedures strategically positioned outside the dense blocks. To ensure the harmonization of feature map lengths within the dense block before concatenation, a convolution layer can be incorporated before each convolution. In the realm of the DenseNet architecture, the transition layer encompasses diverse components, including batch normalization (BN), as well as other essential elements like average pooling layers and convolution layers, all working together to optimize computational efficiency and to minimize the number of input feature maps.
Figure 4 depicts the DenseNet flow chart. The DenseNet design’s dense block comprises a layer of BN, ReLU activation, and convolution, all accompanied by convolution. After the last dense block, an average global pooling layer is performed, and the results are sent to a Softmax classifier.

3.5.3. Multi-Objective Optimization Solution

After evaluating the initial classifier for consistency and complexity, the next step involves optimization, a multi-objective problem-solving process similar to standard optimization in conventional machine learning. This optimization process aims to convert optimal labeling into unique objective efficiency using the weight-sum method. However, it is important to note that this approach is susceptible to weight shifting due to the trade-off between the ensemble’s generalization ability and the two objectives. To address this challenge, we have implemented the AO [], a swarm-based multi-objective optimization method, which systematically explores optimal trade-offs within a swarm population, effectively managing multi-objective reduction complexities without compromising generalization. Our paper introduces the AO as a novel technique inspired by the natural hunting behavior of Aquilas. The AO algorithm employs four strategies: soaring high for exploration, gliding and attacking to explore diverging search spaces, gradually descending to exploit converging areas, and diving by foot to capture insights. These diverse approaches synergize to unlock the algorithm’s optimization potential, as shown in Figure 5.
Figure 5. AO’s working flow.
EDAO excels in improving both accuracy and variation in sentiment analysis, especially in multi-label datasets where multiple sentiments need predicting.
It achieves this by generating diverse base learners, each skilled in capturing different sentiments, enhancing its ability to understand nuanced variations in sentiment across texts. The results consistently demonstrate EDAO’s superiority over benchmark schemes and individual learning methods, with higher precision, recall, F1-score, and accuracy across various datasets. EDAO’s success is attributed to its utilization of DenseNet-AO, a specialized variant designed to target diversity-related objectives, setting it apart from methods that may not explicitly consider diversity. Furthermore, EDAO offers a sensitivity analysis, quantifying the uncertainty and variability in its decision-making process, highlighting its reliability and consistency. The AO’s is used to find the optimal parameters for the DenseNet algorithm, as shown in Algorithm 1.
Table 4 succinctly outlines EDAO’s unique features that distinguish it from other multi-label learning techniques and the benefits they bring. EDAO stands out for its integration of accuracy and diversity in the optimization process, enhancing both precision and variation in sentiment analysis. Notably, it employs DenseNet-AO, a specialized variant, to target diversity-related objectives, setting it apart. The table underscores EDAO’s superior performance compared to benchmarks and highlights its use of sensitivity analysis for assessing reliability. The pseudocode of the proposed model is provided in Algorithm 2.
Algorithm 1 Aquila Optimizer for DenseNet parameters.
1:
Input: Population
2:
Output: BestParametersFound
3:
Initialize Population
4:
EvaluateFitness(Population)
5:
generation ← 0
6:
while not TerminationCriteriaMet do
7:
      selectedParents ← SelectParents(Population)
8:
      offspring ← CrossoverAndMutate(selectedParents)
9:
      EvaluateFitness(offspring)
10:
      nextGeneration ← SurvivalSelection(Population, offspring)
11:
      UpdateParameters(nextGeneration)
12:
      generation ← generation + 1
13:
end while
14:
return BestParametersFound
1:
function EvaluateFitness(Individual)
2:
      Calculate objectiveFunctionValue using TrainAndEvaluateDenseNet(Individual)
3:
      Calculate fitness ← 1 / (1 + objectiveFunctionValue)
4:
      Update Individual.fitness and Individual.objectiveValue
5:
end function
1:
function TrainAndEvaluateDenseNet(Parameters)
2:
                                        ▹ Train DenseNet and evaluate on validation set
3:
                                         ▹ Return objective function value (e.g., validation loss)
4:
end function
1:
function SelectParents(Population)
2:
                                   ▹ Use a selection mechanism to choose individuals for reproduction
3:
end function
1:
function CrossoverAndMutate(Parents)
2:
                             ▹ Apply crossover and mutation to generate offspring from selected parents
3:
                                               ▹ Return new population of offspring
4:
end function
1:
function SurvivalSelection(Population, Offspring)
2:
                                           ▹ Combine current population and offspring
3:
                                          ▹ Select individuals to survive based on fitness
4:
                                           ▹ Return next generation of individuals
5:
end function
1:
function UpdateParameters(NextGeneration)
2:
                                    ▹ Update parameters of individuals in the population
3:
end function
Table 4. Key aspects and benefits of EDAO in multi-label classification and sentiment analysis.
Algorithm 2 Proposed model algorithm (pseudocode).
Require: Data (Input data for sentiment analysis)
Ensure: SentimentResults (Output Classification results)
1:
procedure CollectData                                ▹ Implementation details for collecting data
2:
end procedure
3:
procedure PreprocessData(Data)
4:
      PerformPOSTagging(Data)
5:
      LemmatizeData(Data)
6:
      PerformPOSTagging(Data)
7:
end procedure
8:
procedure ExtractFeatures(Data)                         ▹ Implementation details for feature extraction
9:
end procedure
10:
procedure GenerateWordVectors(Features)                    ▹ Implementation details for word vector generation
11:
end procedure
12:
procedure TrainModel(WordVectors)
13:
      InitializeModel
14:
      FineTuneModel(WordVectors)
15:
      OptimizeModel(AquilaOptimizer)
16:
end procedure
17:
procedure EvaluateModel
18:
      CalculateAccuracy
19:
      CalculateRecall
20:
      CalculateFscore
21:
      PerformSensitivityAnalysis
22:
      CalculatePrecision
23:
      MeasureComputationalComplexity
24:
end procedure
25:
SentimentResults ← AnalyzeData(Data)

4. Simulation Results and Discussion

This section provides an extensive description of the experimental results along with an in-depth analysis of the outcomes. A system with specification a ninth-generation Core i7 processor with 2.4 quad-cores was used for the research. The main programming language used for this study is Python, and the programming environment used was the Spyder IDE.

4.1. Dataset Description

Based on the findings reported in Table 5, our proposed ensembler underwent comprehensive testing across diverse real-world multi-label classification datasets spanning various domains. Among these datasets, the Medicine [] dataset is the first in biology. Its primary objective is to estimate the prevalence of different diseases within the population. The results of this study were obtained from the Kaggle website. The second dataset, called hotel reviews, contains the pertinent hotel booking records from Kaggle []. The other dataset comes from various diseases. The results of this study were obtained from the Kaggle website. The second dataset, called hotel reviews, contains the pertinent hotel booking records from Kaggle []. The other dataset comes from Kaggle [], which aims to identify the specific subject categories to which each document belongs. For the purpose of classification, we have partitioned the dataset into training and testing sets using an 80/20 split. This means that 80% of the data are allocated for training the model, while the remaining 20% are reserved for evaluating the model’s performance.
Table 5. Dataset details.

4.2. Metrics for Performance Evaluation

This section will apply five essential multi-label metrics, as previously introduced and discussed in our work. These metrics have gained significant attention in the current literature, making them fundamental for our analysis. To illustrate, consider the scenario within our multi-label datasets, which consist of multiple instances represented as (ci, pi), where “ci” represents a specific label and “pi” is the associated input. We will use the notation h(pi) for the predictive label collection. Additionally, we will introduce ci, f to denote the individual MLL h function and ci, f for the separate MLL h function, each corresponding to a specific (ci, pi) pair. Regarding the variable “ci”. This variable reflects how the learner rates input “pi” at label “ci”, ensuring that "ci” accurately corresponds to the label assigned to input “pi”. This alignment holds significant importance in our analysis.

4.2.1. Accuracy

The ratio of accurately anticipated observations to all observed data points (or observations plus predictions) is the most easily understood performance metric (or observations). High-precision models are the most accurate because they are the most precise (high accuracy). Precision is essential, but only an asymmetric dataset will have nearly equal-sized negative and false positive results.

4.2.2. Precision

It establishes the percentage of cases correctly diagnosed out of all those that received favorable labels under specific conditions. The ratio of all correctly predicted positive observations to the amount of all foreseen positive sightings plus the number of significant positive observations is known as precision. A low rate of false positives in medical diagnosis is associated with good diagnostic accuracy.

4.2.3. Recall

Recall, also known as the true positive rate or sensitivity, is a crucial metric in multi-labeling. It measures the proportion of correctly predicted positive observations within the context of our multi-label dataset. It is calculated by dividing the total number of observations by the correctly predicted positive observations. This metric is significant, alongside accuracy and other key metrics, for assessing the performance of multi-label classification models.

4.2.4. F-Measure

The F-measure combines precision and recall, providing a balanced evaluation of multi-label classification performance. This metric, also known as the balanced F-score, represents the harmonic mean of precision and recall. The formula for computing it involves the square of the scaling factor divided by the weighted sum. It is essential to note that precision and recall are equally weighted in this metric, commonly referred to as the F1-measure.

4.3. Compared Models

The suggested EDAO was put through its paces in a full comparison with the most commonly used MLL techniques to demonstrate its efficacy. In the ensemble multi-label learning (EnML) domain, two distinct scenarios, ML-NCL and ML-HISC, stand out. These scenarios optimize a single target to validate the provided objective function. By focusing on individual targets, they serve as valuable testbeds for refining and validating the optimization process within the EnML framework. All of these techniques are summarized in the following sections.
  • Proposed Ensembler: DenseNet is used as a DNN algorithm ensemble with AO and as an ensembler. The DenseNet makes use of its bunch of layers.
  • NB: The ML algorithm was utilized in the multi-labeling method.
  • CNN: The multi-label learning (MLL) algorithm is fundamentally rooted in the domain of NNs.
  • ML-RBF []: Additionally, the RBF NN algorithm-based MLL algorithm is the primary trainer utilized by the ENL system and is developed based on RBF NN.
  • RAKEL []: Another approach to multivalued learning. In this approach, a basic learner with a single label makes judgments based on a randomly selected subset of the encountered labels. This limited subset serves as the basis for decision-making, reflecting the nature of most basic learners in this context.
  • ECC [] provides a detailed description of an ensemble approach for MLL based on using classifier chains. To transform EnML into a subproblem, we must first modify the unique goal into the one we used to create EnML.
M L HSIC C = 1 / log ( M L HSIC ) M L NCL L = 1 M L N C L / ( m × L )
Enhanced DenseNet with Aquila Optimizer (EDAO) is a revolutionary advancement in sentiment analysis that has unique characteristics that distinguish it from other approaches. The Aquila Optimizer, a customized optimization method created especially for adjusting DenseNet’s parameters, is one of its main contributions. This new optimizer, which prioritizes effectiveness and efficiency, is essential to improving the convergence and general performance of the model. Moreover, EDAO overcomes the drawbacks of other sentiment analysis techniques, excelling especially in multi-label sentiment analysis. Through its optimization process, EDAO combines accuracy and diversity, in contrast to benchmark techniques that could just favor accuracy. This integrated approach is beneficial, allowing EDAO to perform better than approaches with a more limited accuracy-focused scope and to catch sentiment fluctuations with ease.
More specifically, to identify minute emotion differences across various texts, EDAO carefully use DenseNet-AO, a DenseNet variation. By leveraging DenseNet-AO, this approach greatly improves the model’s capacity to traverse the intricacies of multi-label datasets, hence addressing a prevalent sentiment analysis difficulty. DenseNet-AO integration is also essential for increasing generalization efficiency, which enables the model to easily adjust to new or unknown input. A sensitivity analysis is a crucial assessment indicator for the decision-making process in terms of robustness and stability, according to EDAO. This novel approach solves the issues with the robustness of current techniques and offers a thorough evaluation of the model’s dependability in a range of scenarios. Furthermore, EDAO simplifies label space translation, reducing the computational burden involved in this procedure. Using the LP approach for label space translation improves the model’s efficiency and scalability, enabling it to handle large datasets with ease. Essentially, Enhanced DenseNet with Aquila Optimizer is a ground-breaking sentiment analysis system that combines cutting-edge optimization techniques, DenseNet-AO’s strength, and an all-encompassing strategy to overcome the drawbacks of conventional approaches. Together, these unique qualities improve EDAO’s performance, let it capture sentiment nuances more effectively, and present it as a reliable multi-label sentiment analysis solution.

4.4. Performance of Different Methods

For each experimental data collection, tenfold cross-validation is conducted. Using the provided outputs, the selected technique’s standard deviation and average productivity are computed for the entire dataset. The experiments were carried out utilizing a device powered by a robust 2.24 GHz Intel Xeon processor and bolstered by a generous 16 GB of RAM.
Table 6, Table 7 and Table 8 show the productivity of six different approaches. Based on all metrics and datasets, the proposed model beats the other methods evaluated, including the ML-RBF process and two combination approaches, RAKEL and ECC. The technique described remains constant across all datasets. It has been demonstrated that, compared to non-ensemble methods, our ensembler EDAO can effectively improve the generalization efficiency of DNN.
Table 6. Benchmark scheme vs EDAO (precision).
Table 7. Recall comparison of benchmark scheme vs. SEn.
Table 8. Benchmark scheme vs. EDAO (F1-score).
EDAO’s advantages over other multi-component learning techniques, such as ECC and NB, highlight our concept’s effectiveness. By harnessing the potential of integrated learners in conjunction with machine learning, MLL undergoes a significant improvement in generalization. Within this proposed setting, DenseNet excels, driven by its unmatched ability to address diversity-related objectives that have long challenged the field of multi-label research. Through the ingenious use of the AO algorithm, DenseNet transcends conventional boundaries, leading to a movement of diversification across the MLL domain. This distinctive attribute is a fundamental pillar behind DenseNet’s remarkable ascent. In our proposed approach, we examine the impact of objective functions on outcomes by comparing EDAO with CNN and Naive Bayes, representing two types of objective functions. The technique consistently yields the best results across all datasets, as evidenced in Table 6, Table 7 and Table 8. Visualizations of the model loss and suggested model accuracy are presented in Figure 6 and Figure 7. In our conceptual model, we observe a compelling trend: As the number of epochs increases, our model’s accuracy demonstrates significant improvement while the loss steadily decreases. This pattern reinforces the effectiveness of our approach, highlighting the positive correlation between training epochs and enhanced model performance.
Figure 6. Proposed model accuracy.
Figure 7. Proposed model loss.
Figure 8, Figure 9 and Figure 10 showcase the features of specific datasets. Performance measures such as F1-score, accuracy, precision, and recall are employed to evaluate productivity across diverse datasets. Figure 11 presents an illustrative example of the findings derived from the news dataset. Regarding accuracy, the suggested model EDAO exhibits superior performance compared to other contrasting approaches. Specifically, CNN-CHIO demonstrates higher accuracy on the news dataset than ML-RBF and NB. Notably, NB yields the lowest accuracy ratings among the five models.
Figure 8. Frequency-based features of the emotion dataset.
Figure 9. Frequency-based features of the medical dataset.
Figure 10. Frequency-based features of the news dataset.
Figure 11. Classifier accuracy on the news dataset.
The proposed ensembler achieves a higher overall score on the movie dataset, as seen in Figure 12. Although it is a multi-labeling dataset, even the most complex method performs well. The proposed ensembler outperforms other techniques in terms of precision, f-measurement, accuracy, and recall, as depicted in Figure 13 and Figure 14. The great accuracy implies that the proposed ensembler would produce unexpectedly strong results with multi-labeling data, which is a promising sign. ML-RBF decreases as the running output increases in Figure 11, Figure 12, Figure 13 and Figure 14. From Figure 11, Figure 12, Figure 13 and Figure 14, we can conclude that EDAO is improving. The divergent trends of both aims indicate that they are inherently at odds. It should not come as a surprise. Predicted baseline learner labels intersect with the labels correctly due to the ML-RBF optimization. As a result, all of these fundamental learners appear to be the same.
Figure 12. Classifier accuracy on the movie dataset.
Figure 13. Classifier accuracy on the emotion dataset.
Figure 14. Accuracy of classifiers on medical dataset.
In contrast, ECC maximization enables base learners to exhibit maximum flexibility in terms of training error. In most situations, the stated goals are opposed to one another. The conflict allows ML-RBF to use population optimization to balance the two goals. It is crucial to note that in this situation, NB and ECC assess the basic learners’ diversity and correctness, not the ensemble’s overall performance. The decrease in ML-RBF does not suggest that there has been a decline in the ensemble. Despite this, the increase in EDAO implies that base learners are becoming more varied, improving group performance. An optimization technique is used to fine-tune the tuning parameters of the ensembler. The precision of multi-labeling is improved as a result of these discoveries. As a result, EDAO consistently enhances a multi-label ensemble’s ability to generalize, enhancing multi-label categorization competence.
The sensitivity analysis (SA) is a critical aspect of our study as it introduces the element of uncertainty into our evaluation of both benchmark schemes and our proposed algorithm. SA involves calculating the extent of change required for each uncertain element before altering the initial decision. This analysis provides valuable insights into the robustness and stability of our proposed approach.
In Figure 15, we present the results of the sensitivity analysis specific to our recommended technique, DenseNet-AO. This figure showcases the best estimates for each variable within our approach, highlighting how they deviate from the baseline estimate. By visualizing these variations, we gain a deeper understanding of how our proposed method responds to uncertain component changes, further strengthening our approach’s reliability and effectiveness. The SA in Figure 15 serves as a vital component of our study, allowing us to assess the impact of uncertainties on our proposed method’s performance and providing valuable insights for optimizing our approach.
Figure 15. Proposed ensembler sensitivity analysis.
Furthermore, Figure 16 depicts the SA of the most recent benchmark methods. Through SA comparisons, it is evident that the EDAO variables exhibit the least variation among the baseline-related variables, surpassing benchmark algorithms.
Figure 16. CNN sensitivity analysis.
Figure 17, Figure 18, Figure 19 and Figure 20 show the confusion matrix of DenseNet-AO and other existing models, which demonstrates the best performance among the models.
Figure 17. Confusion matrix of DenseNet-AO.
Figure 18. Confusion matrix of NB.
Figure 19. Confusion matrix of BERT.
Figure 20. Confusion matrix of CNN.
Table 9 determines the limitations the EDAO approach addresses, highlighting its significant advancements in multi-label sentiment analysis. EDAO offers distinct advantages over existing methods through its integration of accuracy and diversity, the utilization of DenseNet-AO, improved generalization efficiency, robust decision-making, and efficient label space transformation. These features contribute to EDAO’s effectiveness in capturing sentiment variations, enhancing performance, and handling complex multi-label datasets.
Table 9. Comparative analysis of limitations and solutions in multi-label sentiment analysis with EDAO approach.
Table 10 compares execution times in seconds for various techniques used in different domains. These values represent approximate execution times based on real-world scenarios. The DenseNet-AO technique stands out with notably lower execution times across all domains, ranging from 20 to 35. It indicates its efficient computational complexity, making it a promising choice for tasks in the news, emotion, medical, bird, hotel, automobile, movie, and protein domains.
Table 10. Comparison of execution times (in seconds) for various techniques.
On the other hand, the RAKEL technique demonstrates varying execution times, with some domains taking longer than others. For example, it requires around 120 s for emotions and automobiles. The ECC technique also exhibits variability in execution times, with the medical domain having the longest execution time of 180. The ML-RBF and CNN techniques generally show execution times in the 70 s to 120 s range, while the NB technique requires approximately 180s for the medical domain. BERT, LSTM, Transformer, GAT, ResNet, SVM, Logistic Regression, and random forest techniques also display varying execution times across domains from the 70 s to 200 s range.
Table 11 presents the Pearson correlation coefficients between various techniques and domains. The coefficients serve as gauges that measure the intricate connection between two variables, spanning a scale from −1 to 1. A coefficient that gravitates towards 1 signifies a robust positive correlation, while a coefficient approaching −1 indicates a compelling negative correlation. Delving into the tabulated data, we discover that DenseNet-AO consistently showcases remarkable positive correlations across various domains. Its values gracefully range from 0.86 to 0.92, portraying unwavering strength and cohesion. It indicates a strong linear relationship between DenseNet-AO and the respective domains, implying that DenseNet-AO performs consistently well across different data types.
Table 11. Pearson correlation coefficients (statistical analysis).
Other techniques such as RAKEL, ECC, ML-RBF, CNN, BERT, LSTM, Transformer, and ResNet also exhibit positive correlations ranging from 0.70 to 0.85, indicating a moderate to strong linear relationship with the domains. These techniques show promising performance in various domains. Conversely, NB reveals negative correlations from −0.32 to −0.35 across most domains. It implies a weak negative linear relationship between NB and the domains, suggesting that NB may not be well-suited for these datasets.
Table 12 presents the results obtained from a comprehensive analysis of variance (ANOVA) conducted on the execution times of various techniques across multiple domains. ANOVA is a robust statistical tool used to determine whether significant differences exist among the means of different groups. Central to this research is the F-value, which indicates the ratio between the observed variability across groups and the variability observed within groups. A higher F-value signifies a more pronounced distinction between the means of these groups. Equally crucial is the p-value, which plays a pivotal role in assessing the statistical significance of the F-value, offering insights into the probability of obtaining such results purely by chance.
Table 12. One-way ANOVA (statistical analysis).
Upon a thorough examination of the table, a notable finding emerges. DenseNet-AO stands out prominently, boasting a substantial F-value of 15.24 and an impressively low p-value of 0.001. It indicates that the execution times of DenseNet-AO exhibit significant variations across the diverse domains considered in our analysis. Similarly, techniques such as CNN, Transformer, and random forest display noteworthy F-values ranging from 5.99 to 7.92, coupled with relatively low p-values ranging from 0.006 to 0.012. It suggests these techniques demonstrate substantial differences in execution times across the analyzed domains. However, it is worth noting that certain techniques, including NB, GAT, SVM, and Logistic Regression, exhibit non-significant F-values and relatively higher p-values. This implies that the execution times of these techniques do not vary significantly across the diverse domains considered in our study.

5. Conclusions

A new ensembler based on an optimization approach is created to handle the problem of multi-labeling. SEn, an integrated multi-objective optimization approach, dynamically optimizes two objective functions to measure the accuracy and variety of multi-label learners simultaneously. By generating a diverse collection of exact multi-label base learners, EDAO enhances the prediction capabilities of the system. The primary goal of this project is to improve the efficiency of the MLL system by incorporating accurate and diverse ML-based learners. To tackle the challenges of multi-labeling, we have developed a novel form of ensembler that leverages optimal algorithms. EDAO combines a swarm multi-objective optimization technique with ML to optimize the two objective functions dynamically, quantifying the accuracy and variety of learners and generating accurate and diverse predictions. Extensive studies have demonstrated that EDAO significantly enhances the generalization capabilities of multi-label classification systems while improving the model’s prediction performance.

Author Contributions

Conceptualization, N.A.; Methodology: N.A. and T.; Formal analysis and investigation, S.H. and J.I.; Writing—original draft preparation, N.A.; Writing—review and editing, S.S.U. and S.H.; Funding acquisition, S.S.U.; Resources: N.A.; Supervision: S.H. and S.S.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The data can be found here: [,,].

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AOAquila Optimizer
BRBinary Relevance
CCClassifier Chain
CNNConvolutional NN
EDAOEnsemble of DenseNet based on Aquila Optimizer
LBLabel Combination
MLMachine Learning
MLLMulti-Label Learner
NN Neural Networks
SA Sensitivity Analysis
SVM Support Vector Machine

References

  1. Singh, P. Money Laundering and Abuse of the Financial System. Indian JL & Legal Rsch. 2023, 5, 1. [Google Scholar]
  2. Wu, C.; Xiong, Q.; Yi, H.; Yu, Y.; Zhu, Q.; Gao, M.; Chen, J. Multiple-element joint detection for aspect-based sentiment analysis. Knowl.-Based Syst. 2021, 223, 107073. [Google Scholar] [CrossRef]
  3. Sun, J.; Lang, J.; Fujita, H.; Li, H. Imbalanced enterprise credit evaluation with DTE-SBD: Decision tree ensemble based on SMOTE and bagging with differentiated sampling rates. Inf. Sci. 2018, 425, 76–91. [Google Scholar] [CrossRef]
  4. Piri, S.; Delen, D.; Liu, T. A synthetic informative minority over-sampling (SIMO) algorithm leveraging support vector machine to enhance learning from imbalanced datasets. Decis. Support Syst. 2018, 106, 15–29. [Google Scholar] [CrossRef]
  5. Zhang, C.; Tan, K.C.; Li, H.; Hong, G.S. A cost-sensitive deep belief network for imbalanced classification. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 109–122. [Google Scholar] [CrossRef]
  6. Collell, G.; Prelec, D.; Patil, K.R. A simple plug-in bagging ensemble based on threshold-moving for classifying binary and multiclass imbalanced data. Neurocomputing 2018, 275, 330–340. [Google Scholar] [CrossRef] [PubMed]
  7. Yu, H.; Sun, D.; Xi, X.; Yang, X.; Zheng, S.; Wang, Q. Fuzzy one-class extreme autoencoder. Neural Process. Lett. 2019, 50, 701–727. [Google Scholar] [CrossRef]
  8. Ben, X.; Zhang, P.; Lai, Z.; Yan, R.; Zhai, X.; Meng, W. A general tensor representation framework for cross-view gait recognition. Pattern Recogn. 2019, 90, 87–98. [Google Scholar] [CrossRef]
  9. Ben, X.; Gong, C.; Zhang, P.; Jia, X.; Wu, Q.; Meng, W. Coupled Patch Alignment for Matching Cross-view Gaits. IEEE Trans. Image Process. 2019, 28, 3142–3157. [Google Scholar] [CrossRef]
  10. Wang, S.; Yang, L.Y. Adaptive bi-weighting toward automatic initialization and model selection for HMM-based hybrid meta-clustering ensembles. IEEE Trans. Cybern. 2019, 49, 1657–1668. [Google Scholar]
  11. Moyano, J.M.; Gibaja, E.L.; Cios, K.J.; Ventura, S. Review of ensembles of multi-label classifiers: Models, experimental study, and prospects. Inform. Fusion 2018, 44, 33–45. [Google Scholar] [CrossRef]
  12. García-Pablos, A.; Cuadros, M.; Rigau, G. W2VLDA: Almost unsupervised system for aspect-based sentiment analysis. Expert Syst. Appl. 2018, 91, 127–137. [Google Scholar] [CrossRef]
  13. Kumar, V.; Pujari, A.K.; Padmanabhan, V.; Kagita, V.R. Group preserving label embedding for multi-label classification. Pattern Recognit. 2019, 90, 23–34. [Google Scholar] [CrossRef]
  14. Deng, M.; Wang, C.; Tang, M.; Zheng, T. Extracting cardiac dynamics within ECG signal for human identification and cardiovascular diseases classification. Neural Netw. 2018, 100, 70–83. [Google Scholar] [CrossRef] [PubMed]
  15. Szymánski, P.; Kajdanowicz, T. Scikit-multilearn: A scikit-based Python environment for performing multi-label classification. J. Mach. Learn. Res. 2019, 20, 209–230. [Google Scholar]
  16. Cevikalp, H.; Benligiray, B.; Gerek, O.N. Semi-supervised robust deep neural networks for multi-label image classification. Pattern Recognit. 2020, 100, 107164. [Google Scholar] [CrossRef]
  17. Charte, F.; Rivera, A.J.; Charte, D.; del Jesus, M.J.; Herrera, F. Tips, guidelines and tools for managing multi-label datasets: The mldr.datasets R package and the cometa data repository. Neurocomputing 2018, 289, 68–85. [Google Scholar] [CrossRef]
  18. Ning, F.; Delhomme, D.; LeCun, Y.; Piano, F.; Bottou, L.; Barbano, P.E. Toward automatic phenotyping of developing embryos from videos. IEEE Trans. Image Process. 2005, 14, 1360–1371. [Google Scholar] [CrossRef]
  19. Mohamed, A.; Dahl, G.E.; Hinton, G. Acoustic modeling using deep belief networks. IEEE Trans. Audio Speech Lang. Process. 2012, 20, 14–22. [Google Scholar] [CrossRef]
  20. Banbhrani, S.K.; Xu, B.; Soomro, P.D.; Jain, D.K.; Lin, H. TDO-Spider Taylor ChOA: An Optimized Deep-Learning-Based Sentiment Classification and Review Rating Prediction. Appl. Sci. 2022, 12, 10292. [Google Scholar] [CrossRef]
  21. Liao, S.; Wang, J.; Yu, R.; Sato, K.; Cheng, Z. CNN for situations understanding based on sentiment analysis of Twitter data. Procedia Comput. Sci. 2016, 111, 376–381. [Google Scholar] [CrossRef]
  22. Severyn, A.; Moschitti, A. Twitter sentiment analysis with deep convolutional neural networks. In Proceedings of the SIGIR ’15: The 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, Santiago, Chile, 9–13 August 2015; pp. 959–962. [Google Scholar]
  23. Ouyang, X.; Zhou, P.; Li, C.H.; Liu, L. Sentiment analysis using convolutional neural network. In Proceedings of the 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing, Liverpool, UK, 26–28 October 2015; pp. 2359–2364. [Google Scholar]
  24. Liu, S.M.; Chen, J.H. A multi-label classification-based approach for sentiment classification. Expert Syst. Appl. 2015, 42, 1083–1093. [Google Scholar] [CrossRef]
  25. Montañes, E.; Senge, R.; Barranquero, J.; Quevedo, J.R.; del Coz, J.J.; Hüllermeier, E. Dependent binary relevance models for multi-label classification. Pattern Recognit. 2014, 47, 1494–1508. [Google Scholar] [CrossRef]
  26. Wu, G.; Zheng, R.; Tian, Y.; Liu, D. Joint Ranking SVM and Binary Relevance with robust Low-rank learning for multi-label classification. Neural Netw. 2020, 122, 24–39. [Google Scholar] [CrossRef] [PubMed]
  27. Lin, X.; Chen, X.W. Mr.KNN: Soft relevance for multi-label classification. In Proceedings of the 19th ACM International Conference on Information and Knowledge Management, Virtual, 19–23 October 2010; pp. 349–358. [Google Scholar]
  28. Wu, G.; Tian, Y.; Zhang, C. A unified framework implementing linear binary relevance for multi-label learning. Neurocomputing 2018, 289, 86–100. [Google Scholar] [CrossRef]
  29. Yan, Y.; Wang, Y.; Gao, W.C.; Zhang, B.W.; Yang, C.; Yin, X.C. LSTM: Multi-Label Ranking for Document Classification. Neural Process. Lett. 2018, 47, 117–138. [Google Scholar] [CrossRef]
  30. Azarbonyad, H.; Dehghani, M.; Marx, M.; Kamps, J. Learning to rank for multi-label text classification: Combining different sources of Information. Nat. Lang. Eng. 2020, 27, 89–111. [Google Scholar] [CrossRef]
  31. Nguyen, T.T.; Dang, M.T.; Luong, A.V.; Liew, A.W.-C.; Liang, T.; McCall, J. Multi-label classification via incremental clustering on an evolving data stream. Pattern Recognit. 2019, 95, 96–113. [Google Scholar] [CrossRef]
  32. Nguyen, T.T.T.; Nguyen, T.T.; Luong, A.V.; Nguyen, Q.V.H.; Liew, A.W.-C.; Stantic, B. Multi-label classification via label correlation and first order feature dependence in a data stream. Pattern Recognit. 2019, 90, 35–51. [Google Scholar] [CrossRef]
  33. Reyes, O.; Ventura, S. Evolutionary strategy to perform batch-mode active learning on multi-label data. ACM Trans. Intell. Syst. Tech. 2018, 6, 46:1–46:26. [Google Scholar] [CrossRef]
  34. Wang, R.; Wang, X.-Z.; Kwong, S.; Xu, C. Incorporating diversity and informativeness in multiple-instance active learning. IEEE Trans. Fuzzy Syst. 2018, 25, 1460–1475. [Google Scholar] [CrossRef]
  35. Wang, X.-Z.; Wang, R.; Xu, C. Discovering the relationship between generalization and uncertainty by incorporating complexity of classification. IEEE Trans. Cybern. 2018, 48, 703–715. [Google Scholar] [CrossRef]
  36. Wei, X.; Yu, Z.; Zhang, C.; Hu, Q. Ensemble of label specific features for multi-label classification. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23–27 July 2018. [Google Scholar]
  37. Yapp, E.K.; Li, X.; Lu, W.F.; Tan, P.S. Comparison of base classifiers for MLL. Neurocomputing 2020, 394, 51–60. [Google Scholar] [CrossRef]
  38. Zhang, M.L.; Li, Y.K.; Liu, X.Y.; Geng, X. Binary relevance for MLL: An overview. Front. Comput. Sci. 2018, 12, 191–202. [Google Scholar] [CrossRef]
  39. Zhou, J.-P.; Chen, L.; Guo, Z.-H. iATC-NRAKEL: An efficient multi-label classifier for recognizing anatomical therapeutic chemical classes of drugs. Bioinformatics 2020, 36, 1391–1396. [Google Scholar] [CrossRef]
  40. Xie, M.-K.; Huang, S.-J. Partial MLL. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018; pp. 4302–4309. [Google Scholar]
  41. Zhou, Z.-H. A brief introduction to weakly supervised learning. Nat. Sci. Rev. 2018, 5, 44–53. [Google Scholar] [CrossRef]
  42. Al-Smadi, M.; Qawasmeh, O.; Al-Ayyoub, M.; Jararweh, Y.; Gupta, B. Deep recurrent neural network vs. support vector machine for aspect-based sentiment analysis of Arabic hotels’ reviews. J. Comput. Sci. 2018, 27, 386–393. [Google Scholar] [CrossRef]
  43. Ghosh, S.; Ekbal, A.; Bhattacharyya, P. A multitask framework to detect depression, sentiment and multi-label emotion from suicide notes. Cogn. Comput. 2022, 14, 110–129. [Google Scholar] [CrossRef]
  44. Al-Smadi, M.; Al-Ayyoub, M.; Jararweh, Y.; Qawasmeh, O. Enhancing aspect-based sentiment analysis of Arabic hotels’ reviews using morphological, syntactic and semantic features. Inf. Process. Manag. 2019, 56, 308–319. [Google Scholar] [CrossRef]
  45. Da’u, A.; Salim, N.; Rabiu, I.; Osman, A. Recommendation system exploiting aspect-based opinion mining with deep learning method. Inf. Sci. 2020, 512, 1279–1292. [Google Scholar]
  46. Kumar, J.A.; Trueman, T.E.; Cambria, E. Gender-based multi-aspect sentiment detection using multilabel learning. Inf. Sci. 2022 606, 453–468. [CrossRef]
  47. Fu, X.; Wei, Y.; Xu, F.; Wang, T.; Lu, Y.; Li, J.; Huang, J.Z. Semi-supervised aspect-level sentiment classification model based on variational autoencoder. Knowl.-Based Syst. 2019, 171, 81–92. [Google Scholar] [CrossRef]
  48. Gu, X.; Gu, Y.; Wu, H. Cascaded convolutional neural networks for aspect-based opinion summary. Neural Process. Lett. 2017, 46, 581–594. [Google Scholar] [CrossRef]
  49. Gargiulo, F.; Silvestri, S.; Ciampi, M.; De Pietro, G. Deep neural network for hierarchical extreme multi-label text classification. Appl. Soft Comput. 2019, 79, 125–138. [Google Scholar] [CrossRef]
  50. Dahou, A.; Ewees, A.A.; Hashim, F.A.; Al-Qaness, M.A.; Orabi, D.A.; Soliman, E.M.; Abd Elaziz, M. Optimizing fake news detection for Arabic context: A multitask learning approach with transformers and an enhanced Nutcracker Optimization Algorithm. Knowl.-Based Syst. 2023, 280, 111023. [Google Scholar] [CrossRef]
  51. MultiLabel Dataset. Available online: http://www.uco.es/kdis/mllresources/ (accessed on 15 June 2023).
  52. MultiLabel Dataset. Available online: www.booking.com/hotel-reviews (accessed on 16 June 2023).
  53. MultiLabel Dataset. Available online: https://sci2s.ugr.es/keel/multilabel.php (accessed on 17 June 2023).
  54. Abualigah, L.; Yousri, D.; Abd Elaziz, M.; Ewees, A.A.; Al-Qaness, M.A.; Gandomi, A.H. Aquila optimizer: A novel meta-heuristic optimization algorithm. IEEE Access 2021, 9, 107250. [Google Scholar] [CrossRef]
  55. Charte, F. A comprehensive and didactic review on multi-label learning software tools. IEEE Access 2020, 8, 50330–50354. [Google Scholar] [CrossRef]
  56. Mai, L.; Le, B. Aspect-based sentiment analysis of Vietnamese texts with deep learning. In Proceedings of the Asian Conference on Intelligent Information and Database Systems, Dong Hoi City, Vietnam, 19–21 March 2018; Springer: Cham, Switzerland, 2018; pp. 149–158. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.