Next Article in Journal
Improving Imbalanced Land Cover Classification with K-Means SMOTE: Detecting and Oversampling Distinctive Minority Spectral Signatures
Next Article in Special Issue
Cluster Appearance Glyphs: A Methodology for Illustrating High-Dimensional Data Patterns in 2-D Data Layouts
Previous Article in Journal
Combine-Net: An Improved Filter Pruning Algorithm
Previous Article in Special Issue
Quantitative and Qualitative Comparison of 2D and 3D Projection Techniques for High-Dimensional Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visual Active Learning for Labeling: A Case for Soundscape Ecology Data

by
Liz Huancapaza Hilasaca
1,
Milton Cezar Ribeiro
2 and
Rosane Minghim
3,*
1
Department of Computer Science, University of São Paulo, São Paulo, SP 13566590, Brazil
2
Biodiversity Department, São Paulo State University—UNESP, São Paulo, SP 13506900, Brazil
3
School of Computer Science and Information Technology, University College Cork, T12 XF62 Cork, Ireland
*
Author to whom correspondence should be addressed.
Information 2021, 12(7), 265; https://doi.org/10.3390/info12070265
Submission received: 1 June 2021 / Revised: 18 June 2021 / Accepted: 22 June 2021 / Published: 29 June 2021
(This article belongs to the Special Issue Trends and Opportunities in Visualization and Visual Analytics)

Abstract

:
Labeling of samples is a recurrent and time-consuming task in data analysis and machine learning and yet generally overlooked in terms of visual analytics approaches to improve the process. As the number of tailored applications of learning models increases, it is crucial that more effective approaches to labeling are developed. In this paper, we report the development of a methodology and a framework to support labeling, with an application case as background. The methodology performs visual active learning and label propagation with 2D embeddings as layouts to achieve faster and interactive labeling of samples. The framework is realized through SoundscapeX, a tool to support labeling in soundscape ecology data. We have applied the framework to a set of audio recordings collected for a Long Term Ecological Research Project in the Cantareira-Mantiqueira Corridor (LTER CCM), localized in the transition between northeastern São Paulo state and southern Minas Gerais state in Brazil. We employed a pre-label data set of groups of animals to test the efficacy of the approach. The results showed the best accuracy at 94.58 % in the prediction of labeling for birds and insects; and 91.09 % for the prediction of the sound event as frogs and insects.

1. Introduction

Active Learning (AL) is considered as a special case of machine learning [1]. AL is also called “query learning” because it actively requests information of selected data from the set of unlabeled data, from which the model will learn. It is widely used in hard cases for big data learning, an example of which is evidenced by [2] when data have only 2000 labeled instances and 250,000 unlabeled instances. There are different strategies used for each stage AL, such as those described and categorized by [1,3,4]. The effectiveness of AL was demonstrated in various tasks, such as: (a) automatic speech recognition [5]; (b) classification of voicemail messages [6]; (c) malicious code detection [7]; (d) text classification [8]; (e) speech emotion classification [9]; and (f) audio retrieval [10]. When the user takes part in the process, there is a cooperation between computer and analyst, and it becomes part of the Human In The Loop (HITL) machine learning paradigm [11], targeted at improving learning processes by employing the user’s expertise as well as computing strategies.
The main goal of AL is that the algorithm learns using the smallest possible number of instances during training, but generating the best predictions of unlabeled data [1]. There are several sampling strategies, but at least three categories can be generalized according to [1]: (1) certainty-based sampling, (2) query-by-committee, and (3) expected error reduction. In the first sampling strategy, a small set of selected samples is annotated at the beginning, and then manually annotated labels are used to train a classifier, which classifies unlabeled samples. The second type of sampling strategy involves two or more classifiers, who may disagree with respect to some instances; if they agree, those instances are delivered for human annotation or validation. The third type of strategy aims to estimate and select the instances that can have a high impact on the expected model error for human annotation; however, this last strategy may be the most computationally expensive.
The final goal of Active Learning is to learn from a small set of labeled samples. However, a large number of current applications require a large subset of labelled data from the start. Examples are the building of certain machine learning models, such as those applied in deep learning, and applications targeted at selecting and reducing the set of attributes to represent a phenomenon (examples are certain biological data sets, medical records, fraud detection, etc.). Our aim in this work is to find a stable balance of the active learning and label propagation strategies to support the actual labeling process, so that the stage of the data science pipeline can be made more effective.
In order to allow user participation in the labeling processes, visualization tools are necessary, since they can help the user navigate in large data with multiple attributes, by facilitating data interpretation through graphical and interactive representations. Multidimensional projections or embeddings allow the reduction of a multidimensional space of m dimensions to another reduced space with p dimensions ( X m X p , p < m ) , while trying to preserve as much information as possible. The AL approach together with visualization techniques can facilitate the acquisition of knowledge from the data and support interpretation as well as sample labeling by the user [12,13].
In this study, we propose a method of user centered active learning that aims to optimize and give more dynamism to the entire process of labeling, by including visual strategies in the several stages of AL. In support of the strategy, we have compared different types of sampling strategies and estimated the appropriate number of samples that must be labeled by the annotator. Thus, we also performed an exhaustive assessment of learning power and performance of the models, in relation to the different sampling strategies and the defined number of clusters. We have also reflected on how the model learns based on a dataset with numerous features against and the same dataset taking as a starting point known discriminative features (as illustrated in Figure 4). After data samples are annotated by expert users, a classification model is trained with the recorded data. Finally, the predictions of the labels for the data set are visualized and evaluated (see Table 1). The projections afford visual analysis and interaction in each stage of the active learning process to give the user a better understanding of the data and of the process of labeling. (see Figure 5). Users can confirm or correct predicted labels. Therefore, this is a visual analytics approach that is meant to blend learning and user supervision in the process of labeling samples in general and in soundscape ecology data in particular (see Figures 3 and 5).
We realize the framework and its methods in the context of environment monitoring using sound recordings. Recordings are becoming central in understanding and describing the condition of natural environments. They are involved in a large number of studies, such as monitoring environmental noise [14], measuring biodiversity integrity and environmental health [15,16,17], freshwater lakes [18,19], and species identification [20]. Soundscape ecology studies [21] have increased in the recent years [22] and, between other things, this research field aims to monitor and understand how different environments respond to changes induced by human activities [23,24,25] and to assess the impact through altered soundscapes.
Extracting information from such data is both very challenging and expensive. Therefore, it is necessary build a bridge between ecoacoustics, machine learning, and visualization, which requires a multidisciplinary approach [26]. One important phase of soundscape ecology studies is the annotation or tagging of events of interest [27]. Therefore, developing new methods that improve the quality of labeling is essential for the success of many soundscape ecology studies worldwide. Here, we test our framework on the task of labeling soundscape ecology data and employing real data on birds, frogs, and insects to evaluate the performance of the method.
The main contributions of this work are summarized as follows:
  • Proposal, implementation, and testing of a Visual Active Learning strategy for labeling and application of such strategy to soundscape ecology data;
  • Anchoring the process of user centered labeling through visualization by multidimensional projections. Anchoring data detailing in visualizations of summarized data; in the case of our sample application that is done by the proposal of “Time Line Spectrogram” (TLS) visualizations (Figure 3).
  • New sampling strategies incorporated in the process of Active Learning for labeling, and their evaluation (Figure 4) using this strategy, we demonstrate reduced annotation costs (Table 1);
  • Presentation of extensive experimental results that demonstrate the relevance of the proposed methods (Figures 3–5 and Table 1).

2. Background

2.1. Visualization in Active Learning and Labeling

Visualization in support of active learning strategies has been observed for some time in a variety of applications such as image processing [28,29]. Years later, the work [30] presented Visalix as a good alternative that also combines AL and visualization, where the user can make the annotation of classes by managing the attributes in a 3D space, but with certain limitations. Reference [31] also contributed in the field by presenting the Case Base Topology Viewer for Active Learning (CBTV-AL), where they considered density, uncertainty, and diversity in the sampling strategies. In the case of diversity and uncertainty, the CBTV-AL needs to be recalculated until the last instance can be labeled. CBTV-AL allows for visualizing the entire labeling process through layouts based on force-directed graph drawing algorithms. The authors emphasize the importance of visualization techniques for user participation in AL. Reference [12] enhances sample selection in AL based on visualization using scatter plots and iso-contours, employing the semi-supervised metric learning method to train with data annotated by the user. Reference [32] presents a pilot study using the t-Distributed Stochastic Neighbor Embedding (t-SNE) [33], force-directed graph layout, and chord diagrams to visualize data and facilitate labeling. According to the authors, this improves the text document labeling process. In these last two approaches, the determination of samples is carried out manually on the visualization (scatter plot) by the user, and this leads to the conclusion that there is no sample selection and AL suggestion strategy directly involved as described by [13]. In addition, there are other strategies as incrementally involved within active learning models as can be seen in the approaches of [34,35,36].
Here, we present an approach to predict labels using label propagation by sample selection over a clustering process. The visualizations in our case are based on multidimensional projections, which aim to map data in 2D based on their similarity, giving an extra layer of effectiveness of the current stage of the labeling. In comparison to previous methods, we contribute both in terms of the interaction method and on the sample selection process.
To test the method, we built a framework for labeling of acoustic landscapes in soundcape ecology. We have employed a data set for animal group identification whose discriminant features we studied before [37] to evaluate the impact on results. Our results prove efficiency in the prediction task of labels and show reduced manual annotation effort with the methodology proposed for soundscape data.

2.2. Labeling of Sound Data

Sound data labeling is a key task that, in general, precedes most of the remaining data analysis tasks or the development of new approaches to automatic interpretation. For labeling sound data, there are some computer-based solutions, such as the ones presented recently by [27]. Manual labeling is very costly in time and degree of complexity, and studies in AL for labeling are meant to minimize manual annotation effort of samples by the users. In line with this objective, the study conducted by [38] addresses a methodology based on the combination of AL and self-training by considering the level of confidence of instances. Thus, instances with low confidence scores are delivered to expert users to be labeled and instances with high confidence scores are used in the prediction automatically. Because it is based on defining the scores, the determination of the confidence threshold is crucial. Reference [38] used the FindSounds (FindSounds: https://www.findsounds.com/, accessed on 22 June 2021) database with duration ranging from 1 to 10 s for each instance of audio. Reference [39] proposed a new medoids-based active learning (MAL) method for generating clusters by K-Medoids; afterwards, the medoids from each cluster are presented to expert users and labeled by them. After obtaining the labels, instances are fed to a classifie. Reference [39] used the UrbanSound8k [40] database with maximum duration of four seconds for each instance of audio. Another study conducted by [41] proposed an AL strategy that combines two stages: (i) The first stage implements the same strategy as [39]; (ii) In the second stage, they proposed a selection of samples based on the prediction mismatch, looking mainly for segments with incorrect labels; labels are then corrected and predictions are updated in the groupings using a nearest-neighbor approach. Reference [42] presents an AL-based framework for classifying soundscape recordings. According to the methodology, the authors first generate 60 clusters, then randomly selected 10 instances of each cluster to be manually labeled. It allowed to define the most appropriate class names for each cluster. The reduced manual annotation effort with the active learning methodology in the paper was demonstrated empirically. Reference [42] used a database with maximum duration of 1 min for each instance of audio. As presented above, several venues can be pursued by researchers when advancing the labeling task, with a varying degree of automation and confidence. On the other handl, Reference [43] proposed an active learning system for detecting sound events where the main objective is to improve the process of selection of samples based on the identification of audio segments with the presence of sound activity for annotation; and, in the same line of work, reference [44] proposed a new strategy to determine samples from the audio database, thus asserting the high influence of this sample selection task in reducing the manual labeling effort by the user.
Our approach focuses on displaying the data as submitted to several strategies or techniques for each stage of active learning process, with the main goal of labeling data with improved accuracy, but with less effort by users than with manual labeling. Here, we propose a visual analytics approach that is meant to blend learning and user supervision in the process of labeling samples in general and in soundscape ecology in particular. The goals are: (i) include the learning and supervision of the user in the model for labeling; (ii) know what the appropriate strategies or techniques are in this labeling process; and (iii) support evaluating the quality of labeling to soundscape data.

3. The Labeling Method

In Figure 1, we illustrate the main steps of our proposed method for labeling strategy. Given a data set that has no labels, the Clustering stage is initially undertaken. Then, samples are extracted from each cluster at the stage named Sampling. Expert users interact with the samples by listening and labeling the audios in the Annotation step. Considering the samples labeled by the user, a learning model (classifier) is trained. Then, the model is then used to predict the rest of the labels of the instances of data set (excluding the samples), and these predictions are performed in the step learning–prediction. The visualization task is intrinsically present in all stages. Finally, the results obtained with the proposed method are evaluated. The steps of each stage of our method are given in the Algorithm 1, and each step is described in the next subsections. We implemented our method as a framework with application in soundscape ecology. A description of the interface of the system is shown in Appendix A.
Algorithm 1: Graph building.
Information 12 00265 i001

3.1. Clustering

To start this stage, a data set of unlabeled instances for which features have been extracted is required. In the case of soundscape data, numerical features are computed from the audio itself and from the image of the audio spectrogram. In general, the input to clustering is only the features extracted for all instances. Firstly, k clusters are extracted from the data employing the Euclidean distance. Hierarchical Agglomerative Clustering and K-Means were used in our framework, both using the Scikit-learn package (available in scikit-learn: https://scikit-learn.org/, accessed on 22 June 2021). We consider that, through clustering, it will be possible to identify patterns between the instances of audio data, and, consequently, these patterns will allow the segregation of the event categories from the soundscape and produce good samples for labeling.

3.2. Sampling

The main goal of the sampling step is to extract additional representative samples from the data set, to be used later in learning tasks. Thus, p samples are extracted from each of the clusters, employing for this the following sample extraction methods: random (r), medoid (m) and contour (c); as well as their combinations: (random-medoid (rm), random-contour (rc), medoid-contour (mc) and random-medid-contour (rmc)). In the first method, samples are taken randomly. In the case of medoid, samples are the instances closest to the cluster centroid. The method contour takes samples furthest from the centroid of the clusters. Figure 2 illustrates the three types of sampling methods.

3.3. Annotation

In this step, the induction of learning from the AL paradigm is initiated, through the interaction of expert users. The goal of this step is to abstract information from users who are experts in recognizing and differentiating sound categories. Thus, this stage deals with the tasks of listening and labeling audio files corresponding to the most representative samples. To perform these tasks, multidimensional projections are employed to allow visualization and interaction with the samples. In a parallel view, the same projection facilitates the visualization of previously generated clustering between instances.
In order to assist interaction with projections, visualizations from data summarization should be presented on the top banner of the interface. In our implementation as a new proposal in visual AL, we deal with combined spectrograms over time. They are denominated Time-Line-Spectrogram (TLS)—see the top of Figure 3. The goal of the TLS visualization is to provide more visual information about samples while the user interacts with the projections, in the tasks of labeling by listening. TLS is a supportive visual representation that summarizes the spectrograms of each audio recording (instance) according to the time of recording in the landscape. Other types of data, such as documents, image collections, and videos, can also be summarized by tag clouds or representative pictures.

3.4. Learning-Prediction

This step aims to train a learning model from features and labels of samples, and this learning is used to predict the labels of the other instances to the data set. To accomplish this step, the user needs to perform the following tasks:
(i)
Learning: Model training. In this case, the model to be used is Random Forest Classifier (RFC).
(ii)
Prediction: After the learning, labels of the instances other than the samples are predicted. Then, by examining the results using the same visualizations, and the criteria of the application, the steps of the proposed method can be repeated starting from the Clustering step (Section 3.1).

3.5. Validation

To validate the proposed method, it is necessary to use a data set that has true labels, that is to say, that all instances of the audios had been previously labeled by expert users. Therefore, the validation in this stage of the method consists of comparing the real labels with predicted labels. In order to validate the prediction, we use the classification Accuracy (AC), which is defined by Equation (1):
A C = T P + T N T P + F P + F N + T N
where TP are the true positives, TN are the true negatives, FP are the false positives, and FN are the false negatives.

3.6. Visualization

Visualization techniques—particularly multidimensional projections—are used in all stages of the proposed method (Clustering, Sampling, Annotation, and Learning), to support verification of results and user interaction where feasible.
By employing projections, users have the same mental model of all steps, and, by interpreting similarity between samples and their neighbors on the projections, the user is equipped with a powerful tool for browsing through data. This should increase the performance of expert users tasks.

4. Data Description and Case Study

To validate our proposed method, we used a data set provided in partnership with the Spatial Ecology and Conservation Lab (LEEC) of São Paulo State University (UNESP—Rio Claro). The soundscape recordings are part of Long-Term Ecological Research within the Ecological Corridor of Cantareira-Mantiqueira (LTER CCM or PELD CCM in Portuguese). The audio data were recorded within 22 landscapes distributed in the LTER CCM region, where the following types of environments were sampled: forest, swamps, and open area (mainly pasture). Originally, the region had been covered by forest, but, due to the expansion of agriculture, pasture, and urbanization, the region shows varying forest cover from 16 % to 85 % in different portions of study areas. The recordings occurred between October 2016 and January 2017, and each landscape was surveyed during three consecutive months (30 days for forest, 30 days for swamps, and 30 days for open areas). Half of the recordings were collected in the morning (from sunrise to 8:30 a.m.) and half in the evening (6:30 p.m. to 10:00 p.m.). For the purpose of the current study, the raw data were re-sampled, in order to represent the soundscape heterogeneity of all the sampled environments. Therefore, the re-sampled data set has more than 40,000 sound files of one minute each. A total of 2277 sound files were labeled by experts. To assess our method, we chose to work with the data set containing those 2277 instances of audio. The sound files were labeled according to the most dominant sound in each minute, which were divided into three labels: 615 for frogs, 822 for birds, and 840 for the insects. Hereafter, we will refer to this data set as DS1. In the same region of LTER CCM, two other soundscape ecology studies were conducted: (1) [22], which aimed to assess how spatial scale (i.e., extents) influences acoustic indices responses and how these indices behave according to natural vegetation cover (%); and the authors in (2) [37] developed a method to identify the most discriminant features for categorizing sound events in soundscapes. This present study and its realization in the field of soundscape ecology represents an important tool for further ecological studies in this and other natural areas.
For data analysis, we follow the methodology described in article [37], which first suggests a preprocessing of the audios with a set of parameters to generate the Spectrum creation. soundfile (Available in: https://pypi.org/project/SoundFile, accessed on 22 June 2021) and librosa (Available in: https://librosa.github.io/librosa, accessed on 22 June 2021) in Python were used for this. The real part of the spectrum was used for obtaining spectrograms. The study also suggests that feature extraction from three different sources (descriptors based on acoustic indices, descriptors based on cepstral information, and descriptors based on the image of spectrogram) can be used together to the benefit of the analysis. For feature extraction, we employed Essentia (Available in: https://essentia.upf.edu, accessed on 22 June 2021), Python, Cython, and C.
In our experiments, for each audio minute, a total of 238 features were extracted. For the experiments, we set up four data sets (DS): (DS1) frogs, birds, and insects, with 2277 instances; (DS2) frogs and birds, with 1437 instances, (DS3) frogs and insects, with 1455 instances; and (DS4) birds and insects, with 1662 instances. Each data set has two feature settings: the first configuration has 102 features in total for each set (original features). For the second configuration, we select best features using the feature selection method based on important features known as Extra-Trees-Classifier (see [37]). After this step, we remained with 30 features for DS1, 30 for DS2, 46 for DS3, and 31 for DS4.

Data Availability and Bioethics

All the raw data used in this study are available on the following platform: https://github.com/LEEClab/soundscape_CCM1_exp01, accessed on 22 June 2021. The Spatial Ecology and Conservation lab of Biodiversity Department of UNESP is in charge of keeping this repository available and updated over time. By the nature of data (audio data only), the UNESP bioethical committee does not require any specific authorization, as no live animals were handled during the sampling period.

5. Results and Discussion

The predictions of data set labels were computed by performing all of the steps described in Section 3, which are: Clustering, Sampling, Annotation, Learning-Prediction, and Visualization. In this scenario, it is important to highlight that, for the purpose of our analysis, user interaction in the step Annotation is recreated or simulated using exceptionally the true labels only to assign the labels to the initial samples. This was so to guarantee the observation of learning effectiveness in the context of the experiments.
For the Learning-Prediction stage, the classifier used was Random Forest (RFC), for showing robust classification in soundscape data in comparison with other classifiers that had been tested such as: Support Vector Classifier (SVC), K-nearest Neighbor Classifier (KNNC), and X-Gradient Boosting classifier (XGBoost). Details of the experiments carried out to analyze the steps of the methodology are described in the next sections.

5.1. Clustering and Sampling Analysis

In order to define the best parameters for the method, the goal of this experiment was to evaluate the first steps of Clustering and Sampling. The accuracy of these steps contributed to increasing the accuracy in the prediction of labels. Therefore, an accurate prediction of the labels can, at the same time, mean a certain segregation of categories of events in data of soundscapes.
In this experiment, we evaluated the following parameters: the number of clusters (k), the total number of samples (p), the number of samples per cluster ( p k ), and the types of strategies for extracting samples (r, m, c, r m , r c , m c ; and r m c —see Section 3.2). The setup of the experiment was defined as follows: (1) The four data sets DS1, DS2, DS3, and DS4 were used; (2) The number of clusters was analyzed in the range of k = { 5 , 6 , . . . , 49 , 50 } ; (3) The number of samples per cluster was established considering the smallest possible number of instances per cluster, so the number of samples per cluster was fixed at p k = { 5 , 10 , | . | } , where | . | represents the total number of instances per grouping and is usually used to define samples.
The visual active learning method was executed 7728 times for the four data sets, and eight sets of results were obtained in the form of tables; four for data sets with all 102 features, and four for data sets using the best features. The results for the DS1 data set are displayed in Table 1, where each row presents the seven types of samples (r, m, c, r m , r c , m c , r m c ) in the form of heatmaps. The maximum accuracy values are displayed in yellow, the minimum accuracy values are displayed in blue; and the intermediate accuracy values are displayed in a color gradient between yellow and blue.
Initially, the clusters were computed using the K-Means (KM) and Hierarchical Agglomerative Clustering (HAC) algorithms, but the best results were obtained with HAC for clusters larger than 20. As expected, we noticed that the larger the number of samples, the higher the accuracy. However, ideally, one wishes to use as few samples as possible to later predict most of the other instances. Therefore, the number of samples is limited by a threshold for the smallest possible number of samples. In this scenario, from the results, we can infer that the proposal to set the number of samples per clustering in | . | is the best option; thus, this parameter can be calculated automatically. Regarding the method for determining initial samples, the analysis focuses on the predominance of maximum accuracy values. After doing a visual analysis of the information in the table using heatmaps, the most suitable strategies for extracting samples in order (from best to worst) are: r, r c , r m , m c , r m c , m, and c.
In order to make a more clear and specific analysis, Figure 4 illustrates the comparison of results between the best features and all 102 features, specifically considering: the two best sample strategies (r and r c ) and the worst sampling method (c) with 10 samples per cluster. In the results presented in Figure 4, we can observe the superiority of the samples r and r c for the data sets under analysis. Thus, through these experiments, the discriminatory capacity of the selected features was verified, achieving high accuracy in prediction of sound categories.

5.2. Visual Analysis via Projections

Some of the results presented in Table 1 can be visualized in Figure 5. A set of the resulting visualizations from the framework is presented: (1) k = 24 as the number of clusters; (2) p k = 5 as the number of samples per cluster; and (3) r c was selected as the method or strategies for extracting samples. The visualizations were generated using t-SNE projections. Figure 5 illustrates each step of the proposed method: Clustering, Sampling, Annotation, and Learning-Prediction. The ground truth is added for comparison.
For each data set, the points of varying colors in Figure 5a,e,i,m, represent the 24 clusters generated in the Clustering stage. The colored dots with up to three colors of Figure 5b,f,j,n, represent the samples from the Sampling step, specifically the colors represent the user’s labeling interaction in the step Annotation. The colored dots in Figure 5c,g,k,o represent the instances with labels that were determined by the prediction in the step Learning-prediction. It is important to report that—for the training of learning—the labels of the samples were considered, and, in the prediction of the rest of the instances, unlabeled of data sets were also used. Finally, the colors of the points in Figure 5d,h,i,p represent the actual true labels of the data set instances. Visually, we can observe the great similarity between the pairs of instances labeled in Figure 5c,d,g,h,k,l,o,p, which, in reality, translates as the visual degree of similarity of the labels in the prediction in relation to the true labels respectively for the four data sets. The accuracy achieved in the labeling task for each of the four data sets varied from high to pretty high: in DS1 equal to 72.6 % (Figure 5c), DS2 equal to 72.36 % (Figure 5g), DS3 equal to 91.09 % (Figure 5k), and DS4 equal to 94.58 % (Figure 5o). Based on these results—at least for our experiments—we can say that, in order to achieve high accuracy in automatic labeling, the experts mostly provide manual annotation for the following percentage of the data: 5.3 % for DS1, 8.35 % for DS2, 8.2 % for DS3, and 7.2 % for DS4.

6. Conclusions, Future Work, and Opportunities

In this study, we presented a method for support sample labeling by employing visual active learning as a label prediction strategy. A framework was tested in the context of a framework for labeling soundscape ecology data. Experiments evaluated each step of the process by the employment of a pre-labeled data set provided by our application partners. We have evaluated the number of clusters, the number of samples per cluster, and the sample strategy within each cluster. Results of the experiments were evaluated according to classification accuracy, which determines the level of prediction of the labels. For clustering unlabeled data, Hierarchical Agglomerative Clustering (HAC) was adapted to the application since it allows for starting the process from user labeled samples, building groups incrementally. According to the results, the best sample strategies were r c and r because these strategies reached the most representative and informative samples from each cluster. The tables in the form of heatmaps with all cluster results identify that trend.
Therefore, we identify an effective parameter configuration for the method and for the current experiment. The best parameters were: (1) which clustering strategy to employ, (2) the number of clusters to generate, (3) the method of strategy to extract samples, (4) the number of samples per clustering, and (5) the technique of visualization where the user will interact in the labeling. Thus, the optimal configuration was: (1) Agglomerative Hierarchical Clustering (AHC) as a strategy to generate clusters; (2) clusters greater than 20; (3) samples of the method random or random combined with contours; and (4) number of samples | . | , per cluster, where | . | is the number of instances per cluster.
The main visual tool employed for user tracking of the process and interpreting the results of each step was multidimensional projections, which can reflect the clustering, sample selection, result of label prediction, and comparison with the ground truth.
With the results achieved in our study, we demonstrate that AL and multidimensional visualization can play an important role in achieving favorable performance with significantly less manual annotation effort. In addition, the accuracy of automatic labeling for our case study varied from high ( 72.36 % ) to very high ( 94.58 % ) for the four data sets, also reflecting a very good advance for the field of soundscape interpretation in the task of discriminating categories of sounds (in our case, groups of animals).
While our approach can be applied to any vector representation of the data set undergoing labeling, the best success in terms of accuracy is achieved when the data set is described by a set of features that has a good potential to discriminate target labels. Our experiments presented here show that aspect of the problem. It is our plan to tackle this problem in the next progression of our efforts in visual analytics for labeling.
Another contribution of our study is the framework itself (see Figure A1) that encapsulates all steps of our proposed method and its calculations. The framework is openly available. Although the framework is dedicated to soundscape data, it should be applicable to implementing the strategies for other data sets where a set of attributes can be extracted which consistently discriminates labels of interest.
In the particular case of soundscape ecology, recordings are being collected constantly, and the amount of data to be labeled increases exponentially in biodiversity monitoring worldwide [22]. Thus, developing solid and easy-to-use methods are of utmost importance for project managers that want to speedily extract information from their data. After data recording in the field, labeling is one of the most time-consuming tasks that precedes extracting knowledge. Therefore, contributions such as automatic labeling combined with a good visualization tool can be crucial for the success of conservation projects. The application of this approach was limited to experimenting with categorical sound events. While this could be construed as one of its limitations, the applicability of the results of the study can bring benefits to science such as understanding the behavior of certain environments, which can lead to the development of environmental monitoring strategies and policies of conservation.
In the next steps of the application studies, we will try to focus the framework in other categories, such as primates, bats, rain, dogs, human conversation, cars, airplane, guns, stream, wind, and background noises, and also verify applicability to another level of audio data resolution, such as identifying species.
More crucial to the development of visual analytics strategies in the future, it essential that the problem of labeling, which drives data analysis and understanding, machine learning, and crucial feature selection in many different applications, is dedicated more effort in order to find a balance and also to generalize the approach so that it can be applicable to a large variety of domains. Examples of areas where accelerating labeling can have a real impact are document analysis, biological data interpretation, and image and video interpretation, to name a few.

Author Contributions

Conceptualization: L.H.H. and R.M.; all authors participated in regular meetings during the project and contributed to ideas to shape the concepts and complement their the execution. Implementation: L.H.H.; Contextualization and verification: L.H.H. and R.M.; Testing and Use Case Scenarios: All authors.; Writing: all authors.; Revision: M.C.R. and R.M. All authors have read and agreed to the published version of the manuscript.

Funding

L.H.H. was financed in part by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior—Brasil (CAPES)—Finance Code 001; MCR was partly funded by the São Paulo Research Foundation—FAPESP (processes 2013/50421-2; 2020/01779-5) and CNPq (processes 312045/2013-1; 312292/2016-3; 442147/2020-1); RM was partly funded by Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq) Grant No. 307411/2016-8.

Data Availability Statement

The data presented in this study are available on the following platform: https://github.com/LEEClab/soundscape_CCM1_exp01 (accessed on 28 June 2021).

Acknowledgments

The authors acknowledge the work of students in LEEC lab, in particular Lucas Gaspar, for the labeling of the data set.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Overview of the Visual Active Learning Framework for Soundscape Ecology

Our methodology was realized by a framework that we call SoundscapeX (code available at https://github.com/hhliz/vizactivelearning, accessed on 22 June 2021). The software implementation is based on a Client/Server architecture developed in Python that mainly employs the following libraries for the back-end: Tornado, Scikit-learn, Librosa, Pandas, and MongoDB. For visualizations, we employ the D3 Javascript library.
On the server, algorithms written in Python are used to perform clustering, sampling, learning, prediction, and projection tasks. On the client-side, algorithms written in D3/js are used to generate the layout of the views, mainly the projections, and, with these projections, the user interacts to perform the manual labeling. The application window is a view created with D3/JS in the web browser (client), but the visualization geometry is created by complex calculations from algorithms written in Python (server). Python and D3 communicate via JSON message passing.
Here, we provide an overview of the data exploration and labeling functionalities. Figure A1 presents an overview of SoundscapeX and the interface functions for exploring and labeling the data. The main interface components are shown in the regions labeled: (A) The process of labeling starts in the configuration panel, where the user issues a query of a data set to be used, the set of features to represent the data, type of the normalization, clustering technique to be used (k-means, HCA), number of clusters, and visualization (t-SNE or Uniform Manifold Approximation and Projection UMAP [45]). (B) At the top, the first seven buttons are used to interact and explore the data within the projection in region D. Then, the remaining buttons are a mini configuration panel to determine the type of sampling, number of samples, and launch of a small interface where the user can listen to the audio and label selected data samples. (C) The region presents the spectrogram of the whole set of audios in the form of a timeline; it is named the “Time-Line-Spectrogram” (TLS) and offers an overview of the audios under analysis. In addition, TLS is coordinated with the projections data allowing exploration. (D) In this region, the visualizations of the active learning process can be observed, by displaying the projection with colored clustering, the selection of samples to be labeled by the user, and finally the prediction of the labels. An example of the alternative visualizations in this region is seen in Figure 5. (E) The region presents the ground truth of the data when one is available, or the final labeling in normal circumstances. (F) Polar histogram of features. (G) Current set of features used in the data set; colors indicate the source (acoustic indices, spectrograms, or sound signal). (H) Legend of labels in the data set.
Figure A1. Screenshot of the main interface of the Framework. (A) Initially, a configuration parameter panel (B) Allows interact and explore the data within the projection in region D. (C) TLS is coordinated with the projection’s data, allowing exploration interactive. (D) Multiple visualizations of the active learning process. (E) The ground truth. (F) Polar histogram of features. (G) Features used in the experiment. (H) Legend of labels in the data set.
Figure A1. Screenshot of the main interface of the Framework. (A) Initially, a configuration parameter panel (B) Allows interact and explore the data within the projection in region D. (C) TLS is coordinated with the projection’s data, allowing exploration interactive. (D) Multiple visualizations of the active learning process. (E) The ground truth. (F) Polar histogram of features. (G) Features used in the experiment. (H) Legend of labels in the data set.
Information 12 00265 g0a1

References

  1. Settles, B. Active Learning Literature Survey. In Computer Sciences Technical Report 1648; University of Wisconsin–Madison: Madison, WI, USA, 2009. [Google Scholar]
  2. Piczak, K.J. ESC: Dataset for Environmental Sound Classification. In Proceedings of the 23rd ACM International Conference on Multimedia (MM ’15); Association for Computing Machinery: New York, NY, USA, 2015; pp. 1015–1018. [Google Scholar] [CrossRef]
  3. Tuia, D.; Volpi, M.; Copa, L.; Kanevski, M.; Munoz-Mari, J. A Survey of Active Learning Algorithms for Supervised Remote Sensing Image Classification. IEEE J. Sel. Top. Signal Process. 2011, 5, 606–617. [Google Scholar] [CrossRef]
  4. Pereira-Santos, D.; Prudêncio, R.B.C.; de Carvalho, A.C. Empirical investigation of active learning strategies. Neurocomputing 2019, 326–327, 15–27. [Google Scholar] [CrossRef]
  5. Riccardi, G.; Hakkani-Tur, D. Active learning: Theory and applications to automatic speech recognition. IEEE Trans. Speech Audio Process. 2005, 13, 504–511. [Google Scholar] [CrossRef]
  6. Kapoor, A.; Horvitz, E.; Basu, S. Selective Supervision: Guiding Supervised Learning with Decision-Theoretic Active Learning. In Proceedings of the 20th International Joint Conference on Artifical Intelligence (IJCAI’07); Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2007; pp. 877–882. [Google Scholar]
  7. Moskovitch, R.; Nissim, N.; Stopel, D.; Feher, C.; Englert, R.; Elovici, Y. Improving the Detection of Unknown Computer Worms Activity Using Active Learning. In KI 2007: Advances in Artificial Intelligence; Hertzberg, J., Beetz, M., Englert, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 489–493. [Google Scholar]
  8. Hu, R. Active Learning for Text Classification. Ph.D. Thesis, Technological University Dublin, Dublin, Ireland, 2011. [Google Scholar] [CrossRef]
  9. Abdelwahab, M.; Busso, C. Active Learning for Speech Emotion Recognition Using Deep Neural Network. In Proceedings of the 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII), Cambridge, UK, 3–6 September 2019; pp. 1–7. [Google Scholar]
  10. Mandel, M.I.; Poliner, G.E.; Ellis, D.P. Support Vector Machine Active Learning for Music Retrieval. Multimed. Syst. 2006, 12, 3–13. [Google Scholar] [CrossRef] [Green Version]
  11. Xin, D.; Ma, L.; Liu, J.; Macke, S.; Song, S.; Parameswaran, A. Accelerating Human-in-the-loop Machine Learning: Challenges and opportunities. In Conjunction with the 2018 ACM SIGMOD/PODS Conference (DEEM 2018), 15 June 2018, Proceedings of the 2nd Workshop on Data Management for End-To-End Machine Learning; Association for Computing Machinery, Inc.: New York, NY, USA, 2018; pp. 1–4. [Google Scholar] [CrossRef]
  12. Liao, H.; Chen, L.; Song, Y.; Ming, H. Visualization-Based Active Learning for Video Annotation. IEEE Trans. Multimed. 2016, 18, 2196–2205. [Google Scholar] [CrossRef]
  13. Limberg, C.; Krieger, K.; Wersing, H.; Ritter, H. Active Learning for Image Recognition Using a Visualization-Based User Interface. In Artificial Neural Networks and Machine Learning—ICANN 2019: Deep Learning; Tetko, I.V., Kůrková, V., Karpov, P., Theis, F., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 495–506. [Google Scholar]
  14. Maijala, P.; Shuyang, Z.; Heittola, T.; Virtanen, T. Environmental noise monitoring using source classification in sensors. Appl. Acoust. 2018, 129, 258–267. [Google Scholar] [CrossRef]
  15. Servick, K. Eavesdropping on Ecosystems. Science 2014, 343, 834–837. [Google Scholar] [CrossRef] [PubMed]
  16. Farina, A.; Pieretti, N.; Piccioli, L. The soundscape methodology for long-term bird monitoring: A Mediterranean Europe case-study. Ecol. Inform. 2011, 6, 354–363. [Google Scholar] [CrossRef]
  17. Farina, A.; Pieretti, N. Sonic environment and vegetation structure: A methodological approach for a soundscape analysis of a Mediterranean maqui. Ecol. Inform. 2014, 21, 120–132. [Google Scholar] [CrossRef]
  18. Putland, R.; Mensinger, A. Exploring the soundscape of small freshwater lakes. Ecol. Inform. 2020, 55, 101018. [Google Scholar] [CrossRef]
  19. Kasten, E.P.; Gage, S.H.; Fox, J.; Joo, W. The remote environmental assessment laboratory’s acoustic library: An archive for studying soundscape ecology. Ecol. Inform. 2012, 12, 50–67. [Google Scholar] [CrossRef]
  20. LeBien, J.G.; Zhong, M.; Campos-Cerqueira, M.; Velev, J.; Dodhia, R.; Ferres, J.; Aide, T. A pipeline for identification of bird and frog species in tropical soundscape recordings using a convolutional neural network. Ecol. Inform. 2020, 59, 101113. [Google Scholar] [CrossRef]
  21. Pijanowski, B.C.; Farina, A.; Gage, S.H.; Dumyahn, S.L.; Krause, B.L. What is soundscape ecology? An introduction and overview of an emerging new science. Landsc. Ecol. 2011, 26, 1213–1232. [Google Scholar] [CrossRef]
  22. Scarpelli, M.D.; Ribeiro, M.C.; Teixeira, F.Z.; Young, R.J.; Teixeira, C.P. Gaps in terrestrial soundscape research: It’s time to focus on tropical wildlife. Sci. Total. Environ. 2020, 707, 135403. [Google Scholar] [CrossRef]
  23. Hu, W.; Bulusu, N.; Chou, C.T.; Jha, S.; Taylor, A.; Tran, V.N. Design and Evaluation of a Hybrid Sensor Network for Cane Toad Monitoring. ACM Trans. Sen. Netw. 2009, 5, 4:1–4:28. [Google Scholar] [CrossRef]
  24. Joo, W.; Gage, S.H.; Kasten, E.P. Analysis and interpretation of variability in soundscapes along an urban-rural gradient. Landsc. Urban Plan. 2011, 103, 259–276. [Google Scholar] [CrossRef]
  25. Parks, S.E.; Miksis-Olds, J.L.; Denes, S.L. Assessing marine ecosystem acoustic diversity across ocean basins. Ecol. Inform. 2014, 21, 81–88. [Google Scholar] [CrossRef]
  26. Sueur, J.; Farina, A. Ecoacoustics: the Ecological Investigation and Interpretation of Environmental Sound. Biosemiotics 2015, 8, 493–502. [Google Scholar] [CrossRef]
  27. Bellisario, K.; Broadhead, T.; Savage, D.; Zhao, Z.; Omrani, H.; Zhang, S.; Springer, J.; Pijanowski, B. Contributions of MIR to soundscape ecology. Part 3: Tagging and classifying audio features using a multi-labeling k-nearest neighbor approach. Ecol. Inform. 2019, 51, 103–111. [Google Scholar] [CrossRef]
  28. Abramson, Y.; Freund, Y. SEmi-automatic VIsuaL LEarning (SEVILLE): A tutorial on active learning for visual object recognition. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), San Diego, CA, USA, 20–26 June 2005; p. 11. [Google Scholar]
  29. Turtinen, M.; Pietikänien, M. Labeling of Textured Data with Co-Training and Active Learning. Proc. Workshop on Texture Analysis and Synthesis. 2005, pp. 137–142. Available online: https://core.ac.uk/display/20963071 (accessed on 22 June 2021).
  30. Lecerf, L.; Chidlovskii, B. Visalix: A web application for visual data analysis and clustering. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, Paris, France, 29 June–1 July 2009. [Google Scholar]
  31. Namee, B.M.; Hu, R.; Delany, S.J. Inside the Selection Box: Visualising active learning selection strategies. nips2010. 2010, p. 15. Available online: https://cseweb.ucsd.edu/~lvdmaaten/workshops/nips2010/papers/namee.pdf (accessed on 22 June 2021).
  32. Huang, L.; Matwin, S.; de Carvalho, E.J.; Minghim, R. Active Learning with Visualization for Text Data. In Proceedings of the 2017 ACM Workshop on Exploratory Search and Interactive Data Analytics (ESIDA ’17); Association for Computing Machinery: New York, NY, USA, 2017; pp. 69–74. [Google Scholar] [CrossRef]
  33. van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  34. Ristin, M.; Guillaumin, M.; Gall, J.; Van Gool, L. Incremental Learning of Random Forests for Large-Scale Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 490–503. [Google Scholar] [CrossRef] [PubMed]
  35. Tasar, O.; Tarabalka, Y.; Alliez, P. Incremental Learning for Semantic Segmentation of Large-Scale Remote Sensing Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2019, 12, 3524–3537. [Google Scholar] [CrossRef] [Green Version]
  36. Shin, G.; Yooun, H.; Shin, D.; Shin, D. Incremental Learning Method for Cyber Intelligence, Surveillance, and Reconnaissance in Closed Military Network Using Converged IT Techniques. Soft Comput. 2018, 22, 6835–6844. [Google Scholar] [CrossRef]
  37. Hilasaca, L.M.H.; Gaspar, L.P.; Ribeiro, M.C.; Minghim, R. Visualization and categorization of ecological acoustic events based on discriminant features. Ecol. Indic. 2021, 107316. [Google Scholar] [CrossRef]
  38. Han, W.; Coutinho, E.; Ruan, H.; Li, H.; Schuller, B.; Yu, X.; Zhu, X. Semi-Supervised Active Learning for Sound Classification in Hybrid Learning Environments. PLoS ONE 2016, 11, 1–23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Shuyang, Z.; Heittola, T.; Virtanen, T. Active learning for sound event classification by clustering unlabeled data. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 751–755. [Google Scholar] [CrossRef]
  40. Salamon, J.; Jacoby, C.; Bello, J.P. A Dataset and Taxonomy for Urban Sound Research. In Proceedings of the 22nd ACM International Conference on Multimedia (MM ‘14); Association for Computing Machinery: New York, NY, USA, 2014; pp. 1041–1044. [Google Scholar] [CrossRef]
  41. Shuyang, Z.; Heittola, T.; Virtanen, T. An Active Learning Method Using Clustering and Committee-Based Sample Selection for Sound Event Classification. In Proceedings of the 2018 16th International Workshop on Acoustic Signal Enhancement (IWAENC), Tokyo, Japan, 17–20 September 2018; pp. 116–120. [Google Scholar] [CrossRef]
  42. Kholghi, M.; Phillips, Y.; Towsey, M.; Sitbon, L.; Roe, P. Active learning for classifying long-duration audio recordings of the environment. Methods Ecol. Evol. 2018, 9, 1948–1958. [Google Scholar] [CrossRef]
  43. Shuyang, Z.; Heittola, T.; Virtanen, T. Active Learning for Sound Event Detection. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 2895–2905. [Google Scholar] [CrossRef]
  44. Wang, Y.; Mendez Mendez, A.E.; Cartwright, M.; Bello, J.P. Active Learning for Efficient Audio Annotation and Classification with a Large Amount of Unlabeled Data. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 880–884. [Google Scholar] [CrossRef]
  45. McInnes, L.; Healy, J.; Melville, J. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv 2020, arXiv:1802.03426. [Google Scholar]
Figure 1. Framework of the proposed method: visual active learning for labeling data.
Figure 1. Framework of the proposed method: visual active learning for labeling data.
Information 12 00265 g001
Figure 2. Types of sampling (random, medoid and contour). C1 (cluster 1) and C2 (cluster 2).
Figure 2. Types of sampling (random, medoid and contour). C1 (cluster 1) and C2 (cluster 2).
Information 12 00265 g002
Figure 3. Example of visual iteration with Time-Line-Spectrogram (TLS): the points selected in the yellow color area in the projection are also visualized with yellow color in the spectrogram timeline (top of figure). In this view, projection multidimensional t-SNE is used for visualizing each data point.
Figure 3. Example of visual iteration with Time-Line-Spectrogram (TLS): the points selected in the yellow color area in the projection are also visualized with yellow color in the spectrogram timeline (top of figure). In this view, projection multidimensional t-SNE is used for visualizing each data point.
Information 12 00265 g003
Figure 4. Comparison of accuracy results (in percentage 100 % ) according to the types of features and the sampling strategies: the graphs show results when the best features are used (continuous lines) and when all 102 features are used (dashed lines). This comparison is made considering the types of samples r, r c , and c, with 10 samples per cluster for data sets DS1 (a), DS2 (b), DS3 (c), and DS4 (d). In the graphs, the x-axis corresponds to the cluster values and the y-axis corresponds to the accuracy values.
Figure 4. Comparison of accuracy results (in percentage 100 % ) according to the types of features and the sampling strategies: the graphs show results when the best features are used (continuous lines) and when all 102 features are used (dashed lines). This comparison is made considering the types of samples r, r c , and c, with 10 samples per cluster for data sets DS1 (a), DS2 (b), DS3 (c), and DS4 (d). In the graphs, the x-axis corresponds to the cluster values and the y-axis corresponds to the accuracy values.
Information 12 00265 g004
Figure 5. Sample visualizations of label predictions: the lines from top to bottom show the t-SNE projections of datasets DS1, DS2, DS3, and DS4 with their best features. Along the columns from left to right, colors represent each of the steps of the method: Clustering; Sampling and Annotation; and Learning-Prediction, as well as the Ground-truth.
Figure 5. Sample visualizations of label predictions: the lines from top to bottom show the t-SNE projections of datasets DS1, DS2, DS3, and DS4 with their best features. Along the columns from left to right, colors represent each of the steps of the method: Clustering; Sampling and Annotation; and Learning-Prediction, as well as the Ground-truth.
Information 12 00265 g005
Table 1. Accuracy results (in percent 100 % ) of the labeling for the DS1 set with the best 30 features: each line represents a different number of clusters (k). In the columns from left to right, three groups of columns are indicated ( p k ), which refers to the sample numbers per clustering (5, 10 e | . | ). In each block of columns from left to right are indicated: (1) the total number of samples (p); (2) the percentage of samples (%) in relation to the total number of instances in the data set; and (3) the accuracy for each of the sampling strategies (random (r), medoid (m), contour (c)) and their combinations r m , r c , m c e r m c In the heatmap, the yellow color represents the maximum accuracy values achieved (labels prediction), the blue color represents the minimum values, the intermediate accuracy values are visualized with a color gradient between the yellow, orange, and blue colors.
Table 1. Accuracy results (in percent 100 % ) of the labeling for the DS1 set with the best 30 features: each line represents a different number of clusters (k). In the columns from left to right, three groups of columns are indicated ( p k ), which refers to the sample numbers per clustering (5, 10 e | . | ). In each block of columns from left to right are indicated: (1) the total number of samples (p); (2) the percentage of samples (%) in relation to the total number of instances in the data set; and (3) the accuracy for each of the sampling strategies (random (r), medoid (m), contour (c)) and their combinations r m , r c , m c e r m c In the heatmap, the yellow color represents the maximum accuracy values achieved (labels prediction), the blue color represents the minimum values, the intermediate accuracy values are visualized with a color gradient between the yellow, orange, and blue colors.
k pk = 5 pk = 10 pk = | . |
(p)(%)(r)(m)(c)( rm )( rc )( mc )( rmc )(p)(%)(r)(m)(c)( rm )( rc )( mc )( rmc )(p)(%)(r)(m)(c)( rm )( rc )( mc )( rmc )
5251.1062.7958.3032.5059.6447.3858.5362.79502.2064.6664.1231.8860.6658.6953.8458.151034.5267.2562.7934.6466.6562.9355.2457.41
6301.3265.2462.4431.5160.2143.2159.2860.17602.6464.1063.7833.6062.0758.0552.6458.371124.9264.4363.7440.0567.4869.2455.6657.14
7351.5471.1961.4638.8567.5359.2862.3662.62703.0771.9560.6737.2066.3865.4357.5963.251195.2372.8563.3037.2668.4067.9855.2861.03
8401.7665.6261.4237.8265.5359.1064.6864.33803.5167.9662.9032.1364.5467.0560.3161.721265.5367.5563.6939.0168.6268.0661.1364.53
9451.9860.2664.6140.1462.4664.2566.7166.58903.9567.0864.6133.7468.5468.2265.1165.571365.9769.8764.2741.7171.1863.8064.8366.14
10502.2063.3664.8943.6562.7366.9167.8567.491004.3970.5664.1740.1966.6568.0867.9865.551426.2471.5265.4348.0173.4071.0568.9968.29
11552.4264.1362.1546.9865.8467.9666.6165.981104.8368.7663.5443.7968.9470.1467.7465.391496.5473.5064.2943.7071.9569.4169.0866.17
12602.6470.5560.6245.2067.3468.2063.7865.311205.2775.1562.8746.1364.7271.1267.4564.861556.8176.4864.8949.6269.3272.2067.8666.21
13652.8567.1861.3544.7166.5960.6764.8767.091305.7168.7062.5550.3067.5468.3768.7065.161607.0370.0064.3860.2368.0270.1569.2566.98
14703.0769.8760.1758.1367.2066.3865.8465.291406.1569.9162.5660.6068.7968.6968.6564.251667.2973.1462.7259.0266.5176.6970.8267.69
15753.2968.9461.2254.1366.1765.6763.4063.901506.5971.9859.7158.4468.9771.3769.4465.261717.5172.9862.6858.9770.8573.1269.9467.57
16803.5166.6461.3656.9466.1867.5964.5463.311607.0370.8162.2157.9667.4573.5069.4967.601777.7770.4864.3860.8671.1075.2469.9569.14
17853.7367.3862.5561.8269.2168.4867.4765.881707.4771.9562.7059.8068.1571.8671.2967.681807.9173.7263.6157.5171.0173.0671.4469.19
18903.9566.9462.8759.3568.2767.8668.7764.291807.9172.9663.0960.7571.6372.1071.3968.431878.2173.2564.3162.5467.1372.8770.2968.04
19954.1769.4362.7461.1867.5168.6568.7066.681908.3473.8463.4461.0470.9673.7471.0168.281908.3474.8965.2662.4871.3572.4070.2969.38
201004.3970.5663.3057.8367.9863.9969.1368.262008.7872.8564.5263.3172.2273.5773.1872.321968.6172.8064.4462.0970.9874.2072.0370.40
211054.6167.5063.4961.6066.4469.1169.2066.852109.2272.5764.2561.4471.0271.4673.3972.182028.8770.8064.0058.3169.5473.7872.9671.28
221104.8369.5962.6764.2867.1068.9068.4467.012209.6676.5263.9362.0871.6675.7971.8569.422059.0073.9464.4357.7772.3975.8771.4369.88
231155.0569.2961.8962.8665.5470.4068.1365.8223010.1072.7462.7761.6571.1876.8072.7971.522089.1372.2164.5258.0571.5373.6671.5871.00
241205.2770.4762.3163.6567.2272.6068.4766.3424010.5472.4663.4865.4471.1874.3773.1071.282139.3574.1364.5863.8171.3275.5372.2970.69
251255.4971.2863.4864.9264.7869.5668.7767.8025010.9874.4965.0266.0170.3576.7672.7271.882169.4973.7064.6860.8969.0076.7172.6872.97
261305.7171.5064.2862.2369.2171.6370.4267.2626011.4273.1366.8364.9570.7075.3174.6273.332209.6673.9466.8061.8469.5275.5573.6571.46
271355.9371.9964.4361.0667.0469.3369.0966.2927011.8673.3966.7265.6270.3076.2373.5971.102279.9772.5965.6661.4172.7374.5473.0771.41
281406.1571.3164.1660.9369.3569.0770.4767.0128012.3073.4666.8064.1573.5674.7173.8671.7123010.1072.5066.0562.2972.6974.5073.8270.35
291456.3773.2664.5958.7265.9571.6270.7367.6829012.7475.7466.5364.4773.2876.6073.7371.9723510.3275.3764.8462.1069.5975.8172.9770.23
301506.5973.2564.2764.1767.4772.2169.2568.3630013.1875.3766.2666.8770.9777.1474.0073.8823710.4173.4866.9663.4371.2774.8571.7669.56
311556.8172.0163.8164.1469.9374.4670.4568.9431013.6172.7067.1666.5071.7375.1973.9773.8424110.5873.0466.7065.2872.5973.0473.4371.41
321607.0373.3664.6766.2369.4471.9970.2968.6332014.0577.1167.2568.7372.1577.7273.8972.8424510.7675.8467.4764.2273.1376.2873.3872.59
331657.2572.0265.3465.5366.7674.7670.7469.7433014.4974.4767.2867.9572.3777.4075.3074.6524910.9474.5166.9667.5071.6572.6873.8273.67
341707.4772.9565.5467.4470.7276.7070.0070.5334014.9377.5966.9168.9771.6675.3275.1274.5725411.1675.4366.5869.2572.9678.2572.6672.86
351757.6974.4165.0867.9872.3672.8469.3669.8935015.3776.4968.1971.5674.2678.3175.3074.1325711.2974.6067.6770.2574.0175.5474.5071.63
361807.9172.5364.4368.1472.2571.2070.1569.1936015.8176.2166.8271.3672.9877.7875.5975.1426011.4273.6767.4871.0070.9077.9473.2870.80
371858.1275.0566.6866.3072.2875.0071.9470.9837016.2577.2469.0170.7974.0478.6675.6274.9926211.5175.8868.4473.1573.1075.9373.4072.16
381908.3475.1364.6466.9471.6373.2671.7370.9238016.6976.7068.9072.3275.2876.0775.1275.1826411.5974.4268.2171.9875.5674.9672.7872.48
391958.5673.1564.9466.7174.0275.2271.8571.6639017.1377.1668.7370.4874.0376.3174.4674.3126811.7775.9668.1970.6872.6277.2072.3271.38
402008.7873.9564.2867.6075.1675.1671.9871.2739917.5279.0768.0070.0273.4876.8974.8173.5427111.9077.5266.1571.4475.6274.6372.7872.38
412059.0071.7264.2967.3771.0976.1172.1070.2440917.9677.4667.9371.1575.5476.2874.2074.5727412.0375.0967.3070.0072.7474.8473.3471.24
422109.2271.6064.9765.9970.3976.1572.0970.1641918.4076.6469.0569.5471.9178.5374.8774.4927712.1774.3567.9568.7572.6077.6075.3071.85
432159.4475.4666.2566.8872.4175.5671.9269.9542918.8478.3568.7269.0572.7378.9074.5773.7628012.3074.8167.8070.3674.0177.5273.8670.56
442209.6675.2166.6567.8772.7873.3671.9070.1243919.2879.8768.9969.9175.5777.0474.9275.2028312.4377.0367.3069.4172.9275.6374.0272.77
452259.8874.2265.9869.0171.3073.6872.9570.0444919.7277.2468.5470.4672.9276.1575.4475.0028612.5673.9367.4071.5273.7376.3474.2371.52
4623010.1075.3865.5169.4774.2676.5071.8170.2145920.1678.4968.0470.5277.0677.6774.7074.5128812.6576.6267.6770.2973.8174.5674.7672.40
4723510.3273.9065.3869.8370.9675.1771.3569.6046920.6076.5568.3670.5875.2877.4375.7774.5129012.7474.9467.0471.3674.1375.8472.9771.62
4824010.5475.7565.9368.5371.8775.1172.1270.3647921.0477.2568.9170.6974.5376.2575.3674.1029312.8776.4166.9971.0773.7475.3574.9072.18
4924510.7675.6966.1969.1472.1577.4672.4470.3948921.4876.6868.4069.3074.5579.5377.4675.8629613.0075.4266.3372.3475.4276.7375.2772.69
5025010.9873.0667.0068.1874.1074.9972.7271.1549921.9179.5370.1370.3675.7078.4076.4376.1130013.1876.5367.1770.2173.6077.7974.5172.23
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hilasaca, L.H.; Ribeiro, M.C.; Minghim, R. Visual Active Learning for Labeling: A Case for Soundscape Ecology Data. Information 2021, 12, 265. https://doi.org/10.3390/info12070265

AMA Style

Hilasaca LH, Ribeiro MC, Minghim R. Visual Active Learning for Labeling: A Case for Soundscape Ecology Data. Information. 2021; 12(7):265. https://doi.org/10.3390/info12070265

Chicago/Turabian Style

Hilasaca, Liz Huancapaza, Milton Cezar Ribeiro, and Rosane Minghim. 2021. "Visual Active Learning for Labeling: A Case for Soundscape Ecology Data" Information 12, no. 7: 265. https://doi.org/10.3390/info12070265

APA Style

Hilasaca, L. H., Ribeiro, M. C., & Minghim, R. (2021). Visual Active Learning for Labeling: A Case for Soundscape Ecology Data. Information, 12(7), 265. https://doi.org/10.3390/info12070265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop