Identifying Patterns of Human and Bird Activities Using Bioacoustic Data

: In general, humans and animals often interact within the same environment at the same time. Human activities may disturb or a ﬀ ect some bird activities. Therefore, it is important to monitor and study the relationships between human and animal activities. This paper proposed a system able not only to automatically classify human and bird activities using bioacoustic data, but also to automatically summarize patterns of events over time. To perform automatic summarization of acoustic events, a frequency–duration graph (FDG) framework was proposed to summarize the patterns of human and bird activities. This system ﬁrst performs data pre-processing work on raw bioacoustic data and then applies a support vector machine (SVM) model and a multi-layer perceptron (MLP) model to classify human and bird chirping activities before using the FDG framework to summarize results. The SVM model achieved 98% accuracy on average and the MLP model achieved 98% accuracy on average across several day-long recordings. Three case studies with real data show that the FDG framework correctly determined the patterns of human and bird activities over time and provided both statistical and graphical insight into the relationships between these two events.


Introduction
Humans often interact with forests, which may disturb the activities of birds and other animals which live there; this interaction might also disturb the biodiversity in these forests [1][2][3].Therefore, it is important to understand the relationships between bird and human activities in forests [4].Due to the large size of forests, it is expensive to send experts to each forest to monitor these activities manually.Thus, there is a demand for methods able to summarize the patterns of bird chirping and human activities in forest areas over time.Recently, bioacoustic data have been utilized to monitor different phenomena [5] in natural environments, such as forests, for many different applications, such as species classification [6] and environmental monitoring [7,8].Bioacoustics analysis could also be used to address the challenge of monitoring and summarizing interactions between human and bird activity in forests.Setting up bioacoustic sensors in forests provides an economic and efficient way to collect bioacoustic data, providing rich information [9] with which to study animal communities.However, analyzing and summarizing these data to provide insight into how human and bird activities affect each other has not been addressed until now.
Most researchers take advantage of remote bioacoustic sensors to work on classifying specific species, especially birds.Salamon and Bello [10] applied a deep convolutional neural network to bird flight calls to classify 43 different species based on the features of mel-frequency cepstral coefficients (MFCCs).Their overall classification accuracy was 96%.Zhao and Zhang [11] applied a support vector machine (SVM) to classify 10 species with an average precision of 95%.Moreover, Zamanian and Pourghassem [12] fed 29 features extracted from both the temporal and spectral domains into a multi-layer perceptron (MLP) model to classify insects and achieved an average of 99% accuracy.The first limitation of the previous works is that they focused only on classifying a specific animal group, such as a bird species [7,10], insect species [13,14], etc.However, it is also meaningful to study mixed category classifications, such as bird chirping and human activities.In some environmental studies, details about which species of birds are present in a location may not be as important as understanding how overall bird chirping changes over time as a whole event.The aim of such studies might be to study or compare this kind of event with other events, like human activities [15,16].The second limitation is that most previous works on bioacoustics have focused only on researching the appropriate methods by which to classify or detect events [17][18][19] and comparing the results with different methods, without studying how to summarize the patterns of these events over time (for example, how the patterns change by month or season).This means that these analyses, while accurate, require thorough expert analysis to gain real insight into species interactions.Effective summarization is required to reduce costs and quickly inform organizations about how to best interact with forests with minimal disturbance to other species, such as birds.
In this paper, we proposed an integrated system to not only classify bird chirping and human activities, but also to provide a frequency-duration graph (FDG) framework to summarize the patterns of these activities over time.This system will help researchers who are interested in studying the relationships between bird and human activities to gain a deeper insight into the patterns of these events as they evolve in forests at different time periods over time.
The remainder of this paper is organized as follows.Section 2 introduces the materials and methods, Section 3 gives the case-study-based evaluation, Section 4 shows the results and discussion, and Section 5 offers the conclusions of the whole study.

Materials and Methods
This system proposed consists of two stages.The first stage recognizes different events in a natural forest environment based on bioacoustic data, and the second stage summarizes the patterns of these events during any given time period.In this system, data are segmented based on expert annotations and input into our proposed system, although the system should be compatible with the automatic segmentation methods derived in the existing literature [19].In this work, we focused only on how to recognize events based on already-event-segmented periods, and how to summarize patterns of events over time.Figure 1 shows a system overview.first limitation of the previous works is that they focused only on classifying a specific animal group, such as a bird species [7,10], insect species [13,14], etc.However, it is also meaningful to study mixed category classifications, such as bird chirping and human activities.In some environmental studies, details about which species of birds are present in a location may not be as important as understanding how overall bird chirping changes over time as a whole event.The aim of such studies might be to study or compare this kind of event with other events, like human activities [15,16].The second limitation is that most previous works on bioacoustics have focused only on researching the appropriate methods by which to classify or detect events [17][18][19] and comparing the results with different methods, without studying how to summarize the patterns of these events over time (for example, how the patterns change by month or season).This means that these analyses, while accurate, require thorough expert analysis to gain real insight into species interactions.Effective summarization is required to reduce costs and quickly inform organizations about how to best interact with forests with minimal disturbance to other species, such as birds.
In this paper, we proposed an integrated system to not only classify bird chirping and human activities, but also to provide a frequency-duration graph (FDG) framework to summarize the patterns of these activities over time.This system will help researchers who are interested in studying the relationships between bird and human activities to gain a deeper insight into the patterns of these events as they evolve in forests at different time periods over time.
The remainder of this paper is organized as follows.Section 2 introduces the materials and methods, Section 3 gives the case-study-based evaluation, Section 4 shows the results and discussion, and Section 5 offers the conclusions of the whole study.

Materials and Methods
This system proposed consists of two stages.The first stage recognizes different events in a natural forest environment based on bioacoustic data, and the second stage summarizes the patterns of these events during any given time period.In this system, data are segmented based on expert annotations and input into our proposed system, although the system should be compatible with the automatic segmentation methods derived in the existing literature [19].In this work, we focused only on how to recognize events based on already-event-segmented periods, and how to summarize patterns of events over time.Figure 1 shows a system overview.First, the raw bioacoustic data were converted into the WAV format with a 22.05 kHz sample rate.Next, the background noises of the preprocessed data were removed by applying low and highpass filters [7,20].The low and high-pass filters were used to filter out the noisy frequency regions that we were not interested in.According to the dataset and Pijanowski et al. [21], human noises and bird chirping lie in the frequency range of 1 kHz to 10 kHz.As such, the high-pass filter was set to 10 kHz, and the low-pass filter was set to 1 kHz.This meant that the frequency regions above 10 kHz and below 1 kHz were filtered out.The background noises to be filtered out included wind noises, which are in the low-frequency range (from 20 Hz to 200 Hz), and special machine noises, which are above 10 kHz.After background denoising, features were extracted based on the pre-segmented periods of events.Different classifiers were then implemented to classify different events using the extracted features.The final step was to summarize the patterns of events based on the given period First, the raw bioacoustic data were converted into the WAV format with a 22.05 kHz sample rate.Next, the background noises of the preprocessed data were removed by applying low and high-pass filters [7,20].The low and high-pass filters were used to filter out the noisy frequency regions that we were not interested in.According to the dataset and Pijanowski et al. [21], human noises and bird chirping lie in the frequency range of 1 kHz to 10 kHz.As such, the high-pass filter was set to 10 kHz, and the low-pass filter was set to 1 kHz.This meant that the frequency regions above 10 kHz and below 1 kHz were filtered out.The background noises to be filtered out included wind noises, which are in the low-frequency range (from 20 Hz to 200 Hz), and special machine noises, which are above 10 kHz.After background denoising, features were extracted based on the pre-segmented periods of events.Different classifiers were then implemented to classify different events using the extracted features.
The final step was to summarize the patterns of events based on the given period of bioacoustic data.To make the system flexible, the number of events and classifiers can be modified to suit different researchers' interests.
In summary, the goal of this system was, given a set of bioacoustic recordings as the input, to output a summary of different events' patterns over time.

Feature Extraction
A set of features were extracted based on each period of events based on the bioacoustic signal.These features are the mean of zero-crossing rate (zcrMean), standard deviation of zero-crossing rate (zcrSD), mean of spectral centroid (specMean), standard deviation of spectral centroid (specSD), mean of root mean square (rmsMean), mel-frequency cepstral coefficients (MFCCs), and spectral entropy (spectralEntropy).

zcrMean, zcrSD, specMean, specSD, and rmsMean
The features of zcrMean and zcrSD were calculated based on the zero-crossing rate.The zero-crossing rate is a commonly used feature in bioacoustic analysis, which measures the rate of signal changes from positive to zero and from negative to zero [22].The features of specMean and specSD were calculated based on the spectral centroid.The spectral centroid is another commonly used feature in signal analysis, which measures the gravity of the spectrum in a given signal period [12].The rmsMean feature was calculated based on the root mean square of the signal in each period.This feature is commonly used in signal analysis to measure the average power of a signal [23].For each event signal period, the zero-crossing rate, spectral centroid, and root mean square features were extracted and calculated using a short-time Fourier transform (STFT) with a Hann window, using 2048 samples with a 75% overlap.After extracting a series of these features for each event signal period, the zcrMean, specMean, and rmsMean were calculated based on the mean value of the zero-crossing rates, the mean value of the spectral centroids, and the mean value of the root mean squares, respectively, while the zcrSD and specSD were calculated based on the sample deviation values of the zero-crossing rates and spectral centroids, respectively.Figure 2 summarizes the process used to extract such features, showing how each signal was segmented into frames and multiple samples. of bioacoustic data.To make the system flexible, the number of events and classifiers can be modified to suit different researchers' interests.
In summary, the goal of this system was, given a set of bioacoustic recordings as the input, to output a summary of different events' patterns over time.

Feature Extraction
A set of features were extracted based on each period of events based on the bioacoustic signal.These features are the mean of zero-crossing rate (zcrMean), standard deviation of zero-crossing rate (zcrSD), mean of spectral centroid (specMean), standard deviation of spectral centroid (specSD), mean of root mean square (rmsMean), mel-frequency cepstral coefficients (MFCCs), and spectral entropy (spectralEntropy).

zcrMean, zcrSD, specMean, specSD, and rmsMean
The features of zcrMean and zcrSD were calculated based on the zero-crossing rate.The zerocrossing rate is a commonly used feature in bioacoustic analysis, which measures the rate of signal changes from positive to zero and from negative to zero [22].The features of specMean and specSD were calculated based on the spectral centroid.The spectral centroid is another commonly used feature in signal analysis, which measures the gravity of the spectrum in a given signal period [12].The rmsMean feature was calculated based on the root mean square of the signal in each period.This feature is commonly used in signal analysis to measure the average power of a signal [23].For each event signal period, the zero-crossing rate, spectral centroid, and root mean square features were extracted and calculated using a short-time Fourier transform (STFT) with a Hann window, using 2048 samples with a 75% overlap.After extracting a series of these features for each event signal period, the zcrMean, specMean, and rmsMean were calculated based on the mean value of the zerocrossing rates, the mean value of the spectral centroids, and the mean value of the root mean squares, respectively, while the zcrSD and specSD were calculated based on the sample deviation values of the zero-crossing rates and spectral centroids, respectively.Figure 2 summarizes the process used to extract such features, showing how each signal was segmented into frames and multiple samples.

Mel-Frequency Cepstral Coefficients (MFCCs)
Mel-frequency cepstral coefficients (MFCCs) are features in the bioacoustics domain [10,24,25].First, the STFT is calculated based on the raw signal.The Mel-frequency cepstrum (MFC), which represents the short-term energy, is then calculated by passing the STFT to the filter bank and calculating the real logarithm of the short-term energy spectrum at each of the mel-frequencies.Next, discrete cosine transform (DCT) is applied to each mel-frequency.The MFCC is thus obtained.The first 13 cepstral coefficients are used to describe the signal information.

Spectral Entropy (SpectralEntropy)
The feature of spectralEntropy is used to measure the spectral entropy of a signal's period.Spectral entropy is derived from information entropy [26], which measures the complexity of a signal.Based on the concept of information entropy, spectral entropy measures the complexity of spectral power distribution [27].It has been used in some acoustic classification problems [28,29].Specifically, spectral entropy is calculated as follows: Calculate the power spectral density (PSD) of the signal by squaring the amplitude normalized by the number of bins (1): where X(W i ) is the spectrum of the signal.Normalize the P(W i ) (2): Calculate the spectral entropy (3):

Recognize Events
There were two groups of target events we wanted to recognize: bird chirping and human/machine events.The 19 features extracted based on different periods of events were used as input to the classifiers to recognize these two events.The classifiers evaluated for this project were support vector machine (SVM) and multi-layer perceptron (MLP).Both of these classifiers have already been used in the bioacoustics domain [6,13,30,31] to perform species classification tasks.

Support Vector Machine
A support vector machine (SVM) is a commonly used classifier.The rationale of SVM is to map the data points to a higher dimension, and its classification is defined by finding the optimal hyperplane to separate the data points [32].

Multi-Layer Perceptron
Multi-layer perceptron (MLP) is a type of feedforward artificial neural network (ANN) that consists of an input layer, some hidden layers, and an output layer.In each layer, there are some neurons that can be activated by non-linear activation functions.This model is trained by following the backpropagation method using the gradient descent algorithm derived by Seiffert [33].

Proposed Framework for Pattern Summarization
We proposed a novel FDG (frequency-duration graph) framework to summarize the patterns of different events in a forest environment based on bioacoustic data over time.Given a bioacoustic signal as an input to the system, the output will show how frequently different events happen (F), the durations of different events (D), and a graphical analysis showing how different events are distributed over a given time period (G).
Frequency (F) calculates how many times a particular event happens in a given time period.Many events (such as bird chirping events, human activities, wind, and rain) may happen during a given time period.The frequencies of different events are important to researchers devoted to working on the interrelationships between different events.Monitoring the frequencies over different time periods may help in the study of how the patterns of events evolve over time.
Duration (D) calculates how long each event lasts in seconds.Combining duration with the frequency of each event, more features such as D/F can then be obtained for different events.D analysis provides information for studying the patterns related to how long events last over time.
Graphic (G) analysis provides a visualization tool to show how different events are distributed over the given time period.The x-axis represents the time and the y-axis represents different events.It is obvious how different events might evolve in a time-based graph.An important point is that G analysis shows the time-series features.Different events may happen sparsely during some time periods but densely in others.Based on the graphic analysis, more features can be derived to describe the patterns of events, which is meaningful for further study.

Case-Study-Based Evaluation
This section first introduces the dataset we used to evaluate the system.The SVM and MLP models were then applied to classify bird chirping and human events.Next, the patterns were summarized using the proposed FDG framework, and this summarization was analyzed to determine the FDG's usefulness.

Data Source
Bioacoustic data were collected from the Samford Ecologic Research Facility (SERF), which recorded bioacoustics information in Queensland, Australia over 4 days from October 13th, 2010 to October 16th, 2010.The dataset was 44.1 kHz, 128 kbps, and in MP3 format.The dataset utilized was also 4 days in length and was broken up into day-long recordings, each of which was around 1.28 GB in size.For convenience, the data were transformed to a 22.05 kHz WAV format for analysis.Some previous works have already done research based on this dataset [34][35][36][37].
The recordings are diverse, including the chirping of different birds and human-made sounds.Accordingly, this dataset was suitable for testing the proposed pattern recognition system.As there have been several works [18,25] already done on event detection itself, for the evaluation of the proposed recognition system, we assumed that the identified events were available to be input into the system.For the experiments, bird experts listened to the audio files and recorded the different events' start times and end times.We compared the expert´s annotations with the output of the FDG framework to assess its effectiveness.Table 1 shows a summary of the different activities present in the dataset.For this project, we aimed to classify bird chirping activity and human activity in the first stage, and summarize the event patterns in the second stage.

Event Classification Evaluation
A total of 19 features (zcrMean, zcrSD, specMean, specSD, spectralEntropy, rmsMean, and 13 MFCCs), explained in the previous section, were fed into the SVM and MLP models.Ten-fold cross-validation tests were conducted to evaluate the FDG framework.
Features were normalized using the min-max feature scaling approach before they were fed to the classification models.The SVM and MLP models were then used to classify different events.For the SVM model, the linear kernel was used and the penalty parameter C was set to 1.For the MLP model, the input layer had 19 neurons because of its 19 features, the hidden layer 1 had 12 neurons with the activation function set to "relu", the dropout layer rate was set to 0.2, hidden layer 2 had 8 neurons with the activation function set to "relu", and the output layer had one neuron with the activation function set to "sigmoid".The loss function used binary cross-entropy and its optimizer used "adam".
Ten-fold cross-validation tests were conducted on the NE20101013 data.The NE20101014 data were used for evaluation.The accuracy rate, precision rate, recall rate, and F 1 -score [38,39] were used to evaluate the system (4), ( 5), ( 6), (7): where TP stands for true positive, TN stands for true negative, FP stands for false positive, and FN stands for false negative.In this work, these variables referred to how closely the classification output compared to expert annotations.For example, for bird chirping, TP referred to the number of instances labeled as bird chirps by both the classifier and the expert annotator.

Pattern Summarization Evaluation
Figure 3 summarizes the whole system, including the FDG framework.In this framework, different classifiers were applied to perform event classification.Next, patterns were summarized by F (frequency), D (duration), and G (graph) over time.In this work, MLP and SVM models were applied to perform event classification in order to compare the two classifiers' qualities.To make the whole system more flexible, more classifiers could be added to the system for comparison, some of which might be better suited to different scenarios.

Figure 3.
The FDG (frequency-duration graph) framework.The pattern summarization has been discussed using the FDG framework on three case studies.

Event Classification
A 10-fold cross-validation test was used for both the SVM and MLP models.Table 2 and Table 3 show the classification results of the two models in scenarios with and without denoising.

Event Classification
A 10-fold cross-validation test was used for both the SVM and MLP models.Tables 2 and 3 show the classification results of the two models in scenarios with and without denoising.According to Tables 2 and 3, the overall accuracies for the two models with denoising and without denoising were very similar.However, in terms of precision, recall, and F 1 -Score, the MLP model performed better than the SVM model when denoising was not used.When using denoising, the SVM model showed higher precision but a lower recall and F 1 -score than the MLP model.Brown et al. [20] tried to use a high-pass filter to denoise the data to detect rain sounds and cicada sounds based on the same dataset.The results showed that denoising worked well for detecting rain sounds, but, in contrast, classification models were not sensitive to noise when it came to cicada chorus sound detection.Moreover, removing background noise did not improve the classification results.

Case Studies for Pattern Summarization Using FDG Framework
The FDG framework's summarization of patterns over 2 days (13 Oct 2011 and 14 Oct 2011) is shown in Table 4 and Figure 4.According to Table 2 and Table 3, the overall accuracies for the two models with denoising and without denoising were very similar.However, in terms of precision, recall, and F1-Score, the MLP model performed better than the SVM model when denoising was not used.When using denoising, the SVM model showed higher precision but a lower recall and F1-score than the MLP model.Brown et al. [20] tried to use a high-pass filter to denoise the data to detect rain sounds and cicada sounds based on the same dataset.The results showed that denoising worked well for detecting rain sounds, but, in contrast, classification models were not sensitive to noise when it came to cicada chorus sound detection.Moreover, removing background noise did not improve the classification results.

Case Studies for Pattern Summarization Using FDG Framework
The FDG framework's summarization of patterns over 2 days (13 Oct 2011 and 14 Oct 2011) is shown in Table 4 and Figure 4. Comparing the time period from 0 to 6:00:00 (Table 3 and Figure 4a) with the time period from 18:00:00 to 23:59:59 pm (Table 3 and Figure 4b) on 13 Oct 2010, it was concluded that the patterns were significantly different between the two periods.According to Table 3, the frequency of bird chirping in the morning (0 to 6:00:00) was around 11 times the frequency in the evening (18:00:00 to 23:59:59).The duration of bird chirping in the morning (1047 seconds) also lasted much longer than that in the evening (60 seconds).By contrast, human activities are not frequent in the morning (9 times) but are frequent in the evening (38 times).Moreover, the duration of human activities in the evening was 4605 seconds, which was significantly higher than the duration in the morning (187 seconds).It can be concluded that bird chirping and human activities have different patterns between morning and evening.The graphic analyses shown in Figure 4a and Figure 4b give clearer insight into different events' patterns over different time periods throughout a day.According to Figure 4a, in the morning, starting around 1:30:00, birds began to chirp, but these events appeared sparsely over

Case 1: Comparing Different Time Periods in One Day
Comparing the time period from 0 to 6:00:00 (Table 3 and Figure 4a) with the time period from 18:00:00 to 23:59:59 pm (Table 3 and Figure 4b) on 13 Oct 2010, it was concluded that the patterns were significantly different between the two periods.According to Table 3, the frequency of bird chirping in the morning (0 to 6:00:00) was around 11 times the frequency in the evening (18:00:00 to 23:59:59).The duration of bird chirping in the morning (1047 s) also lasted much longer than that in the evening (60 s).By contrast, human activities are not frequent in the morning (9 times) but are frequent in the evening (38 times).Moreover, the duration of human activities in the evening was 4605 s, which was significantly higher than the duration in the morning (187 s).It can be concluded that bird chirping and human activities have different patterns between morning and evening.The graphic analyses shown in Figure 4a,b give clearer insight into different events' patterns over different time periods throughout a day.According to Figure 4a, in the morning, starting around 1:30:00, birds began to chirp, but these events appeared sparsely over time.From around 4:30:00, bird chirping became very dense over time.Human activities seldomly appeared after around 4:00:00.According to Figure 4b, fewer birds chirped but more human activities appeared in the evening.Human activities happened densely around 20:30:00 and 21:50:00.Thanks to the FDG framework, at least some of the following information could be concluded.During a day (i.e., 13 Oct 2010), bird chirping and human activities have significantly different patterns between morning and evening times.Bird chirping happens more frequently and densely in the morning than in the evening.However, human activities show a contrasting pattern, as they happen more frequently starting at 19:00:00 in the evening than in the morning.

Case 2: Comparing Same Time Periods in Different Days
Comparing how different events evolve during the same time period for different days could provide valuable insights into how soundscapes change over long durations.Comparing Table 3 and Figure 4b with Table 3 and Figure 4d, it can be concluded that in the evening (18:00:00 to 23:59:59), two events showed very similar patterns on two different days.There was much less bird chirping activity during the evening on both days (56 times on 13 Oct 2011 and 49 times on The total time durations of the bird chirping were also similar between the two days, with 60 s on 13 Oct 2010 and 54 s on 14 Oct 2010.In addition, the graphic analysis shows that bird chirping activities were densely distributed around 18:00:00 on both days.In contrast, human activities lasted longer than bird chirping in the evenings on both days (4605 s on 13 Oct 2010 and 2403 s on 14 Oct 2010).In this case, it can be concluded that during the evening time, bird chirping and human activities had similar patterns over the two different days analysed.Bird chirping activities happened less in the evening, with some activity at around 18:00:00, while more human activities occurred in the evening.

Case 3: Temporal Correlation Analysis of Patterns over 4 Days
With the help of the FDG framework, some basic temporal analysis of the patterns as conducted.According to the duration and Ggraphic analysis, we assigned value 1 to the timepoints when the activities were happening, and value 0 to the timepoints when the activities were not happening.For example, for the period from 4:30:25 to 4:30:30, if the birds chirped from 4:30:26 to 4:30:28, but a human activity lasted from 4:30:25 to 4:30:27, then we can assign different values, as shown in Table 5.The Pearson correlation coefficient and its p-value were calculated based on 4 days' (from 13 Oct 2010 to 16 Oct 2010) data to analyze the temporal correlations between the two activities.Table 6 shows the results.
Table 6 shows that there was a very low temporal correlation between birds chirping and human activities.

Conclusions and Future Directions
To conclude, this study addressed the problem of comparing and summarizing different event patterns (bird chirping and human activities) over time by proposing a new system utilizing a novel frequency-duration graph (FDG) framework to automatically summarize how frequencies of different activities change over time.The proposed framework is flexible and allows users to add different events.For example, users might be interested in how weather changes over so that specific weather events, such as rain and wind, can be added to the system in the event classification stage.The FDG framework will generate detailed pattern summaries for these weather events over time to give researchers more insight into the environment.Finally, classification approaches are developing rapidly, but these new approaches can be easily incorporated into the classification process described here.
One future research direction could be to add more events of interests and evaluate the system further.In this work, we focused on analyzing the patterns of bird chirping activities and human activities over time.In recent years, climate change has also attracted much attention.Different climates might have different impacts on forest health and evolution.This work could be extended by including more climate phenomena able to be captured by bioacoustic data, from sources such as wind, thunder, and rain, to show how the patterns of these phenomena evolve over time.
The FDG framework could also be extended to different locations.Bioacoustic data collected from different locations could be processed by the proposed system and FDG framework to automatically compare patterns over time for different locations.For those who are experts in comparing different eco-environments, this extension work could be meaningful and helpful.Additionally, more temporal analyses, such as a concept drift analysis for different activity patterns, could be conducted in different eco-environments based on the FDG framework in future works.
One of the assumptions made in this paper was that some denoising has already been done and events have already been detected before using the proposed system.Even though there have been several works on event detection, new methods for when the environment is too noisy for current event detection methods must still be researched.

Figure 2 .
Figure 2. Extraction of the zero-crossing rate (zcrMean), standard deviation of zero-crossing rate (zcrSD), mean of spectral centroid (specMean), standard deviation of spectral centroid (specSD), and mean of root mean square (rmsMean), with 2048 samples each and a hop length of 512 samples.

Figure 3 .
Figure 3.The FDG (frequency-duration graph) framework.The pattern summarization has been discussed using the FDG framework on three case studies.

Table 1 .
Different activities in the dataset.

Table 2 .
Human activity and bird chirping activity classification results for the SVM and MLP models with the denoising technique applied.

Table 2 .
Human activity and bird chirping activity classification results for the SVM and MLP models with the denoising technique applied.

Table 3 .
Human activity and bird chirping activity classification results of the SVM and MLP models without the denoising technique applied.

Table 3 .
Human activity and bird chirping activity classification results of the SVM and MLP models without the denoising technique applied.

Table 5 .
Example of assigning 0 and 1 for the temporal correlation analysis of patterns.

Table 6 .
Temporal correlation analysis for patterns.