You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

21 November 2024

Automated Pipeline for Robust Cat Activity Detection Based on Deep Learning and Wearable Sensor Data

,
,
,
and
1
Institute of Digital Anti-Aging Healthcare, Inje University, Gimhae 50834, Republic of Korea
2
Department of Computer Engineering, Inje University, Gimhae 50834, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Sensing-Based Animal Biomechanics

Abstract

The health, safety, and well-being of household pets such as cats has become a challenging task in previous years. To estimate a cat’s behavior, objective observations of both the frequency and variability of specific behavior traits are required, which might be difficult to come by in a cat’s ordinary life. There is very little research on cat activity and cat disease analysis based on real-time data. Although previous studies have made progress, several key questions still need addressing: What types of data are best suited for accurately detecting activity patterns? Where should sensors be strategically placed to ensure precise data collection, and how can the system be effectively automated for seamless operation? This study addresses these questions by pointing out whether the cat should be equipped with a sensor, and how the activity detection system can be automated. Magnetic, motion, vision, audio, and location sensors are among the sensors used in the machine learning experiment. In this study, we collect data using three types of differentiable and realistic wearable sensors, namely, an accelerometer, a gyroscope, and a magnetometer. Therefore, this study aims to employ cat activity detection techniques to combine data from acceleration, motion, and magnetic sensors, such as accelerometers, gyroscopes, and magnetometers, respectively, to recognize routine cat activity. Data collecting, data processing, data fusion, and artificial intelligence approaches are all part of the system established in this study. We focus on One-Dimensional Convolutional Neural Networks (1D-CNNs) in our research, to recognize cat activity modeling for detection and classification. Such 1D-CNNs have recently emerged as a cutting-edge approach for signal processing-based systems such as sensor-based pet and human health monitoring systems, anomaly identification in manufacturing, and in other areas. Our study culminates in the development of an automated system for robust pet (cat) activity analysis using artificial intelligence techniques, featuring a 1D-CNN-based approach. In this experimental research, the 1D-CNN approach is evaluated using training and validation sets. The approach achieved a satisfactory accuracy of 98.9% while detecting the activity useful for cat well-being.

1. Introduction

Recognition of activity and the analysis of diseases of different types of pets and humans has been increasing simultaneously in recent years. For their cuteness and kindness, cats are the most popular choice for household or outdoor pets. Recently, primitive commercial services have started to record cats’ activities and report them to their owners or health instructors. It is vital to assess the activity of undisturbed animals to properly comprehend their ecosystem or offer indices of their welfare, but this is a difficult undertaking [1]. Multiple physical conditions and physiological factors may be monitored simultaneously with the position of the animals, due to the inclusion of numerous sensors in transmitters [2]. Due to recent advances in data loggers that capture body movements using acceleration signals, it is possible to develop and install a variety of animal sensors, allowing researchers to monitor varied behaviors [3,4,5]. In addition to direct observation by humans, the rapid development of digital information-processing technologies provides a lot of chances to investigate animal behavior through more accurate activity analysis [6]. There is no secret in the world of bio-signal processing, because understanding data from several sensors linked to one’s body results in high precision and accuracy [7,8,9,10,11]. Automated systems for the detection of behaviors have become increasingly popular in recent years, owing to the ability of the associated sensors in the automated systems to discriminate between various activity patterns. Research has demonstrated that combining data from multiple sensors can lead to more effective and efficient activity detection processes [12,13,14]. There are many aspects to consider while making decisions about one’s health, state of mind, and activity, and hence utilizing several body sensors is thought vital in this scenario. However, throughout the last decade, the use of various body sensors to resolve a specific decision was a point of contention due to a lack of resources and effective frameworks for multi-sensory data handling. The process of quick prototyping and deployment has grown considerably easier with the introduction of body sensor networks (BSNs) and business cloud architectures. Body activity types are important for the identification of behavior patterns.
Consequently, accelerometers, gyroscopes, and magnetometers are generally used in automated systems. Accelerometers are small sensors that use micro-machined structures to measure acceleration along three axes denoted by the letters X, Y, and Z. When the sensor moves in a specific direction, it returns a value in meters per second squared for how fast it traveled along all its axes. Behavior such as resting and walking can be identified efficiently using accelerometer sensor data. Cat Activity Recognition (CAR) is classifying the activity of a cat using responsive sensors that are affected by cat movement. The gyroscope’s ability to gauge rotational speed around a specific axis helps it continue to function as intended. In our study, the gyro sensor works with three axes (X, Y, and Z) to measure the activity rate, combined with the accelerometer. On the other hand, the magneto sensor also works with three axes to measure the angular rate, combined with the accelerometer. According to previous studies [15,16,17,18,19,20,21,22,23,24,25], systems are implemented and evaluated with the use of deep learning methods, namely, Artificial Neural Networks (ANNs), Long Short-Term Memory (LSTM), and a One-Dimensional Convolutional Neural Network (1D-CNN). Most of these past studies focused on traditional machine learning techniques for the activity detection of different animals using wearable accelerometers and some gyroscopes. Simultaneously, a few other researchers looked at accelerometer data for activity, but very few studies have looked at the combination of accelerometer, gyroscope, and magnetometer sensors for activity detection and classification. We opted to employ neck-wearable sensors to detect the activity patterns of domestic pets after gathering a lot of evidence from previous studies.
In our research, we attempted to create an automated system that analyzes data from neck sensors to classify activity patterns. For the performance comparison, we used standard deep learning approaches as well as incremental machine learning techniques. We also evaluated our system using validation sets of data, and the machine learning models provided the best prediction performance, indicating that our system has the potential to be employed in real-world scenarios. Finding a strong solution is the primary goal of this study, using deep learning algorithms for monitoring, detecting, and classifying the activities of domesticated cats, and carrying out an activity analysis of home-kept cats by leveraging the 1D-CNN-based deep learning approach. More specifically, we use a computer-assisted methodology for the classification of different types of activities.
Hence, the objectives of this study are as follows:
  • To develop an automated system that can accept accelerometer, gyroscope, and magnetometer multi-axis data as input and distinguish different cat activity patterns by applying a deep learning algorithm.
  • With the incorporation of multi-sensor devices on cats’ necks, to use feature engineering for best real-time accuracy.
  • To address imbalance issues by employing the class weight approach, and to carry out hyperparameter tuning to achieve the optimal performance of the model.
  • To test the automated system developed using a 1D-CNN model to accurately detect cat activities in real time.

3. Materials and Methods

In this section, we are going to describe the methodologies that are used in this research, including information about cats, sources of data, feeding and husbandry environment, and physical sensors.

3.1. Data Acquisition

3.1.1. Cats

The total number of cats was 10. Among those, 4 were males and 6 were females with different ages, sizes, and breeds. The cats were healthy and in good condition during the experimental period. Table 2 below shows the details about the experimental cats.
Table 2. Experimental cats summary.

3.1.2. Source of Data

The data were collected from 10 cats from 1 November 2021 to 30 November 2021. The dataset contained 1,284,789,349 samples of tri-axial data from accelerometer, gyroscope, and magnetometer sensors. We took the cats’ real-time data from Ujura Company (Seoul, Republic of Korea) and we used it for our experiment on cat activity detection.

3.1.3. Husbandry

The cats were guided by the expertise of husbandry and veterinary professionals; they designed a diet that catered to their health needs. Their living place was very clean, and a spacious environment with proper lighting was provided for all the cats. The rooms were 4.0 m × 3.5 m in size. To provide a better environment and to keep them active, the cats were provided with balls to play with. For the scratching of the cats, proper rugs were placed in each room. In every room, a box filled with water was provided for drinking purposes in case the cats were thirsty, and a different box was kept likewise for urinating and defecating. Figure 1 shows the experimental environment of the cats.
Figure 1. Housing, monitoring, and husbandry environment of the cats.

3.1.4. Sensor Device

The wearable devices consisted of three sensors, namely, accelerometer, gyroscope, and magnetometer sensors, and one device was placed on the neck of each cat. These sensors were able to measure the linear motions, rotational motions, and magnetic motions in all three axes, i.e., x, y, and z. The gyro sensor has a range of ±2000 DPS and a sampling rate of 0.001–100 Hz, and the accelerometer has a range of ±16 g and a sampling rate of 0.001–100 Hz. Likewise, the magnetometer has a range of ±1300 μT (x, y-axis), ±2500 μT (z-axis), and a sampling rate of 0.001–25 Hz. The devices weighed 0.2 oz and the dimensions were 27 mm × 27 mm × 4 mm. The sensor devices are designed to detect cat movements through rotational, linear, or magnetic motions. The data were timestamped for synchronization purposes. Each device has a Lipo battery with 70–100 mAH and a charging time of 2 h. Figure 2 shows the internal details of the sensor and the outfit of the sensor.
Figure 2. Wearable sensors with internal features.

3.1.5. Video Recording

The activities of cats were recorded through CCTV cameras as well. The collected data from the sensor devices were synchronized with the video for the ground truth and to ensure the correct labeling of different activities. Ten researchers classified the videos using manual classification, and their work was verified by the senior specialist and CTO from Ujura Company, Republic of Korea.

3.1.6. Data Collection

The data used in this research were collected with prior authorization from the cats’ owner (Ujura Company). The data were generated by the movement of the cats, and three kinds of data have been generated: linear motion data from the accelerometer, rotational motion data from the gyroscope, and magnetic motion data from the magnetometer sensors. The data were sent to the server via Bluetooth and were stored on the server. At the same time, video recording was also performed for the corresponding sensor data, and the sensor data were labeled while using the video recording. In Figure 3, we show all the steps of the data collection procedure.
Figure 3. Data collection procedure. (A) Server room for real-time monitoring and storing data, (B) sensor device, (C) sensor device on the cat’s neck, (D) cat living space, including surveillance cameras, (E) transferring sensor data to the server.
The data were processed and analyzed using a system with the following specifications: Windows 11, 2.50 GHz 64-bit, 12th Gen, Intel Core i5-12400 processor, 32 GB RAM, NVIDIA GeForce RTX 3080 GPU, Python 3.9, and TensorFlow 2.4.0.
Figure 4 shows that the data distribution across all the five classes is highly imbalanced, so this behavior of data may render the model overfitting. This issue must be addressed before the data are used for the training of the model. Figure 5 below represents a sample of the bio-signals obtained from the sensor devices attached to the cats.
Figure 4. Data distribution of activity detection.
Figure 5. Samples of bio-signals from the wearable devices on the cats.

3.1.7. Preprocessing of the Data

This process is vital because sensor data is often noisy, contains missing values, and requires proper formatting before it is fed it into machine learning models for activity detection tasks [46]. We take the raw sensor data, which is initially labeled, and then proceed to enhance its quality and prepare it for training in an artificial intelligence model. This involves crucial steps to handle noise and anomalies that might be present in the data, particularly in the bio-signals reflecting the cats’ activities. To address the issue of noisy data, we employ the Butterworth low-pass filter, effectively eliminating unwanted high-frequency noise components while preserving the essential bio-signals. This ensures that the actual activities of the cat are accurately represented in the processed data. By refining and cleansing the data through these smart preprocessing techniques, we have developed a robust and accurate dataset suitable for training the machine learning model. This essential step significantly enhances the model’s ability to interpret the cats’ activities and paves the way for more reliable artificial intelligence predictions.

3.1.8. Feature Engineering

To extract valuable insights from the sensor data, we applied feature engineering techniques. By analyzing accelerometer, gyroscope, and magnetometer data, we derived key features, most important features such as standard deviation, mean absolute deviation, mean, minimum, maximum, interquartile range, energy measure, skewness, kurtosis, etc. Zheng et al. have presented windowing techniques for activity detection using sensor devices, which involve segmenting the continuous stream of sensor data into smaller time windows or frames [47]. This windowing technique was used to divide the data into 2 s windows with 50 data samples per window, overlapping 25 samples from the previous window. This allowed us to create new features while preserving temporal information. The transformed data were labeled based on the most frequent activity within each sliding window. We applied a fast Fourier transform to convert time domain data into the frequency domain, gaining deeper insights and enriching the dataset. In the end, our feature engineering efforts yielded 312 informative features. Kempa-Liehr et al. showed that the goal of feature engineering is to provide algorithms with informative input features that capture the essential patterns and characteristics of the activities, enabling accurate and efficient activity detection and classification [48,49].

3.1.9. Class Weight Approach

There are different approaches for balancing the data, for example, Random Oversampling, the synthetic minority oversampling technique (SMOTE), Adaptive Synthetic Sampling (ADASYN), etc. [50]. To keep a balance among the classes, a threshold should be defined so that class weights can be increased or decreased. With this technique, we take more care of the minority samples while training the model and, to calculate the loss function, a weighting mechanism is developed. Different weights are assigned to majority and minority classes according to the imbalance scenario in the dataset. To keep a balance among the classes, a threshold should be defined so that class weights can be increased or decreased. This will help in preventing the biasing of the algorithm towards any specific class. The formula for class weight can be defined as
W i = n _ i n s t a n c e s     ( n _ c l a s s e s     n _ i n s t a n c e s i )
where ( W i ) represents the weight of each class and (i) represents the class. n_instances denotes the total number of instances or rows in our dataset, whereas n_classes represents the overall number of unique classes in the class label. The total number of rows in each class is denoted as n _ i n s t a n c e s i . The weighting mechanism adopted in this study is shown in Table 3 below.
Table 3. Class weights for activity detection model training.

3.2. Methods

3.2.1. Proposed Activity Detection Algorithm

A One-Dimensional Convolutional Neural Network is a powerful algorithm for activity detection based on sensor data from wearable devices [51,52]. A proposed pet activity detection algorithm was developed, which included the collection of bio-signals from wearable devices, i.e., an accelerometer, a gyroscope, and a magnetometer. The bio-signals were preprocessed by applying different preprocessing techniques like data filtration, data normalization, etc. The activities of the cats were predicted using the CNN-based algorithm as a well-known deep learning approach used for different purposes like classification and detection. It automatically extracts the highly relevant features without any human intervention or handcrafted methods and uses those features for classification and detection purposes. We used sensor data consisting of x, y, and z values, and we transformed them into vector magnitude data. This vector was used to develop a 1D-CNN for the classification of the different activities of the cats.

3.2.2. The Network Architecture of 1D-CNN

Convolutional Neural Networks (CNNs) have garnered widespread acclaim in the realm of deep learning due to their remarkable capabilities in various applications, such as classification and detection tasks. Unlike traditional Artificial Neural Networks, CNNs possess a unique ability to autonomously extract essential features from data without relying on manual intervention or handcrafted methods. This intrinsic feature extraction capability empowers CNNs to excel in both feature extraction and subsequent classification processes. In our research, we harnessed the potential of CNNs to explore the activities of cats using sensor data that comprised x, y, and z values. To facilitate a more efficient representation of the data, we transformed it into vector magnitude data. This transformed vector served as the basis for the development of a sophisticated 1D-CNN model tailored specifically for the classification of different feline activities. Our 1D-CNN architecture was thoughtfully designed, encompassing key components such as convolutional layers to learn intricate patterns, dropout layers for enhanced generalization and robustness, flattened layers to reshape the data for seamless processing, fully connected layers to establish meaningful connections, and SoftMax layers for accurate probability distribution across the various cat activities. By leveraging this comprehensive 1D-CNN model, we were able to gain deeper insights into the distinct activities exhibited by cats, paving the way for future advancements in understanding and analyzing animal behavior. The results of our study not only underscore the effectiveness of CNNs in the realm of activity classification but also shed light on the tremendous potential of deep learning architectures for pushing the boundaries of knowledge discovery and problem-solving in diverse domains.
  • Input layer: The input layer of the model received three-axis data from each accelerometer, gyroscope, and magnetometer sensor in the form of vector magnitude.
  • Convolutional layer: The convolutional operations were used with a stride size of 1. The kernels used in the convolutional layers were 128, 128, 128, 256, and 256, while the strides were kept at 1 in each layer.
  • Dropout: To avoid overfitting and to reduce the complexity of the model, dropout layers were used while the dropout value was set to 0.5.
  • Output: In deep learning, activation functions play an important role in the prediction of any task. The right and wise choice of activation function results in good prediction. Rectified Linear Unit (ReLu) was used in this experiment. Since we had activities from five cats, which is a multiclass classification, we used the SoftMax function for the classification of all five activities. A Stochastic Gradient Descent (SGD) optimizer was applied, and the learning rate was set to 0.0001. Categorical cross-entropy was used as a loss function, which calculates the loss between the actual and predicted values. The smaller the difference between the values, the higher the performance of the model. Figure 6 illustrates the architecture of the model and Figure 7 shows the classification of the activities.
    Figure 6. The deep learning model architecture of our experimental research work.
    Figure 7. Classification of the five activities.

3.2.3. The Proposed Research

In our approach to detecting cat activities using data from wearable sensor devices, the overall process can be summarized as follows:
Firstly, we meticulously extracted data pertaining to five distinct activities and concurrently captured corresponding videos, synchronizing them at a precise frame rate for each activity. Subsequently, we undertook rigorous data preprocessing, effectively eliminating noise and unwanted bio-signals from the dataset. Employing a Butterworth low-pass filter, we successfully eliminated noise, thereby enhancing the dataset’s quality. To harness the full potential of the dataset, we engaged in feature engineering, a critical step that extracted pertinent information while discarding unnecessary data. This strategic maneuver facilitated the construction of an efficient algorithm, primed for accurate activity classification. Furthermore, we recognized the importance of data normalization to ensure that all data points fell within the same value range. The raw data extracted from the sensor devices were preprocessed and filtered by applying a Butterworth low-pass filter. The filters removed the noise and unwanted signals from the data and, as a result, we obtained refined data. Data normalization was applied to the dataset to normalize the range of all the data and bring them to the same scale. This process played a crucial role in optimizing the subsequent stages of our analysis. To evaluate the performance of our model, we diligently split the data into an 80% training set and, after preprocessing the dataset, a 20% testing set. Acknowledging the data’s inherent imbalance, we employed data oversampling techniques on the training dataset, ensuring a balanced representation of all classes. We leveraged the class weight technique, further enhancing the model’s sensitivity to minority classes during training. Next, we developed a sophisticated 1D-CNN model and diligently trained it using the class-weighted training dataset. As we prioritized performance, we continuously monitored the model during hyperparameter tuning, ensuring that any potential issues were swiftly addressed. The culmination of our efforts yielded promising results. The experimental outcomes showcased the model’s exceptional performance, with the class weight technique playing a pivotal role in enhancing accuracy and reliability. Overall, our approach not only demonstrated the efficacy of the 1D-CNN architecture but also highlighted the importance of thoughtful data preprocessing and the strategic handling of class imbalances in activity classification tasks. Figure 8 shows the complete process of the development of the activity detection systems for household pets.
Figure 8. The complete process of the automated pipeline.

4. Experimental Results with Discussion

The experimental results are discussed in detail in this section. We conducted experiments using class weights for our class labels to balance the activities of the cats.

4.1. Evaluation Methods

The performance of the model is based on accuracy, precision, recall, F-score, and ROC.
Accuracy: accuracy states how close our nearest value is to the known value:
A c c u r a c y = T P + T N T P + T N + F P + F N × 100 %
Precision: this is the fraction of relevant instances among the retrieved instances:
P r e c i s i o n = T P T P + F P × 100 %
Recall: this is the fraction of relevant instances that were retrieved:
R e c a l l = T P T P + F N × 100 %
F1-Score: this is a way of combining the precision and recall of the model, and it is defined as the harmonic mean of the model’s precision and recall:
F - s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l × 100 %
where TP represents a true positive, TN is a true negative, FN is a false negative, and FP is a false positive. Precision indicates the degree of accuracy of the model in predicting the correct classification of activities. For instance, eating was positive, and all other activities of the cats were negative. In this scenario, the correct classification of eating is divided by the sum of the correct classification, and the incorrect classification of eating gives the precision value.

4.2. Evaluation Methods (Graphical)

The prediction outcome of a classification model is summarized in the confusion matrix. By displaying the number of predictions that were accurate and inaccurate for each class, it displays the model’s performance. It presents details regarding the model’s actual and expected classifications. The model properly predicts the values in the diagonals while misclassifying the values outside of the diagonals.

4.2.1. Matrix

The confusion matrices for all five of the behaviors of the cats studied in our research utilizing class weight are shown in Figure 9 and Figure 10, respectively, with and without normalization.
Figure 9. Confusion matrix without normalization using the test dataset.
Figure 10. Confusion matrix with normalization using the test dataset.

4.2.2. Accuracy and Loss

Figure 11 and Figure 12 illustrate, respectively, the accuracy and loss of the model. Figure 11 demonstrates how the training accuracy rose to 98.9% after 700 epochs and 96.85% of the validations were accurate.
Figure 11. Accuracy graph for the validation and training.
Figure 12. Loss graph for the validation and training.

4.2.3. AUC and RUC Curve

The AUC-ROC curve helps us to visualize the performance of our proposed model. In other words, it is a measurement of evaluation that displays each class’s performance while drawing a graph between the true positive rate (TPR) and the false positive rate (FPR). The AUC-ROC is displayed in Figure 13. The better the model performs, the closer its graph is to the left corner and close to value 1. The graph below demonstrates that all the curves for each class are closer to 1, indicating that the model is performing at almost 100%. The AUC value of our suggested model is 100% for all five classes, demonstrating that our model correctly distinguishes between positive and negative class points.
Figure 13. Receiver operating characteristic (ROC) curves and AUCs for each class.

5. Discussion

We have developed a highly efficient activity detection pipeline specifically designed for monitoring household pets through the utilization of wearable sensors attached to the neck. The pipeline incorporates data from accelerometers, gyroscopes, and magnetometers to extract various features. These features serve as input for accurately recognizing different activities, resulting in low misclassification rates.
Our findings underscore the intricacies involved in detecting the behavior of household pets. Notably, similar features extracted from diverse signal types prove effective in identifying various activities. Importantly, our model demonstrates impeccable performance. Although the accelerometer, gyroscope, and magnetometer data-based model has not been extensively studied for animals, a few research studies on humans, using different state-of-the-art techniques, yielded similar results to ours, establishing our proposed pipeline as state-of-the-art for animals, especially for cat activity detection.
This study introduces several novelties: (1) This is the first study to incorporate data from three different types of neck-wearable sensors—accelerometric, gyroscopic, and magnetometers for detecting activity patterns. This approach is particularly recommended for real-life situations involving pets as their activities, such as body movements involving the neck, are accurately captured using both types of sensors. (2) This study presents a robust pipeline utilizing an incremental learning-based deep learning algorithm for activity detection. The proposed activity detection system in this study outperforms previous research efforts. A comprehensive comparison with state-of-the-art models for activity detection is detailed in Table 4. We trained several machine learning and deep learning models; from the trained models, the 1D-CNN, the ANN, and LSTM have given the best results and, among these three, the 1D-CNN gave the best accuracy, as we have shown in Table 4. A 1D-CNN is ideal for cat activity detection as it excels in capturing temporal patterns in sequential sensor data, providing effective recognition of localized behaviors, and allows the efficient processing of wearable sensor information.
Table 4. Class weights of three models for activity detection model training.
In this study, we collected data from a limited number of cats (10 cats). In the future, we will use more data for model training. This time, we worked on household cats; in the future, we will work on outdoor cats who live without limitations or borders. Also, in the future, we intend to focus on investigating activities separately and adding many more activities.

6. Conclusions

We introduced a 1D-CNN-based automated activity detection system to predict and classify five fundamental activities of cats. The data collected from cats of various ages, breeds, and genders were meticulously preprocessed and transformed into a suitable format for model training. Through feature engineering, we extracted the most meaningful and essential features from the raw data. Our proposed pipeline for detecting cat activities enables efficient monitoring of their well-being and overall health. This groundbreaking research proposes a 1D-CNN-based approach for cat activity detection, utilizing data from multiple sensors. We achieved a high accuracy of 98.9%, due to the integration of multiple sensors, which captures a comprehensive view of cat movements. This system holds great promise for enhancing our understanding of feline behavior and ensuring the welfare of our feline companions.

Author Contributions

Conceptualization, M.A.I.M.; methodology, M.A.I.M., T.P.T.A. and R.I.S.; software, S.M.I.U.; validation, M.A.I.M., T.P.T.A. and S.M.I.U.; formal analysis, M.A.I.M. and T.P.T.A.; investigation, H.-C.K.; resources, M.A.I.M.; data curation, M.A.I.M.; writing—original draft preparation, M.A.I.M.; writing—review and editing, R.I.S., T.P.T.A. and S.M.I.U.; visualization, M.A.I.M.; supervision, H.-C.K.; project administration, H.-C.K.; funding acquisition, H.-C.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science ICT), Korea, under the National Program for Excellence in SW, supervised by the IITP (Institute of Information and Communications Technology Planning and Evaluation) in 2022 (2022-0-01091, 1711175863).

Data Availability Statement

The internal dataset analyzed for this study is not publicly available at this moment. Data can be given upon request to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Watanabe, S.; Izawa, M.; Kato, A.; Ropert-Coudert, Y.; Naito, Y. A new technique for monitoring the detailed behavior of terrestrial animals: A case study with the domestic cat. Appl. Anim. Behav. Sci. 2005, 94, 117–131. [Google Scholar] [CrossRef]
  2. Gillingham, M.P.; Bunnell, F.L. Reliability of motion-sensitive radio collars for estimating the activity of black-tailed deer. J. Wildl. Manag. 1985, 49, 951–958. [Google Scholar] [CrossRef]
  3. Tanaka, H.; Takagi, Y.; Naito, Y. Swimming speeds and buoyancy compensation of migrating adult chum salmon Oncorhynchus keta revealed by speed/depth/acceleration data logger. J. Exp. Biol. 2001, 204, 3895–3904. [Google Scholar] [CrossRef] [PubMed]
  4. Crowell-Davis, S.L. Understanding Cats; COMPENDIUM; The University of Georgia: Athens, GA, USA, 2007. [Google Scholar]
  5. Jalal, A.; Quaid, M.A.K.; Tahir, S.B.U.D.; Kim, K. A study of accelerometer and gyroscope measurements in physical life-log activities detection systems. Sensors 2020, 20, 6670. [Google Scholar] [CrossRef] [PubMed]
  6. Lima, W.S.; Souto, E.; El-Khatib, K.; Jalali, R.; Gama, J. Human activity recognition using inertial sensors in a smartphone: An overview. Sensors 2019, 19, 3213. [Google Scholar] [CrossRef]
  7. Baratchi, M.; Meratnia, N.; Havinga, P.J.M.; Skidmore, A.K.; Toxopeus, B.A.G. Sensing solutions for collecting spatio-temporal data for wildlife monitoring applications: A review. Sensors 2013, 13, 6054–6088. [Google Scholar] [CrossRef] [PubMed]
  8. Zhu, J.; San-Segundo, R.; Pardo, J.M. Feature extraction for robust physical activity recognition. Hum.-Centric Comput. Inf. Sci. 2017, 7, 16. [Google Scholar] [CrossRef]
  9. Alonso-Español, A.; Bravo, E.; Ribeiro-Vidal, H.; Virto, L.; Herrera, D.; Alonso, B.; Sanz, M. Prying into the intimate details of animal lives: Use of a daily diary on animals. Endanger. Species Res. 2008, 4, 123–137. [Google Scholar]
  10. Foster, M.; Mealin, S.; Gruen, M.; Roberts, D.L.; Bozkurt, A. Preliminary evaluation of a wearable sensor system for assessment of heart rate, heart rate variability, and activity level in working dogs. In Proceedings of the 2019 IEEE SENSORS, Montreal, QC, Canada, 27–30 October 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar] [CrossRef]
  11. Lemasson, G.; Lucidarme, P.; Duhaut, D. Real-time detection of the activity of a dog. In Nature-Inspired Mobile Robotics; World Scientific: Singapore, 2013; pp. 815–821. [Google Scholar] [CrossRef]
  12. Pires, I.M.; Garcia, N.M.; Pombo, N.; Flórez-Revuelta, F.; Spinsante, S. Data Fusion on Motion and Magnetic Sensors embedded on Mobile Devices for the Identification of Activities of Daily Living. arXiv 2017, arXiv:1711.07328. [Google Scholar] [CrossRef]
  13. Liu, X.; Wang, Q.; Liu, D.; Wang, Y.; Zhang, Y.; Bai, O.; Sun, J. Human emotion classification based on multiple physiological signals by wearable system. Technol. Health Care 2018, 26, 459–469. [Google Scholar] [CrossRef]
  14. Masum, A.K.M.; Bahadur, E.H.; Shan-A-Alahi, A.; Chowdhury, A.U.Z.; Uddin, M.R.; Al Noman, A. Human activity recognition using accelerometer, gyroscope and magnetometer sensors: Deep neural network approaches. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar] [CrossRef]
  15. Khan, A.; Berlin, E.; Mellor, S.; Thompson, R.; Hammerla, N.; McNaney, R.; Olivier, P.; Plötz, T. How did I do? Automatic Skill Assessment from Activity Data. GetMobile Mob. Comput. Commun. 2016, 19, 18–22. [Google Scholar] [CrossRef]
  16. Hooker, S.K.; Biuw, M.; McConnell, B.J.; Miller, P.J.; Sparling, C.E. Bio-logging science: Logging and relaying physical and biological data using animal-attached tags. Deep.-Sea Res. Part II 2007, 54, 177–182. [Google Scholar] [CrossRef]
  17. Guan, Y.; Plötz, T. Ensembles of deep lstm learners for activity recognition using wearables. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2017, 1, 1–28. [Google Scholar] [CrossRef]
  18. Quaranta, A.; D’ingeo, S.; Amoruso, R.; Siniscalchi, M. Emotion recognition in cats. Animals 2020, 10, 1107. [Google Scholar] [CrossRef] [PubMed]
  19. Ladha, C.; Hammerla, N.; Hughes, E.; Olivier, P.; Ploetz, T. Dog’s life: Wearable activity recognition for dogs. In Proceedings of the 2013 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Zurich, Switzerland, 8–12 September 2013. [Google Scholar] [CrossRef]
  20. Kumpulainen, P.; Cardó, A.V.; Somppi, S.; Törnqvist, H.; Väätäjä, H.; Majaranta, P.; Gizatdinova, Y.; Antink, C.H.; Surakka, V.; Kujala, M.V.; et al. Dog behaviour classification with movement sensors placed on the harness and the collar. Appl. Anim. Behav. Sci. 2021, 241, 105393. [Google Scholar] [CrossRef]
  21. Kumpulainen, P.; Valldeoriola, A.; Somppi, S.; Törnqvist, H.; Väätäjä, H.; Majaranta, P.; Surakka, V.; Vainio, O.; Kujala, M.V.; Gizatdinova, Y.; et al. Dog activity classification with movement sensor placed on the collar. In Proceedings of the Fifth International Conference on Animal-Computer Interaction, Atlanta, GA, USA, 4–6 December 2018. [Google Scholar] [CrossRef]
  22. Hammerla, N.Y.; Halloran, S.; Plötz, T. Deep, convolutional, and recurrent models for human activity recognition using wearables. arXiv 2016, arXiv:1604.08880. [Google Scholar] [CrossRef]
  23. Chambers, R.D.; Yoder, N.C.; Carson, A.B.; Junge, C.; Allen, D.E.; Prescott, L.M.; Bradley, S.; Wymore, G.; Lloyd, K.; Lyle, S. Deep learning classification of canine behavior using a single collar-mounted accelerometer: Real-world validation. Animals 2021, 11, 1549. [Google Scholar] [CrossRef]
  24. Crowell-Davis, S.L. Cat behavior: Social organization, communication, and development. In The Welfare of Cats; Springer: Dordrecht, Switzerland, 2007; pp. 1–22. [Google Scholar] [CrossRef]
  25. Khan, D.; Al Mudawi, N.; Abdelhaq, M.; Alazeb, A.; Alotaibi, S.S.; Algarni, A.; Jalal, A. A wearable inertial sensor approach for locomotion and localization recognition on physical activity. Sensors 2024, 24, 735. [Google Scholar] [CrossRef]
  26. Alphonce, E.; Kisangiri, M.; Kaijage, S.; Seshaiyer, P. Design and Analysis of smart sensing system for animal emotions recognition. Int. J. Comput. Appl. 2017, 169, 46–50. [Google Scholar] [CrossRef]
  27. Hussain, A.; Begum, K.; Armand, T.P.T.; Mozumder, A.I.; Ali, S.; Kim, H.C.; Joo, M.-I. Long Short-Term Memory (LSTM)-Based Dog Activity Detection Using Accelerometer and Gyroscope. Appl. Sci. 2022, 12, 9427. [Google Scholar] [CrossRef]
  28. Sanhudo, L.; Calvetti, D.; Martins, J.P.; Ramos, N.M.; Mêda, P.; Gonçalves, M.C.; Sousa, H. Activity classification using accelerometers and machine learning for complex construction worker activities. J. Build. Eng. 2020, 35, 102001. [Google Scholar] [CrossRef]
  29. Aich, S.; Chakraborty, S.; Sim, J.-S.; Jang, D.-J.; Kim, H.-C. The design of an automated system for the analysis of the activity and emotional patterns of dogs with wearable sensors using machine learning. Appl. Sci. 2019, 9, 4938. [Google Scholar] [CrossRef]
  30. Maza, S.; Buchert, R.; Brenner, W.; Munz, D.L.; Thiel, E.; Korfel, A.; Kiewe, P. Brain and whole-body FDG-PET in diagnosis, treatment monitoring and long-term follow-up of primary CNS lymphoma. Radiol. Oncol. 2013, 47, 103–110. [Google Scholar] [CrossRef] [PubMed]
  31. Chakraborty, S.; Aich, S.; Joo, M.-I.; Sain, M.; Kim, H.-C. A multichannel convolutional neural network architecture for the detection of the state of mind using physiological signals from wearable devices. J. Healthc. Eng. 2019, 2019, 5397814. [Google Scholar] [CrossRef] [PubMed]
  32. Kiyohara, T.; Orihara, R.; Sei, Y.; Tahara, Y.; Ohsuga, A. Activity Recognition for Dogs Using Off-the-Shelf Accelerometer. In Proceedings of the International Conference on Agents and Artificial Intelligence, Lisbon, Portugal, 10–12 January 2015; SCITEPRESS: Setúbal, Portugal, 2015; Volume 2. [Google Scholar] [CrossRef]
  33. Vehkaoja, A.; Somppi, S.; Törnqvist, H.; Cardó, A.V.; Kumpulainen, P.; Väätäjä, H.; Majaranta, P.; Surakka, V.; Kujala, M.V.; Vainio, O. Description of movement sensor dataset for dog behavior classification. Data Brief 2022, 40, 107822. [Google Scholar] [CrossRef]
  34. Hussain, A.; Ali, S.; Joo, M.I.; Kim, H.C. A deep learning approach for detecting and classifying cat activity to monitor and improve cat’s well-being using accelerometer, gyroscope, and magnetometer. IEEE Sens. J. 2023, 24, 1996–2008. [Google Scholar] [CrossRef]
  35. Hussain, A.; Ali, S.; Abdullah; Kim, H.-C. Activity Detection for the Wellbeing of Dogs Using Wearable Sensors Based on Deep Learning. IEEE Access 2022, 10, 53153–53163. [Google Scholar] [CrossRef]
  36. Venkatraman, S.; Long, J.D.; Pister, K.S.; Carmena, J.M. Wireless inertial sensors for monitoring animal behavior. In Proceedings of the 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Lyon, France, 22–26 August 2007. [Google Scholar] [CrossRef]
  37. Yen, C.-T.; Liao, J.-X.; Huang, Y.-K. Human daily activity recognition performed using wearable inertial sensors combined with deep learning algorithms. IEEE Access 2020, 8, 174105–174114. [Google Scholar] [CrossRef]
  38. Mao, A.; Huang, E.; Wang, X.; Liu, K. Deep learning-based animal activity recognition with wearable sensors: Overview, challenges, and future directions. Comput. Electron. Agric. 2023, 211, 108043. [Google Scholar] [CrossRef]
  39. Minati, L.; Li, C.; Bartels, J.; Chakraborty, P.; Li, Z.; Yoshimura, N.; Frasca, M.; Ito, H. Accelerometer time series augmentation through externally driving a non-linear dynamical system. Chaos Solitons Fractals 2023, 168, 113100. [Google Scholar] [CrossRef]
  40. Decandia, M.; Giovanetti, V.; Molle, G.; Acciaro, M.; Mameli, M.; Cabiddu, A.; Cossu, R.; Serra, M.; Manca, C.; Rassu, S.; et al. The effect of different time epoch settings on the classification of sheep behaviour using tri-axial accelerometry. Comput. Electron. Agric. 2018, 154, 112–119. [Google Scholar] [CrossRef]
  41. Rahman, A.; Smith, D.; Little, B.; Ingham, A.; Greenwood, P.; Bishop-Hurley, G. Cattle behaviour classification from collar, halter, and ear tag sensors. Inf. Process. Agric. 2018, 5, 124–133. [Google Scholar] [CrossRef]
  42. Hammond, T.T.; Springthorpe, D.; Walsh, R.E.; Berg-Kirkpatrick, T. Using accelerometers to remotely and automatically characterize behavior in small animals. J. Exp. Biol. 2016, 219, 1618–1624. [Google Scholar] [CrossRef] [PubMed]
  43. Pan, Z.; Chen, H.; Zhong, W.; Wang, A.; Zheng, C. A CNN-Based Animal Behavior Recognition Algorithm for Wearable Devices. IEEE Sens. J. 2023, 23, 5156–5164. [Google Scholar] [CrossRef]
  44. Arablouei, R.; Wang, Z.; Bishop-Hurley, G.J.; Liu, J. Multimodal sensor data fusion for in-situ classification of animal behavior using accelerometry and GNSS data. Smart Agric. Technol. 2023, 4, 100163. [Google Scholar] [CrossRef]
  45. Shahbazi, M.; Mohammadi, K.; Derakhshani, S.M.; Koerkamp, P.W.G.G. Deep Learning for Laying Hen Activity Recognition Using Wearable Sensors. Agriculture 2023, 13, 738. [Google Scholar] [CrossRef]
  46. Chakravarty, P.; Cozzi, G.; Ozgul, A.; Aminian, K. A novel biomechanical approach for animal behavior recognition using accelerometers. Methods Ecol. Evol. 2019, 10, 802–814. [Google Scholar] [CrossRef]
  47. Zheng, X.; Wang, M.; Ordieres-Meré, J. Comparison of data preprocessing approaches for applying deep learning to human activity recognition in the context of industry 4.0. Sensors 2018, 18, 2146. [Google Scholar] [CrossRef]
  48. Banos, O.; Galvez, J.-M.; Damas, M.; Pomares, H.; Rojas, I. Window size impact in human activity recognition. Sensors 2014, 14, 6474–6499. [Google Scholar] [CrossRef]
  49. Alonso-Español, A.; Bravo, E.; Ribeiro-Vidal, H.; Virto, L.; Herrera, D.; Alonso, B.; Sanz, M. Feature engineering workflow for activity recognition from synchronized inertial measurement units. In Proceedings of the Pattern Recognition: ACPR 2019 Workshops, Auckland, New Zealand, 26 November 2019; Proceedings 5. Springer: Singapore, 2020. [Google Scholar]
  50. Mozumder, M.A.I.; Sumon, R.I.; Uddin, S.M.I.; Kim, H.C. Domesticated cat activity detection using deep learning technique. In Proceedings of the IEEE 2024 International Conference on Electrical, Computer and Energy Technologies (ICECET), Sydney, Australia, 25–27 July 2024. [Google Scholar]
  51. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  52. Muralidharan, K.; Ramesh, A.; Rithvik, G.; Prem, S.; Reghunaath, A.A.; Gopinath, M. 1D Convolution approach to human activity recognition using sensor data and comparison with machine learning algorithms. Int. J. Cogn. Comput. Eng. 2021, 2, 130–143. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.