You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

26 October 2024

MPE-HRNetL: A Lightweight High-Resolution Network for Multispecies Animal Pose Estimation

,
,
and
1
School of Software, Henan Polytechnic University, Jiaozuo 454000, China
2
Anyang Institute of Technology, Anyang 455000, China
3
School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo 454000, China
4
CSSC Haifeng Aviation Technology Co., Ltd., No.4, Xinghuo Road, Fengtai Science Park, Beijing 100070, China
This article belongs to the Section Sensor Networks

Abstract

Animal pose estimation is crucial for animal health assessment, species protection, and behavior analysis. It is an inevitable and unstoppable trend to apply deep learning to animal pose estimation. In many practical application scenarios, pose estimation models must be deployed on edge devices with limited resource. Therefore, it is essential to strike a balance between model complexity and accuracy. To address this issue, we propose a lightweight network model, i.e., MPE-HRNet . L , by improving Lite-HRNet. The improvements are threefold. Firstly, we improve Spatial Pyramid Pooling-Fast and apply it and the improved version to different branches. Secondly, we construct a feature extraction module based on a mixed pooling module and a dual spatial and channel attention mechanism, and take the feature extraction module as the basic module of MPE-HRNet . L . Thirdly, we introduce a feature enhancement stage to enhance important features. The experimental results on the AP-10K dataset and the Animal Pose dataset verify the effectiveness and efficiency of MPE-HRNet . L .

1. Introduction

Animal pose estimation aims to identify and locate the key points of animal bodies. It plays a crucial role in assessing animal health [], protecting endangered species [], and analyzing animal behavior []. With the advancement of deep learning technology, it has been an inevitable trend to apply the technology to animal pose estimation. Deep-learning-based solutions can be categorized into bottom-up solutions and top-down solutions []. The bottom-up solutions detect key points and then cluster the key points to generate specific poses [,]. Although these solutions are relatively fast, they cannot guarantee accuracy in complex scenes. Different from the bottom-up solutions, the top-down solutions [,,] identify individual animal objects, and then estimate the poses of these objects, yielding higher accuracy [,].
A lot of effort has been devoted to the top-down solutions [,]. For example, some researchers utilize large network models, such as Hourglass [] and HRNet [], to achieve higher accuracy at the cost of high complexity [,,,,,,]. In many practical application scenarios, pose estimation models must be deployed on edge devices. Edge devices are always equipped with limited resource. They are unable to leverage the advantages of large models since they cannot provide the resource that large models need. Inspired by ShuffleNet [], Yu et al. [] proposed a lightweight version of HRNet, i.e., Lite-HRNet. Lite-HRNet can estimate the poses of human beings with relatively high speed and accuracy. However, it shows an obvious drop in accuracy when estimating the poses of animals of multiple species. That happens because of two reasons. The first reason is that animals of different species possess different anatomical structures, resulting in significant variations in their poses. The second reason is that Lite-HRNet sacrifices accuracy in favor of a light network structure.
In this paper, we propose a lightweight network to estimate the poses of multiple species on edge devices with higher accuracy. We name the lightweight network MPE-HRNet . L and construct it by improving Lite-HRNet from three aspects. Firstly, we improve Spatial Pyramid Pooling-Fast (SPPF) and propose SPPF . + . We apply SPPF and SPPF . + to different branches in order to enrich the receptive fields of feature maps and mitigate the impact of feature loss caused by pooling operations. Secondly, we propose a mixed pooling module (MPM) and a dual spatial and channel attention mechanism (DSCA). Leveraging MPM and DSAM, we construct a mixed feature extraction (MFE) block with two different versions. Thirdly, we design a feature enhancement (FE) stage, and append it as the last stage of MPE-HRNet . L . The main contributions of this work are as follows.
  • We propose SPPF . + . Comparing with SPPF, SPPF . + has a stronger feature extraction capability and lower computational load.
  • We propose MPM and DSCA. Based on MPM and DSCA, we design MFE block which possesses strong feature extraction capabilities.
  • We design the FE stage and take it as the final stage of MPE-HRNet . L . This stage is introduced to emphasize the semantic feature in the output feature map of MPE-HRNet . L .
The rest of this paper is organized as follows. Section 2 introduces the theoretical basis. Section 3 elaborates MPE-HRNet . L . After the experimental results are presented in Section 4, the paper is concluded in Section 5.

3. Proposed Algorithm

In this section, we illustrate the proposed MPE-HRNet . L . We outline the structure of MPE-HRNet . L in Section 3.1, and then elaborate the improved version of SPPF in Section 3.2. After representing the details of MFE block in Section 3.3, we describe the feature enhancement stage in Section 3.4.

3.1. Architecture of MPE-HRNet . L

To construct a lightweight network for accurately estimating the poses of multiple animal species, we improved Lite-HRNet and proposed MPE-HRNet . L . The improvements include optimizing branches with SPPF, SPPF . + and MFE block, and introducing a feature enhancement (FE) stage. The branch optimization is designed to enhance the feature extraction and representation capabilities of MPE-HRNet . L . The introduction of the FE stage aims to enhance the important features in the final outputs of MPE-HRNet . L . Figure 1 and Table 1 show the architecture and the details of MPE-HRNet . L , respectively.
Figure 1. Overall architecture of MPE-HRNet . L . MPE-HRNet . L includes five stages. MFE block and SPPF . + /SPPF are applied to stages 2 to 4. FE stage is appended as the final stage. The yellow cuboids describe the feature maps processed by our improvements.
Table 1. Structure details of MPE-HRNet . L .
As Figure 1 and Table 1 show, stage stem only creates a new branch with a conv2d block and a shuffle block. Both stage 2 and stage 3 create a new branch, and apply SPPF . + to the branches created by the previous stage of each of them. They stack multiple MFE blocks to extract features from all the branches except the ones created by themselves. Similar to stage 2 and stage 3, stage 4 also applies stacked MFE blocks to branches. The difference is that stage 4 does not create any new branches, and it applies SPPF instead of SPPF . + to the branch created by its previous stage. Stage FE is quite different from the former four stages. It is introduced to enhance the features output by stage 4.

3.2. SPPF+

Lite-HRNet utilizes downsampling to create new branches. However, downsampling can only generate feature maps with limited receptive fields. It could probably also lose some important features. SPPF can generate feature maps with rich receptive fields by extracting multiple features of varying scales and fusing them together. However, SPPF has three obvious limitations. Firstly, 1 × 1 convolution struggles to extract complex features. Secondly, pooling operations also lose some features. Finally, concat operation increases computational complexity. To address these issues, we improve SPPF and propose SPPF . + .
In order to enhance the ability to extract features, SPPF . + replaces the 1 × 1 convolution with a depthwise separable convolution. To ensure that the feature maps generated by pooling operations contain rich features, SPPF . + inserts fusion operations among the sequential pooling operations. Concretely, it fuses the results of the initial two pooling operations, and then processes the fused results with the last pooling operation. To reduce the computational complexity caused by the concatenation operation, SPPF . + refrains from concatenating the feature maps produced by the pooling operations with the one yielded by the depthwise separable convolution. Instead, it fuses the results of the last pooling operation with the fused results of the first two pooling operations, and then concatenates the fused results with the results of the depthwise separable convolution. SPPF . + employs ADD operations to achieve fusion.
Figure 2 illustrates the architecture of SPPF . + . s p p f + ( X s ) denotes the outputs of SPPF . + with the given input X s , which includes multiple channels. It is computed according to Equation (1). In the equation, c o n v 1 × 1 ( · ) represents 1 × 1 convolution, c o n c a t ( · ) denotes the concatenation operation, c o n v D C ( · ) describes depthwise separable convolution, and m m a p p ( · ) represents a series of max pooling operations. m m a p p ( · ) is calculated according to Equations (2)–(5).
s p p f + ( X s ) = c o n v 1 × 1 ( c o n c a t ( c o n v D C ( X s ) , m m a p p ( X s ) )
m m a x p ( X s ) = s m a x p ( X s ) + d m a x p ( X s ) + t m a x p ( X s )
s m a x p ( X s ) = m a x p ( c o n v D C ( X s ) )
d m a x p ( X s ) = m a x p ( s m a x p ( X s ) )
t m a x p ( X s ) = m a x p ( s m a x p ( X s ) + d m a x p ( X s ) )
Figure 2. Structures of SPPF . + . © and ⊕ represent concat operation and add operation, respectively.

3.3. Mixed Feature Extraction Block

Lite-HRNet exploits the conditional channel weight block (CCW) to extract features. The CCW includes two weight calculation functions, H s and F s . H s employs average pooling to generate low-resolution feature maps, which may loss some important details. F s neglects spatial information when computing weight matrices, which is crucial for localizing keypoints. To address these issues, we propose WGM along with DSCA, and construct the MFE block based on them, which serves as the foundational module for feature extraction.

3.3.1. Weight Generation Module

To alleviate the feature loss caused by average pooling operation in H s , we propose a mixed pooling module (MPM) and construct WGM by replacing the pooling operation in H s with MPM. Figure 3 shows the structure of WGM.
Figure 3. Structure of WGM. The top part shows the structure of WGM, and the bottom part describes the structure of MPM. Matrix transformation represents a series of operations, e.g., reshaping, unsqueezing, and permuting, on the dimensions or sizes of three−dimensional matrices. H T and W T denote the height and width of a transformed matrix, and N is calculated by dividing H T W T by H W .
In the figure, the area surrounded by dashed lines illustrates the details of MPM. As the figure shows, MPM consists of a context-aware branch, an average pooled branch, and a maximum pooled branch. The three branches are designed to capture spatial contextual features, highlight local prominent features, and preserve global features, respectively.
The context-aware branch uses a point-wise convolution to extract spatial contextual features, followed by a s o f t m a x activation to generate a weight matrix. Subsequently, the weight matrix and the input feature maps undergo matrix transformations and multiplication to map the spatial contextual features into the input feature maps.
Taking X m as the input of MPM, X m can be described as ( X m 0 , X m 1 , , X m c 1 ), where c denotes the total number of channels in X m . The output of MPM is denoted as m p m ( X m ) . It is calculated according to Equation (6). In the equation, m b r c h ( X m , i ) denotes the results of the i-th branch in MPM. It is calculated by Equation (7). In the equation, g a p ( · ) and g m p ( · ) separately describe global average pooling and global max pooling; δ ( · ) denotes the s o f t m a x function, and ( · ) T represents matrix transformation.
m p m ( X m ) = i = 0 2 m b r c h ( X m , i ) × θ i
m b r c h ( X m , i ) = ( ( δ ( ( c o n v 1 × 1 ( X m ) ) T ) ) T × X m T ) T , i = 0 g a p ( X m ) , i = 1 g m p ( X m ) , i = 2
Figure 3 also shows the structure of WGM. According to the figure, WGM first performs mixed pooling on input feature maps and then concatenates the pooled results. After that, it processes the concatenated results using two sequential 1 × 1 convolutions, followed by a sigmoid function to the outputs of the convolutions to generate weight matrices.
Taking X w as the input of WGM, X w can be described as ( X w 0 , X w 1 , , X w b 1 ), where b denotes the total number of the existing branches in the corresponding stage. Moreover, X w i ( X w i X w ) , X w i comprises multiple channel feature maps. For the given input X w , the result of WGM is denoted as w g m ( X w ) . w g m ( X w ) is calculated by Equation (8). In the equation, u p s ( · ) denotes the upsampling operation, and σ ( · ) represents the sigmoid function. X w i ( X w i X w ) , WGM calculates a weight matrix for each channel feature map in X w i .
w g m ( X w ) = u p s ( σ ( c o n 1 × 1 ( c o n 1 × 1 ( c o n c a t ( X w ) ) ) ) )

3.3.2. Dual Spatial and Channel Attention Module

Each stage includes both low-resolution branches and high-resolution branches with channel and spatial features. Channel features are richer than spatial features in low-resolution feature maps, while both channel features and spatial features are rich in high-resolution feature maps. F s calculates weights based solely on channel features. It does not capitalize on spatial features which are vital for keypoint localization. To leverage both channel and spatial features for weight calculation, we introduce a dual spatial attention module (DSA) and a hybrid attention mechanism, i.e., DSCA. DSCA is constructed by combining DSA with a channel attention module. Figure 4 shows the structure of DSCA.
Figure 4. Structure of DSCA. It comprises a channel attention module and a DSA module. The area surrounded by red dashed lines describes the structure of DSA. ⨀ represents element-wise multiplication.
DSA utilizes four branches to extract spatial features. The first branch processes each input feature map with max and average pooling, then performs a weighted sum. The second, third, and fourth branches separately use a 1 × 1 convolution, a 3 × 3 convolution, and a 5 × 5 convolution to process each feature map and generate spatial features of various receptive fields. The first and second branches extract global spatial features, capturing the overall structure and layout relationships in each input feature map to obtain higher-level, more abstract semantic information. The third and fourth branches focus on extracting local spatial features, enhancing the attention to different regions within an input image.
After extracting spatial features, DSA concatenates the output of each branch and employs a 7 × 7 convolution to process the concatenated results. It generates a spatial attention weight matrix by applying sigmoid activation function to the results of the 7 × 7 convolution. Given the input X d with c channels, the result of DSA is calculated according to Equation (9). In the equation, σ denotes sigmoid function, c o n v 7 × 7 describes the 7 × 7 convolution, c o n c a t i = 0 3 ( · ) represents the operation to concatenate the results of the four branches, and d b r c h ( X d , i ) depicts the results of the i-th branch. d b r c h ( X d , i ) is calculated according to Equation (10) in which m p ( · ) and a p ( · ) describe max pooling and average pooling, respectively.
d s a ( X d ) = σ ( c o n v 7 × 7 ( c o n c a t i = 0 3 ( d b r c h ( X d , i ) ) ) )
d b r c h ( X d , i ) = 0.5 ( m p ( X d ) + a p ( X d ) ) , i = 0 c o n v 1 × 1 ( X d ) , i = 1 c o n v 3 × 3 ( X d ) , i = 2 c o n v 5 × 5 ( X d ) , i = 3
d s c a ( · ) denotes the outputs of DSCA. It can be calculated according to Equation (11). In this equation, c a ( · ) represents the output of a channel attention module, and ⊙ denotes element-wise multiplication. To simultaneously utilize both channel and spatial features for weight calculation, we utilize DSCA to replace F s to compute weights.
d s c a ( X d ) = c a ( X d ) d s a ( c a ( X d ) X d )

3.3.3. Mixed Feature Extraction Block

We construct MFE block by substituting H s with WGM and F s with DSCA, respectively. MFE has two architectures, namely, MFE-H and MFE-L, which utilize WGM and DSCA in different manners. MFE-H utilizes both WGM and DSCA, while MFE-L only uses WGM. Figure 5 compares the structures of MFE block with CCW block.
Figure 5. Structures of MFE. Both MFE-H and MFE-L exploit WGM to generate weighted feature maps as the input of the 3 × 3 DWConv. MFE-H also utilizes the proposed DSCA to generate the weighted feature maps involving in the concat operation, while MFE-L still utilizes F s to generated weighted feature maps.
High-resolution branches are rich in both channel and spatial features, while low- resolution branches are rich only in channel features. Thus, we apply MFE-H to the two highest-resolution branches and MFE-L to the others.

3.4. Feature Enhancement Stage

Lite-HRNet uses bilinear interpolation in stage 4 to perform upsampling and feature fusion step by step from the 1/32 branch, until generating the output feature maps of the 1/4 branch, which is used for generating heatmaps and locating keypoints. However, bilinear interpolation only considers the subpixel neighborhood and generates feature maps within small receptive fields, which is incapable of capturing the rich semantic information needed for pose estimation. Therefore, we introduce the FE stage to capture more semantic information.
Figure 6 illustrates the structure of the FE stage. The FE stage utilizes content-aware reassembly of features (CARAFE) [] to upsample input feature maps from 1/8 to 1/4 resolution. Then, it uses the efficient channel attention module (ECA) to process the generated feature maps, and fuses the processed results with the input feature maps at 1/4 resolution. Finally, it processes the fused results sequentially with shuffle block, depthwise separable convolutions, and ECA to create feature maps for heatmap generation and keypoint localization.
Figure 6. Structure of FE stage. The cuboids represent feature maps. © denotes concat operation.

4. Experiments

In this section, we conduct experiments on two public benchmark datasets to evaluate MPE-HRNet . L . Firstly, we represent the details of datasets, training, and evaluation metrics in Section 4.1. Secondly, we analyze the ablation experiments in Section 4.2 and comparative experiments in Section 4.3, respectively. Finally, we discuss the limitation of HRNet . L in Section 4.4.

4.1. Experimental Settings

Dataset. We use two datasets, i.e., AP-10K [] and Animal Pose [], to evaluate our solution. AP-10K dataset is the a large-scale dataset for animal pose estimation. It includes 10,015 images collected from 23 families and 54 species of animals. The AP-10K dataset contains a total of 13,028 instances, each of which is labeled with 17 key points to describe poses. Table 2 shows the details of those keypoints. In our experiments, the AP-10K dataset is divided into train set, validation set, and test set. The three datasets include 7010 images, 1002 images, and 2003 images, respectively. The Animal Pose dataset contains 4608 images collected from dog, cat, cow, horse, and sheep species. It includes 6117 instances annotated with 20 keypoints. The Animal Pose dataset is also divided into train set, validation set, and test set. Each of these sets include 2798 images, 810 images, and 1000 images, respectively.
Table 2. Definitions of keypoints.
Training. All the training is carried out on the same server which is equipped with 5 NVIDIA GeForce RTX 3090 GPUs sourced from Santa Clara, CA, USA, and 376 GB of Kingston memory sourced from Fountain Valley, CA, USA. Our solution and all the comparative solutions are implemented in Python 3.8.19, and executed on PyTorch 2.5.0 framework. Multiple data augmentations are adopted in the training, which include random horizontal flip, random rotation from −80 to 80 degrees, random scaling by a factor of 0.5–1.5, and random body part augmentation generating animals with only the upper or lower part of a body.
To ensure the fairness of training processes, the training parameters, batch_size and the number of epochs, are separately set to 64 and 260, while all the other training parameters are configured with their default values. All parameters are updated by Adam optimizer combined with warm-up learning strategy. Learning rate is configured to 5 × 10 4 at the first 170 epochs and reduced to 5 × 10 5 and 5 × 10 6 at the 170th and 200th epochs, respectively.
Evaluation metrics. Object keypoint similarity (OKS) is a widely accepted criterion for pose estimation. It is calculated according to Equation (12). In the equation, d i describes the Euclidean distance between the model prediction and the ground truth of keypoint i, v i denotes the visibility flag of the ground truth of the key point, s represents the scale of the animal object, and k i describes the constant to control the fall-off of the keypoint.
O K S = i e d i 2 2 s 2 i 2 δ ( ν i > 0 ) i δ ( ν i > 0 )
We use the average precision (AP), A P 50 (AP at OKS = 0.5), A P 75 (AP at OKS = 0.75), A P M (AP for medium objects: 322 < area < 962), A P L (AP for large objects: area > 962), and average recall (AR) to evaluate the accuracy of models. AP is regarded as the main metric for accuracy. In addition, we use Params (M) to evaluate complexity, and use FLOPs (G) and inference time (ms) to evaluate speed.

4.2. Ablation Experiments

To analyze the effectiveness of the improvements in this work, we carry out a series of ablation experiments.

4.2.1. Selection of SPPF+ and SPPF

SPPF+ fuses the feature maps generated by different pooling operations by add operation. Add operation adds feature maps related to the same channel. It can merge multiple groups of feature maps generated by different pooling operations into one group, thus reducing parameters for subsequent operations. However, it also results in feature loss. Fortunately, the feature loss could be alleviated if the feature maps involving the same add operation are much similar to each other.
For the feature maps generated by the stacked pooling operations in SPPF . + , the similarity between the feature maps related to the same channel increases with the increase in their resolutions. Therefore, SPPF . + is more suitable for the branches of high resolutions. SPPF does not suffer from the issue. MPE-HRNet . L contains multiple branches of different resolutions. In order to improve accuracy through the proper use of SPPF . + and SPPF in these branches, we carry out the experiments in which SPPF/SPPF . + are added at the beginning of different branches.
Table 3 records the experiment results of Lite-HRNet and its variants only involving SPPF, SPPF . + , or both of them. +SPPF . 1 / 8 , 1 / 16 , 1 / 32 denotes the variant with SPPF in the 1/8, 1/16, and 1/32 branches. +SPPF . + 1 / 8 and SPPF . 1 / 16 , 1 / 32 describes the variant version with SPPF . + in the 1/8 branch and SPPF in the 1/16 and 1/32 branches. +SPPF . + 1 / 8 , 1 / 16 and SPPF . 1 / 32 represent the variant version with +SPPF . + in the 1/8 and 1/16 branches and SPPF in the 1/32 branch. +SPPF . + 1 / 8 , 1 / 16 , 1 / 32 indicates the variant version with SPPF . + in the 1/8, 1/16, and 1/32 branches. From the experimental results, it can be observed that inserting SPPF . + in the 1/8 and 1/16 branches and SPPF in the 1/32 branch yields the best results. In the following experiments, we configure SPPF and SPPF . + according to the best results.
Table 3. Effectiveness study of SPPF . + and SPPF.

4.2.2. Effectiveness Study of All the Improvements

We design five different models to evaluate the effectiveness of each improvement and describe the corresponding experiment results in Table 4. All these results are generated on the AP-10K dataset. In the table, Model0 denotes the original version of Lite-HRNet, while Model1 to Model4 describe different variants of the original version. Model1 and Model2 separately denote the variant with MFE-L in all the branches and the variant with MFE-L in 1/16 and 1/32 branches and MFE-H in 1/4 and 1/8 branches. Model3 represents the variant with MFE-L in 1/16 and 1/32 branches, MFE-H in 1/4 and 1/8 branches, and FE stage.
Table 4. Effectiveness study of all the improvements.
According to Table 4, Model1 and Model2 achieved improvements in AP, AR, parameters, and FLOPs(G) separately compared to Model0 and Model1, indicating that the introduction of MFE-L and MFE-H helped to improve the performance of Lite-HRNet. Model3 also achieved performance improvements in the above four evaluation metrics compared with Model2, which verified the effectiveness of the collaboration of MFE-L, MFE-H, and FE stage in improving the performance of Lite-HRNet.
Model4 describes the variant with MFE-L, MFE-H, the new stage, and S P P F + /SPPF. It executes with a higher performance compared with Model3, which verified the effectiveness of all the improvements proposed in this work.

4.3. Performance Comparison

4.3.1. Comparison on AP-10K Dataset

We carry out experiments to evaluate the detection accuracy and speed of MPE-HRNet . L by comparing it with nine important pose estimation solutions, including CSPNeXt-t [], CSPNeXt-s, ShuffleNetV1 [], ShuffleNetV2 [], MobileNetV2 [], MobileNetV3 [], Dite-HRNet [], Lite-HRNet, and DARK []. DARK utilizes Lite-HRNet to generate heatmaps but performs post-processing in a different way to the other solutions. All the solutions are lightweight solutions.
Table 5 shows the results of these solutions on the AP-10K validation set. According to the table, MPE-HRNet . L obtained the highest AP, A P 50 , A P 75 , A P M , A P L , and AR. Compared with ShuffleNetV2, ShuffleNetV1, MobileNetV2, MobileNetV3, CSPNeXt-t, Dite-HRNet, Lite-HRNet, DARK, and CSPNeXt-s, it improved AP in animal pose estimation by 8.8%, 8.5%, 6.4%, 5.5%, 5.3%, 2.3%, 2.2%, 2.0%, and 1.5%, respectively. It also achieved improvements when any of the other metrics was used. It is obvious that MPE-HRNet . L outperformed all the comparative solutions in accuracy.
Table 5. Results on the AP-10K validation set.
We aimed to strike a balance between accuracy and complexity, striving to carefully manage the complexity of MPE-HRNet . L during its design process. In all the solutions, DARK and Lite-HRNet have the least Params, and Dite-HRNet has the lowest FLOPs. Comparing with these two solutions, MPE-HRNet . L does not increase Params and FLOPs too much while obtaining an improvement in accuracy. Concretely, it has only 0.43 M more parameters than DARK and Lite-HRNet, and 0.29 G more FLOPs than Dite-HRNet, respectively. Comparing with ShuffleNet, MobileNet, and CSPNeXt, MPE-HRNet . L has the least Params and the lowest FLOPs(G). The inference time of MPE-HRNet . L is 1.74 ms longer than that of DARK and 1.62 ms longer than Lite-HRNet, indicating a slight increase with negligible impact.
Table 6 shows the results on the AP-10K test set. According to the table, MPE-HRNet . L still achieved the highest AP in all the comparison solutions. Comparing with ShuffleNetV2, ShuffleNetV1, MobileNetV2, MobileNetV3, CSPNeXt-t, Dite-HRNet, Lite-HRNet, DARK, and CSPNeXt-s, it improved AP by 8.4%, 7.9%, 4.9%, 4.9%, 4.6%, 1.4%, 1.3%, 1.4%, and 1.1%, respectively. MPE-HRNet . L also exhibited the highest A P 50 , A P 75 , A P M , A P L , and AR in all the other comparison solutions except DARK. Although DARK surpasses our method in the A P M metric, it falls behind MPE-HRNet . L in all the other metrics. Considering all the metrics, MPE-HRNet . L outperforms all comparison solutions on the test set.
Table 6. Results on the AP-10K test set.
We also compared MPE-HRNet . L with ResNet-101, HRNet-W32, and RFB-HRNet. The latter three solutions are based on large models. According to the results in Table 7, these three solutions exhibit a higher accuracy than MPE-HRNet . L . However, they also have many more Params than MPE-HRNet . L . In many practical application scenarios, pose estimation models are required to be deployed on edge devices. Large models struggle to perform optimally on edge devices due to the inability of these devices to provide the resources required by the models.
Table 7. Comparison with solutions based on large models.

4.3.2. Comparison on Animal Pose Dataset

We also conducted experiments on the Animal Pose dataset. In these experiments, we compared MPE-HRNet . L with ShuffleNetV2, MobileNetV2, and CSPNeXt-s. The corresponding results are recorded in Table 8. Similar to the results on the AP-10K dataset, MPE-HRNet . L also achieved the highest AP on the Animal Pose dataset, demonstrating the effectiveness of MPE-HRNet . L in improving accuracy.
Table 8. Results on the Animal Pose dataset.

4.4. Limitation Discussion

While MPE-HRNet . L obtains accuracy improvements, it still suffers from two obvious limitations. The first limitation is that MPE-HRNet . L does not perform well in complex occlusion scenes. The second limitation is that the accuracy of MPE-HRNet . L needs to be improved.
Complex occlusion limitation. We chose twelve images with different scenes and depicted the corresponding results of MPE-HRNet . L in Figure 7. These images were shot at different distances and angles. They include various animals belonging to different species. Some of them even include shadow and object occlusions. According to Figure 7b,e,h, MPE-HRNet . L performed well in the images including multiple objects even if those objects are different sizes. It also performed well in the scenario where shadow exists, according to Figure 7i. MPE-HRNet . L can still work well in the scenario where some parts of bodies are occluded lightly, as Figure 7j,k show. However, it does not work well in the scenario where complex occlusion exists, as Figure 7f shows.
Figure 7. Example qualitative results of the AP-10K.
Accuracy limitation. According to Section 4.3.1 and Section 4.3.2, MPE-HRNet . L obtains accuracy improvements compared with lightweight models such as Dark, Lite-HRNet, and so on. However, MPE-HRNet . L exhibits a relatively low accuracy when compared with the solutions based on large models. In addition, MPE-HRNet . L improves accuracy by 1.4%, 1.3%, and 1.4% compared with Dark, Lite-HRNet, and Dite-HRNet, respectively. However, it increases parameters by 38% (0.43 M), 38% (0.43 M), and 31% (0.37 M) compared with the three solutions, respectively. Compared to the improvement in accuracy, the increase in parameters is more pronounced. Nonetheless, the number of parameters of MPE-HRNet . L is only 1.56 M. MPE-HRNet . L is still a lightweight model. However, MPE-HRNet . L does not exhibit any advantage in accuracy compared to large models. It is necessary to improve MPE-HRNet . L to obtain higher accuracy.

5. Conclusions

In this work, we focused on the problem of estimating the poses of multiple species with high accuracy in real scenarios. To deal with this problem, we constructed a lightweight network by improving Lite-HRNet. Concretely, we proposed SPPF . + and applied it as well as SPPF to all branches. We proposed MPM and DSCA, and constructed MFE block by applying them together. We also designed a new stage to enhance the features in the feature map output by MPE-HRNet . L . The experimental results on the AP-10K dataset and the Animal Pose dataset verified the effectiveness of our work. In the future, we will focus on the problem of animal pose estimation involving complex occlusions.

Author Contributions

Conceptualization, J.S., J.L. and W.W.; methodology, J.S., Y.J. and W.W.; formal analysis, J.S.; investigation, Y.J. and W.W.; resources, J.S. and Y.J.; data curation, Y.J.; writing—original draft preparation, J.S., Y.J. and W.W.; writing—review and editing, J.S., Y.J. and W.W.; visualization, Y.J.; supervision, J.S., J.L. and W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by Junwei Luo and the National Natural Science Foundation of China (Grant No. 61972134) and the Innovative and Scientific Research Team of Henan Polytechnic University (No. T2021-3).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The datasets used in this article are publicly available. The AP-10K dataset can be accessed at https://drive.google.com/file/d/1-FNNGcdtAQRehYYkGY1y4wzFNg4iWNad/view (accessed on 15 May 2024), and the Animal Pose dataset is available at https://openxlab.org.cn/datasets/OpenDataLab/animalpose (accessed on 4 October 2024).

Conflicts of Interest

Author Wei Wang was employed by the company CSSC Haifeng Aviation Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Harding, E.J.; Paul, E.S.; Mendl, M. Cognitive bias and affective state. Nature 2004, 427, 312. [Google Scholar] [CrossRef] [PubMed]
  2. Zuffi, S.; Kanazawa, A.; Berger-Wolf, T.; Black, M.J. Three-D Safari: Learning to Estimate Zebra Pose, Shape, and Texture from Images “In the Wild”. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5359–5368. [Google Scholar]
  3. Anderson, T.L.; Donath, M. Animal behavior as a paradigm for developing robot autonomy. Robot. Auton. Syst. 1990, 6, 145–168. [Google Scholar] [CrossRef]
  4. Jiang, L.; Lee, C.; Teotia, D.; Ostadabbas, S. Animal pose estimation: A closer look at the state-of-the-art, existing gaps and opportunities. Comput. Vis. Image Underst. 2022, 222, 103483. [Google Scholar] [CrossRef]
  5. Cheng, B.; Xiao, B.; Wang, J.; Shi, H.; Huang, T.S.; Zhang, L. Higherhrnet: Scale-aware representation learning for bottom-up human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5386–5395. [Google Scholar]
  6. Chao, W.; Duan, F.; Du, P.; Zhu, W.; Jia, T.; Li, D. DEKRV2: More accurate or fast than DEKR. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 1451–1455. [Google Scholar]
  7. Xiao, B.; Wu, H.; Wei, Y. Simple baselines for human pose estimation and tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 466–481. [Google Scholar]
  8. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5693–5703. [Google Scholar]
  9. Yu, C.; Xiao, B.; Gao, C.; Yuan, L.; Zhang, L.; Sang, N.; Wang, J. Lite-hrnet: A lightweight high-resolution network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 10440–10450. [Google Scholar]
  10. Li, C.; Lee, G.H. From synthetic to real: Unsupervised domain adaptation for animal pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 1482–1491. [Google Scholar]
  11. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  12. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 483–499. [Google Scholar]
  13. Wang, Z.; Zhou, S.; Yin, P.; Xu, A.; Ye, J. GANPose: Pose estimation of grouped pigs using a generative adversarial network. Comput. Electron. Agric. 2023, 212, 108119. [Google Scholar] [CrossRef]
  14. Fan, Q.; Liu, S.; Li, S.; Zhao, C. Bottom-up cattle pose estimation via concise multi-branch network. Comput. Electron. Agric. 2023, 211, 107945. [Google Scholar] [CrossRef]
  15. He, R.; Wang, X.; Chen, H.; Liu, C. VHR-BirdPose: Vision Transformer-Based HRNet for Bird Pose Estimation with Attention Mechanism. Electronics 2023, 12, 3643. [Google Scholar] [CrossRef]
  16. Zhou, F.; Jiang, Z.; Liu, Z.; Chen, F.; Chen, L.; Tong, L.; Yang, Z.; Wang, H.; Fei, M.; Li, L.; et al. Structured context enhancement network for mouse pose estimation. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 2787–2801. [Google Scholar] [CrossRef]
  17. Graving, J.M.; Chae, D.; Naik, H.; Li, L.; Koger, B.; Costelloe, B.R.; Couzin, I.D. DeepPoseKit, a software toolkit for fast and robust animal pose estimation using deep learning. eLife 2019, 8, e47994. [Google Scholar] [CrossRef]
  18. Zhao, S.; Bai, Z.; Meng, L.; Han, G.; Duan, E. Pose Estimation and Behavior Classification of Jinling White Duck Based on Improved HRNet. Animals 2023, 13, 2878. [Google Scholar] [CrossRef]
  19. Gong, Z.; Zhang, Y.; Lu, D.; Wu, T. Vision-Based Quadruped Pose Estimation and Gait Parameter Extraction Method. Electronics 2022, 11, 3702. [Google Scholar] [CrossRef]
  20. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  21. Maselyne, J.; Adriaens, I.; Huybrechts, T.; De Ketelaere, B.; Millet, S.; Vangeyte, J.; Van Nuffel, A.; Saeys, W. Measuring the drinking behaviour of individual pigs housed in group using radio frequency identification (RFID). Animal 2016, 10, 1557–1566. [Google Scholar] [CrossRef] [PubMed]
  22. Liu, S.; Fan, Q.; Liu, S.; Zhao, C. DepthFormer: A High-Resolution Depth-Wise Transformer for Animal Pose Estimation. Agriculture 2022, 12, 1280. [Google Scholar] [CrossRef]
  23. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  24. Sun, K.; Zhao, Y.; Jiang, B.; Cheng, T.; Xiao, B.; Liu, D.; Mu, Y.; Wang, X.; Liu, W.; Wang, J. High-resolution representations for labeling pixels and regions. arXiv 2019, arXiv:1904.04514. [Google Scholar]
  25. Xu, Y.; Zhang, J.; Zhang, Q.; Tao, D. Vitpose: Simple vision transformer baselines for human pose estimation. Adv. Neural Inf. Process. Syst. 2022, 35, 38571–38584. [Google Scholar]
  26. Liao, J.; Xu, J.; Shen, Y.; Lin, S. THANet: Transferring Human Pose Estimation to Animal Pose Estimation. Electronics 2023, 12, 4210. [Google Scholar] [CrossRef]
  27. Hu, X.; Liu, C. Animal Pose Estimation Based on Contrastive Learning with Dynamic Conditional Prompts. Animals 2024, 14, 1712. [Google Scholar] [CrossRef] [PubMed]
  28. Zeng, X.; Zhang, J.; Zhu, Z.; Guo, D. MVCRNet: A Semi-Supervised Multi-View Framework for Robust Animal Pose Estimation with Minimal Labeled Data. 2024; preprint. [Google Scholar] [CrossRef]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  30. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  31. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  32. Qin, Z.; Zhang, P.; Wu, F.; Li, X. Fcanet: Frequency channel attention networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 783–792. [Google Scholar]
  33. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  36. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  37. Jocher, G. YOLOv5 by Ultralytics. 2020. Available online: https://zenodo.org/records/7347926 (accessed on 11 April 2024).
  38. Wang, J.; Chen, K.; Xu, R.; Liu, Z.; Loy, C.C.; Lin, D. Carafe: Content-aware reassembly of features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3007–3016. [Google Scholar]
  39. Yu, H.; Xu, Y.; Zhang, J.; Zhao, W.; Guan, Z.; Tao, D. Ap-10K: A benchmark for animal pose estimation in the wild. arXiv 2021, arXiv:2108.12617. [Google Scholar]
  40. Cao, J.; Tang, H.; Fang, H.S.; Shen, X.; Lu, C.; Tai, Y.W. Cross-Domain Adaptation for Animal Pose Estimation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  41. Jiang, T.; Lu, P.; Zhang, L.; Ma, N.; Han, R.; Lyu, C.; Li, Y.; Chen, K. Rtmpose: Real-time multi-person pose estimation based on mmpose. arXiv 2023, arXiv:2303.07399. [Google Scholar]
  42. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  43. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  44. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  45. Li, Q.; Zhang, Z.; Xiao, F.; Zhang, F.; Bhanu, B. Dite-HRNet: Dynamic lightweight high-resolution network for human pose estimation. arXiv 2022, arXiv:2204.10762. [Google Scholar]
  46. Zhang, F.; Zhu, X.; Dai, H.; Ye, M.; Zhu, C. Distribution-aware coordinate representation for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7093–7102. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.