Next Article in Journal
Correction: Martin et al. The FeetMe® Insoles System: Repeatability, Standard Error of Measure, and Responsiveness. Sensors 2024, 24, 6043
Previous Article in Journal
Gap Measurements in Aerospace Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLOv8 Architectural Scene Section Recognition Method Based on SimAM-EMA Hybrid Attention Mechanism

1
Quzhou Special Equipment Inspection&Testing Research Institute, Quzhou 324000, China
2
School of Information and Software Engineering, East China Jiaotong University, Nanchang 330013, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3060; https://doi.org/10.3390/s25103060
Submission received: 24 March 2025 / Revised: 29 April 2025 / Accepted: 6 May 2025 / Published: 12 May 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

:
To overcome the limitations of current methods in architectural scene segmentation accuracy, this study presents a hybrid attention-enhanced YOLOv8 framework and introduces a dedicated building interior profile dataset. The proposed approach extends the YOLOv8 architecture by integrating SimAM (Simple, Parameter-Free Attention Module) to dynamically evaluate neuron significance and refine feature representations. This is coupled with an Efficient Multi-Scale Attention (EMA) module that synergizes local and global attention mechanisms, enabling robust multi-scale feature fusion while maintaining stable weight optimization. To address the scarcity of domain-specific data, a meticulously annotated dataset encompassing common architectural interior elements is developed. Comprehensive evaluations demonstrate that the enhanced model achieves an 89.9% mAP@0.5 on the proposed dataset, outperforming the baseline YOLOv8 with relative improvements of 4.5% in precision, 5.2% in recall, 5.1% in mAP@0.5, and 4.5% in mAP@0.5–0.95. These advancements underscore the efficacy of hybrid attention mechanisms in architectural scene analysis and establish a benchmark dataset to facilitate future research in intelligent building environment interpretation.

1. Introduction

With the rapid development of 3D acquisition, display, and plotting technology, 3D modeling has become one of the key technologies in computer-aided civil and infrastructure engineering. As a result, the application of large-scale 3D infrastructure models is becoming increasingly important in these industries.
While researchers have extensively investigated the basic techniques used for the feature extraction, segmentation, and rendering of conventional 3D models, there is an increasing demand in the industry for processing techniques dedicated to large-scale 3D infrastructure models. With the development of computer-aided civil engineering, the informatization of various requirements in civil engineering has become one of the mainstream trends.
In civil engineering, a sectional plan is mainly used to represent the internal structure of an object. It is performed by applying a cutting plane (plane or surface) to section the object, removing the part between the observer and the cutting plane, and projecting the rest onto the projection plane. The resulting diagram is referred to as a section view.
Especially in architecture, a sectional plan refers to the cross-section of a building, with one or more upright sections perpendicular to the axis of the exterior walls. The resulting projection is used to represent the internal structure or construction form of the building, along with the layering situation, the connections between each part, the materials and heights, etc.; the sectional plan is one of the most indispensable and essential drawings used in collaboration with the ichnography and elevation drawings. The 3D cutting (or 2D plane cutting) of sectional views is a critical processing technique for CID design analysis and the exploration of data distribution inside the structure, as introduced by Yang [1] et al. Especially for modern building informatization modeling and design, sectional drawings based on 3D cutting are essential references for observing the details and verifying the correctness of the design. For large-scale 3D infrastructure modeling, 3D cutaway-based profiles are necessary for designers’ evaluation and valuable for engineers to review the 3D details. For example, sectional plans can provide an opportunity to perceive the correctness of 3D models enclosed in larger 3D objects.
With the advancement of 3D modeling technology, building information modeling (BIM) [2] has been widely used to produce various 3D building models. Through the 2D plane cutting of 3D building models, the section view of buildings can be efficiently and accurately obtained. Further, by introducing advanced image recognition algorithms, we can automatically analyze and identify each component and detail in the profile. This method improves the accuracy and efficiency of identification. It provides new ideas for extending the identification to the section plane of a building floor and even to the component identification in the section plane of large substation models. Architectural section detection differs significantly from general object detection tasks. Firstly, architectural section images have special structural characteristics; they contain rich geometric shapes, material boundaries, and spatial relationship information, along with multi-scale features, requiring precise identification of both large structural components (such as walls, floor slabs) and smaller detail components (such as doors, windows, furniture). Secondly, architectural section images typically have high information density, with complex layering and nesting relationships between various components, increasing the difficulty of detection.
In the field of computer vision, attention mechanisms have become a key technology for improving model performance by mimicking the selective attention process of the human visual system, enabling models to focus on the most relevant parts of input data. In architectural scene analysis, attention mechanisms are particularly important due to the complex geometric shapes and multi-level spatial relationships in structures. Addressing the unique challenges of architectural section detection, SimAM precisely captures critical lines and geometric features by evaluating neuronal importance, while EMA enhances the understanding of spatial relationships between components by integrating local and global information. The combination of these mechanisms not only improves the model’s ability to recognize complex architectural elements but also enhances the efficiency of multi-scale feature processing, enabling more accurate identification of various components in architectural sections, from small details to large structural elements, effectively addressing detection challenges across different scales.
The main contributions of this paper are as follows: (1) A dataset of architectural section planes was created, which provides ideas for the subsequent identification of components of 3D model cutting planes, such as large-scale building scenes and substations. (2) A hybrid attention mechanism based on quantitative evaluation was proposed. The improved method is more accurate than the original YOLO series model and MM detection model.

2. Related Work

The use of 3D cutting in section-based analysis has been widely studied in various fields, such as manufacturing and medical analysis based on 3D models. Specifically, sectional views are an important tool for design optimization and the visualization of large structural models [3].
A profile is mainly used to represent an object’s internal structure. It is obtained by applying a sectional plane (plane or surface) to cut the object, removing the part between the observer and the sectional plane, and projecting the rest onto the projection plane. The resulting figure is called the section view. Section views have always been an important reference point for designing and analyzing project structures and construction drawings. Holgado-Barco [4] et al. proposed a method for processing the data captured by a moving laser scanning system to obtain the geometric cross-section of a highway. Arias [5] et al. used photogrammetric information to collect the sectional properties of an ancient, irregular piece of wood, further analyzed the geometric data on this basis, and finally realized the three-dimensional reconstruction of the scanned object. Based on the important information in the profile, Lin [6] et al. developed a safe structural form and construction method for underground infrastructure. Karhu [7] et al. developed a general construction process modeling method based on a local view of the model. This approach is particularly valuable for industry practitioners, enabling them to interact with the model and represent scheduling in the construction process. Yuan [8] et al. determined the location of the section plane through medical anatomical points, using a rotation sequence of the long-axis and short-axis views of the anchor points to conduct a three-dimensional reconstruction of the heart. Hu H. [9] et al. analyzed organ profiles to classify organs based on color blocks and reconstructed three-dimensional organs by multi-angle projection, filter-back projection, and other strategies. Shi [10] et al. proposed a deep learning framework, MsvNet, based on a multi-sectional view (MSV) representation for feature segmentation and recognition algorithms. Y. Wen [11] et al. used a profile to reconstruct a 3D model from a 2D profile by enhancing the traditional volume-based method. By locating the object recognition geometric plane of the profile, Gong [12] et al. developed a graphical representation to describe multiple relationships between individual views in a two-dimensional graphical space and implemented an evidence theory-based reasoning technique to validate the relationships used to fold views and profiles in a three-dimensional object space.
In recent years, significant progress has been made in target recognition through the adoption of deep learning techniques, transitioning from traditional feature-based manual algorithms to automatic recognition methods based on deep neural networks [13]. Currently, deep learning technology has made remarkable achievements in the field of target detection. In particular, convolutional neural networks (CNNs), proposed by LeCun et al., first demonstrated their superior performance in handwritten digit recognition tasks [14]. Among them, the YOLO (You Only Look Once) family of models is a class of deep learning models widely used in target detection, with the unique feature of end-to-end detection, enabling the simultaneous completion of target detection and localization in a single forward propagation, with the advantages of high efficiency and accuracy. There have been multiple studies on the application of YOLO series models in the construction industry. Kisaezehra [15] et al. implemented a safety helmet detection system at construction sites, which utilized the YOLOv5 model to achieve real-time detection of safety helmets at construction sites, providing technical support for ensuring worker safety. Feng [16] et al. explored the application of YOLO models in identifying safety risks at construction sites. Wang [17] et al. published ’YOLOv7 Optimization Model Based on Attention Mechanism Applied in Dense Scenes,’ which optimized the YOLOv7 network by combining standard convolution with attention mechanisms, proposing a YOLOv7B-CBAM model that effectively improved object detection accuracy in dense construction scenarios. These studies indicate that the YOLO series models, due to their real-time capability and high accuracy, have broad application prospects in the construction industry, while the introduction of attention mechanisms can further enhance their recognition capabilities in complex construction scenes.
Therefore, it can be inferred that the sectional plan obtained by a 3D model cutting method has significant research value in various fields.

3. Methodology

With the development of information technology in construction, the intelligent generation of building sectional plans and the recognition of sectional information are significant. The recognition of a building model’s sectional plan allows engineers to examine 3D details and evaluate the safety and rationality of the building [18]. A hybrid attention mechanism based on the YOLOv8 model is proposed for the current 3D architecture. This attention mechanism dramatically improves the recognition accuracy of the dataset of the internal profile plane of the building. Firstly, the input images are normalized, and then the feature extraction and other operations are carried out through the backbone network. Finally, the image is detected through the detection head network. The overall workflow of the network is as follows, see Figure 1:

3.1. YOLOv8 Model

The YOLO series model was proposed by Joseph Redmon et al. The original YOLOv1 used a single convolutional neural network to directly predict bounding boxes and categories, achieving efficient object detection. With each iteration, the YOLO model continues to introduce new technologies and improvements, such as the batch normalization of YOLOv2 [19], the multi-scale prediction of YOLOv3 [20], CSPDarknet53 [21], the architecture of YOLOv4 [22], and the lightweight design of YOLOv5. Building on this, YOLOv8 is further optimized by enhancing the model architecture and adopting more efficient training strategies and optimization methods, resulting in significantly improved detection accuracy and speed. The architecture design of YOLOv8 is mainly reflected in the following aspects:

3.1.1. Improved Feature Extraction Network

YOLOv8 has significantly improved feature extraction networks, adopting a deeper and wider network structure—CSPNet (Cross Stage Partial Network)—to improve the processing capability of complex scenes. Implementing CSPNet effectively reduces the computational cost and improves the model’s feature representation ability. CSPNet [23] reduces redundant computation by passing partial features layer by layer and fusing these features at specific layers. YOLOv8 uses improved Backbone networks like CSPDarknet53 [24] to enhance its feature extraction capabilities. The new Backbone network improves the depth and width of the model by adding convolutional layers and optimizing the residual structure.

3.1.2. Multi-Scale Feature Fusion

YOLOv8 introduces multi-scale feature fusion techniques, such as FPN (Feature Pyramid Network) [25], which improves the detection accuracy of small and large targets by constructing bottom-up feature pyramids and combining feature maps at different scales, and PANet (Path Aggregation Network) [26], which further enhances the richness of feature expression and the detection accuracy of targets at different scales by top-down path augmentation feature fusion [27].

3.1.3. New Activation Function

YOLOv8 uses the Mish [28] activation function, which performs better in training deep neural networks than traditional ReLU functions. Compared with ReLU, the Mish function has better smoothness and nonlinear characteristics, which helps to improve the model’s expression ability and training stability.

3.1.4. Attention Mechanism

YOLOv8 introduces the SE (Squeeze and Excitation) [29] module to improve detection accuracy by focusing on important features. The SE module adjusts the weight of the feature map by global information so that the model can pay more attention to important features and improve the detection performance.

3.2. SimAM Attention Mechanism

The SimAM [30] attention mechanism is a simple but very effective attention module for convolutional neural networks. In contrast with the existing channel and spatial attention modules, SimAM does not require additional parameters to the original network. Instead, it infers the 3D attention weights of the feature graph in a single layer. Specifically, SimAM, based on some well-known neuroscience theories, proposes to optimize an energy function to discover the importance of each neuron, further deduces a solution for a fast closed form of the energy function, and shows that the solution can be implemented in less than ten lines of code. Another advantage of this module is that most operators are selected according to the solution of the defined energy function, avoiding too much effort on the structural adjustment. Through the quantitative evaluation of various visual tasks, the flexibility and effectiveness of the module have been demonstrated, and the expressibility of many ConvNets has been improved.
The method defines the following energy function:
e t ( w t , b t , y , x i ) = ( y t ( t ^ ) ) 2 + 1 M 1 i = 1 M 1 ( y 0 x i ^ ) 2
where t ^ = w t t + b t , x i ^ = w i x i + b t denotes the linear transformation of t; w i are the target neurons and the other neurons in a single channel of the input feature X R C × H × W ; i is the index in the spatial dimensions; M = H × W is the number of neurons in this channel; and w t and b t are the weights and biases for the transformation. When n = y t , Formula (1) reaches its minimum value, and all other n = y 0 , where y t and y 0 are two different values. By minimizing the equation, Formula (1) is equivalent to finding the linear separability between the target neuron t and all other neurons within the same channel. For simplicity, we adopt binary labels (i.e., 1 and −1) for y t and y 0 and add a regularizer to Equation (1). The final energy function is as follows:
e t ( w t , b t , y , x i ) = 1 M 1 i = 1 M 1 ( 1 ( w t x i + b t ) ) 2 + ( 1 ( w t t + b t ) ) 2 + λ w t 2
Theoretically, each channel has M energy functions. Solving all these equations with iterative solvers like SGD is a computational burden. Fortunately, Formula (2) has a fast closed solution, which can be easily resolved using the following:
w t = 2 ( t μ t ) ( t μ t ) 2 + 2 σ t 2 + 2 λ
b t = 1 2 ( t + μ t ) w t
Since all the neurons on each channel follow the same distribution, the mean and variance can be computed first for the input features in both the H and W dimensions, while avoiding repeated computations, as follows:
e t * = 4 ( σ ^ 2 + λ ) ( t μ ^ ) 2 + 2 σ ^ 2 + 2 λ )
where μ ^ = 1 M i = 1 M x i , σ ^ 2 = 1 M i = 1 M ( x i μ ^ ) 2 . The lower the energy e t * of a neuron t, the greater difference between it and the surrounding neurons, and the more important it is for visual processing. Therefore, the importance of each neuron can be determined by 1 e t *
The whole process can be expressed as follows:
X ˜ = s i g m o i d 1 E X
where E groups all 1 e t * along the channel and spatial dimensions. Adding a sigmoid function limits the values in e that are too large. The sigmoid is a monotonic function, so it does not affect the relative importance of each neuron.
The SimAM attention mechanism can infer three-dimensional attention weights for feature graphs without increasing the original network parameters. It is proposed to optimize the energy function to tap into each neuron’s importance.

3.3. EMA Attention Mechanism

The EMA [31] attention mechanism is a novel and efficient multi-scale attention mechanism that focuses on preserving information about each channel and reducing computational overhead by realigning some channels into batch dimensions, as well as grouping channel dimensions into multiple sub-features, so that the spatial semantic features are evenly distributed within each feature group. Specifically, in addition to encoding the global information in each parallel branch to recalibrate the channel weights, the output features of these two parallel branches are further aggregated through cross-dimensional interactions to capture pixel-level pair relationships, addressing the potential side effects of modeling cross-channel relationships through channel dimension reduction on extracting depth vision representations.
The EMA attention mechanism optimizes the representation of multi-scale features by introducing the methods for examining cross-dimensional interaction and feature grouping without significantly increasing the computational complexity. It also provides an efficient solution for inferring three-dimensional attention weights from feature graphs. The mechanism enables the following processes:
1. Feature grouping: The input feature map is divided into multiple sub-features according to channel dimensions to learn different pieces of semantic information.
2. Parallel subnetworks: The attention-weight descriptors of feature graphs are extracted by parallel processing of 1 × 1 and 3 × 3 convolution. The 1 × 1 branch encodes channel information through two 1D global averaging pooling operations, while the 3 × 3 branch captures multi-scale features through a 3 × 3 convolution.
3. Cross-space learning: A 2D global average pooling operation encodes global spatial information, and a Softmax nonlinear function fits a Gaussian distribution to generate a spatial attention map.
The global averaging pooling formula for input features with width W is as follows:
z C W ( W ) = 1 H 0 j H x c ( j , W )
where C denotes the number of input channels; H and W represent the spatial dimensions of the input features, respectively; and x c denotes the input feature of the c-th channel.
The formula for the 2D global averaging pooling operation is as follows:
z C = 1 H × W i = 0 H j = 0 W x c ( i , j )
It is designed for encoding global information and modeling remote dependencies. A natural nonlinear function, Softmax with 2D Gaussian mapping, is adapted to fit the aforementioned linear transformation at the output of 2D global average pooling to enhance computational efficiency. We obtain the spatial attention map by multiplying the production of the above-mentioned parallel processing with the matrix dot product operation.
The EMA attention mechanism captures pixel-level pairing relationships and highlights the global context by collecting multi-scale spatial information at the same processing stage. The final output feature map is the same size as the input and can be efficiently stacked into modern architectures. At the same time, the EMA attention mechanism effectively fuses cross-channel and spatial position information to enhance the pixel-level attention of convolutional neural networks on high-level feature maps, providing significant performance improvements for visual tasks.

3.4. Hybrid Attention Mechanism

As mentioned, the SimAM [32] attention mechanism is a simple but very effective attention module for convolutional neural networks. Based on neuroscience theory, it optimizes the energy function to discover the importance of each neuron without adding parameters to the original network. In addition, the EMA attention mechanism is an efficient multi-scale attention module that enhances spatial semantic feature distribution by reshaping channels into batch dimensions and grouping them into sub-features. It also utilizes interdimensional interactions to capture pixel-level relationships. See Figure 2.
To perform feature extraction, input feature maps are processed by the SimAM and EMA modules. The SimAM module optimizes the energy function, discovering the importance of each neuron, while the EMA module optimizes multi-scale feature representation through reshaping and grouping operations. Using the fused feature map can improve the feature representation capability without significantly increasing the computational complexity. This hybrid mechanism can take advantage of the simple efficiency of SimAM and the multi-scale feature representation of EMA to provide a richer feature representation. The fusion process is shown in Figure 3.
Firstly, the SimAM module receives the image features and deduces the 3D attention weights for the feature graphs. Then, the EMA module receives image features reweighted by SimAM, performs grouping and parallel convolution processing on them, and outputs the final image features. This hybrid attention mechanism takes full advantage of the parametric efficiency and simplicity of SimAM, as well as EMA’s multi-scale feature representation capabilities and computational efficiency.
We can express the combination of the SimAM and EMA modules as follows:
H y b r i d ( X ) = α · S i m A M ( X ) + β · E M A ( X )
where α and β are the fusion weights, adjusting the influence of SimAM and EMA in the overall feature map. According to the above-mentioned calculation formula, the hybrid attention mechanism can improve the feature representation ability without significantly increasing the computational complexity and is thus suitable for various visual tasks.

4. Experimental Analysis

This study employs a YOLOv8 architecture based on hybrid attention mechanisms for architectural scene section recognition. The experimental procedure primarily consists of the following four stages: dataset creation, model architecture design, model training and optimization, and performance evaluation.

4.1. Building Scene Internal Profile Dataset Production

Since there is limited research on identifying profiles within architectural scenes, standardized datasets for testing scene profiles are scarce. Therefore, we built some models of the interior of the building using Blender 4.1.2 and other software, incorporating various furniture commonly found in the interior as an example. We generated accurate building interior section images by setting up cutting planes perpendicular to the axis in the three-dimensional model, removing the portion between the observer and the cutting plane, and projecting the remaining portion onto the projection plane. This method provides a clear representation of the building’s internal structure and a consistent data foundation for model training.
When creating the dataset, we followed the following principles: 1. The location is reasonable and in line with human cognition. For example, the models can be physically adjacent, but no collision should result in the table inside the bed, which is not in line with reality. 2. In line with the actual needs of the field of civil engineering construction, only the X, Y, and Z axes are cut because, in practical applications, the model only needs to be cut along the X, Y, and Z directions to meet the needs of civil engineering analysis of the section plane; 3. Since the purpose of section plane identification in this work is to further extend to the identification of buildings, industrial machines, and other equipment; these tools or buildings usually do not have pronounced color characteristics. Unified decolorization is performed on all furniture models to ensure that color-related issues do not affect identification accuracy.

4.2. Implementation Details

The experimental setup in this paper uses Windows 10 as the operating system, an NVIDIA RTX 3060 GPU device, Blender 3.6 for building the scene cross-sectional dataset, and PyTorch 2.1.0 for constructing the experimental code. The initial learning rate for training was set to 0.01, with a weight decay of 0.005, and the number of epochs for training on the scene cross-section dataset was 300. The input image size was 640 × 640 pixels.

4.3. Comparative Experiment of Attention Mechanism

To verify the effectiveness of the hybrid attention mechanism, we introduced the performance of several different attention mechanisms, such as SimAM, EMA, GMA [33], SEA [34], and CA [35], and verified each of them in turn; all conditions were held constant. YOLOv8 was used to compare the different attention mechanisms. The experimental results are shown in Table 1, Table 2 and Table 3.
According to the experimental results in Table 1, adding the EMA attention mechanism improves the accuracy and mAP@0.5. In contrast, adding the SimMA attention mechanism improves the recall rate. To effectively identify the influence of these attention mechanisms on the detection results, the two attention mechanisms were randomly mixed, and the optimal combination was obtained as SimAMEMA, which was superior to the baseline index across all indicators.
The results show that our hybrid attention achieves a 4.5% improvement over the baseline accuracy, a 5.2% improvement over recall, a 5.1% improvement over MAP@0.5, and a 4.5% improvement over the average accuracy of mAP@0.5:0.95. Therefore, the model’s detection performance has improved. In addition, we compare our improved network of mixed attention mechanisms with several other advanced networks, as shown in Table 3:

4.4. Ablation Experiment

To evaluate the effectiveness of the mixed attention mechanism, we conducted ablation experiments on the baseline model using the scene profile dataset that we produced.
Based on the results shown in Table 4, the ablation experiments highlight the importance and effectiveness of integrating each module. Notably, the combination of the SimAM and EMAv2 modules improved the overall accuracy of YOLOv8.

5. Conclusions

In this paper, we proposed a hybrid attention mechanism based on the YOLOv8 architectural scene profile recognition method, which can optimize energy function to discover the importance of each neuron, enhance spatial semantic feature distribution by reshaping channels into batch dimensions and grouping them into sub-features, and capture pixel-level relationships using interdimensional interactions. The hybrid attention-enhanced YOLOv8 model proposed in this study achieved significant results in architectural section recognition, which has broad application prospects. In the field of architectural design, the model can be used in BIM (Building Information Modeling) systems to automatically identify component elements in architectural section drawings, improving design efficiency and accuracy. In terms of building safety monitoring, the model can be used to detect potential safety hazards in building structures in real time, such as structural deformation or non-compliant designs. The experiments show that, compared with the existing methods, our proposed mixed attention mechanism performs better in realizing the architectural scene profile dataset, achieving an mAP@0.5 of 89.9%. Precision, recall, mAP@0.5, and mAP0.5–0.95, respectively, increased by 4.5%, 5.2%, 5.1%, and 4.5%, compared with the baseline of YOLOv8 model.
Although this study has achieved good results in architectural section recognition, there are still some limitations in practical applications. First, the current model is mainly optimized for conventional building interior objects, and its recognition accuracy may decrease when applied to unconventional building structures or special components. Second, the model’s performance may be compromised in highly complex sectional images or under extremely poor lighting conditions. See Figure 4.

Author Contributions

Conceptualization, J.Y. and Z.S.; methodology, W.Z. and W.H.; software, F.X. and J.Y.; validation, J.Q. and H.W.; formal analysis, J.Y.; resources, Z.S.; writing—original draft preparation, G.L. and J.Y.; supervision, J.Q.; funding acquisition, J.Y. and F.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Quzhou City Science and Technology Plan Project (Grant No. 2023K044, 2024K182), the Natural Science Foundation of Jiangxi Province (Grant No. 20224BAB202018, 20232ABC03A04).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, X.; Gu, Y.; Li, Y.; Guan, Z. Fast volume rendering and cutting for finite element model. Comput. Aided Civ. Infrastruct. Eng. 2003, 18, 121–131. [Google Scholar] [CrossRef]
  2. Zou, Y.; Kiviniemi, A.; Jones, S.W. A review of risk management through BIM and BIM-related technologies. Saf. Sci. 2017, 97, 88–98. [Google Scholar] [CrossRef]
  3. Zhou, L.; Fan, M.; Hansen, C.; Johnson, C.R.; Weiskopf, D. A review of three-dimensional medical image visualization. Health Data Sci. 2022, 2022, 9840519. [Google Scholar] [CrossRef]
  4. Holgado-Barco, A.; Riveiro, B.; González-Aguilera, D.; Arias, P. Automatic inventory of road cross-sections from mobile laser scanning system. Comput.-Aided Civ. Infrastruct. Eng. 2017, 32, 3–17. [Google Scholar] [CrossRef]
  5. Arias, P.; Carlos Caamaño, J.; Lorenzo, H.; Armesto, J. 3D modeling and section properties of ancient irregular timber structures by means of digital photogrammetry. Comput.-Aided Civ. Infrastruct. Eng. 2007, 22, 597–611. [Google Scholar] [CrossRef]
  6. Lin, Z.; Chen, X.; Yang, H. Experimental study on structural form and excavation model of urban metro cross transfer station with super large cross section and shallow excavation. Comput. Civ. Infrastruct. Eng. 2020, 2020, 9723927. [Google Scholar] [CrossRef]
  7. Karhu, V. A view-based approach for construction process modeling. Comput. Civ. Infrastruct. Eng. 2003, 18, 275–285. [Google Scholar] [CrossRef]
  8. Yuan, X.; Liu, C.; Feng, F. Slice-mask based 3d cardiac shape reconstruction from ct volume. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 1909–1925. [Google Scholar]
  9. Hu, H. Multi-slice helical CT: Scan and reconstruction. Med. Phys. 1999, 26, 5–18. [Google Scholar] [CrossRef]
  10. Shi, P.; Qi, Q.; Qin, Y. A novel learning-based feature recognition method using multiple sectional view representation. J. Intell. Manuf. 2020, 31, 1291–1309. [Google Scholar] [CrossRef]
  11. Wen, Y.; Zhang, H.; Yu, Z. Reconstructing 3D objects from 2D sectional views of engineering drawings using volume-based method. In Proceedings of the 2010 Shape Modeling International Conference, Aix-en-Provence, France, 21–23 June 2010; pp. 13–24. [Google Scholar]
  12. Gong, J.; Zhang, H.; Jiang, B. Identification of sections from engineering drawings based on evidence theory. In Proceedings of the 2008 ACM symposium on Solid and physical Modeling, Stony Brook, NY, USA, 2–4 June 2008; pp. 13–24. [Google Scholar]
  13. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  14. LeCun, Y.; Bottou, L.; Bengio, Y. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  15. Farooq, M.U.; Bhutto, M.A.; Kazi, A.K. Real-Time Safety Helmet Detection Using YOLOv5 at Construction Sites. Intell. Autom. Soft Comput. 2023, 36, 1. [Google Scholar]
  16. Feng, R.; Miao, Y.; Zheng, J. A YOLO-Based Intelligent Detection Algorithm for Risk Assessment of Construction Sites. J. Intell. Constr. 2024, 2, 1–18. [Google Scholar] [CrossRef]
  17. Wang, J.; Wu, J.; Wu, J.; Wang, J.; Wang, J. YOLOv7 optimization model based on attention mechanism applied in dense scenes. Appl. Sci. 2023, 13, 9173. [Google Scholar] [CrossRef]
  18. Urbieta, M.; Urbieta, M.; Laborde, T.; Villarreal, G.; Rossi, G. Generating BIM model from structural and architectural plans using Artificial Intelligence. J. Build. Eng. 2023, 78, 107672. [Google Scholar] [CrossRef]
  19. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  20. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  21. Gouider, C.; Seddik, H. YOLOv4 enhancement with efficient channel recalibration approach in CSPdarknet53. In Proceedings of the 2022 IEEE Information Technologies & Smart Industrial Systems (ITSIS), Paris, France, 15–17 July 2022; pp. 1–6. [Google Scholar]
  22. Bochkovskiy, A.; Wang, C.; Liao, H. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  23. Lv, L.; Li, X.; Mao, F. A deep learning network for individual tree segmentation in UAV images with a coupled CSPNet and attention mechanism. Remote Sens. 2023, 15, 4420. [Google Scholar] [CrossRef]
  24. Mahasin, M.; Dewi, I. Comparison of CSPDarkNet53, CSPResNeXt-50, and EfficientNet-B0 backbones on YOLO v4 as object detector. Int. J. Eng. Sci. Inf. Technol. 2022, 2, 64–72. [Google Scholar] [CrossRef]
  25. Hu, M.; Li, Y.; Fang, L. A2-FPN: Attention aggregation based feature pyramid network for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 15343–15352. [Google Scholar]
  26. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A deep network architecture for pan-sharpening. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5449–5457. [Google Scholar]
  27. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10781–10790. [Google Scholar]
  28. Misra, D. Mish: A self-regularized non-monotonic activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  29. Jin, X.; Xie, Y.; Wei, X.; Zhao, B.; Chen, Z.; Tan, X. Delving deep into spatial pooling for squeeze-and-excitation networks. Pattern Recognit. 2022, 121, 108159. [Google Scholar] [CrossRef]
  30. Yang, L.; Zhang, R.Y.; Li, L. Simam. A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Washington, DC, USA, 21–24 August 2023; pp. 11863–11874. [Google Scholar]
  31. Ouyang, D.; He, S.; Zhang, G.; Luo, M.; Guo, H.; Zhan, J.; Huang, Z. Efficient multi-scale attention module with cross-spatial learning. In Proceedings of the ICASSP 2023–2023 IEEE International Conference on Acoustics, Speech and Signal Processing, Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  32. Liu, Y.; Shao, Z.; Hoffmann, N. Global attention mechanism: Retain information to enhance channel-spatial interactions. arXiv 2021, arXiv:2112.05561. [Google Scholar]
  33. Wan, Q.; Huang, Z.; Lu, J.; Yu, G.; Zhang, L. SeaFormer++: Squeeze-enhanced axial transformer for mobile visual recognition. Int. J. Comput. Vis. 2025, 133, 3645–3666. [Google Scholar] [CrossRef]
  34. Wan, Q.; Huang, Z.; Lu, J.; Yu, G.; Zhang, L. Seaformer: Squeeze-enhanced axial transformer for mobile semantic segmentation. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  35. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2020; pp. 13713–13722. [Google Scholar]
Figure 1. Workflow of YOLOv8 architectural scene profile recognition method based on hybrid attention mechanism.
Figure 1. Workflow of YOLOv8 architectural scene profile recognition method based on hybrid attention mechanism.
Sensors 25 03060 g001
Figure 2. Hybrid attention mechanism module workflow.
Figure 2. Hybrid attention mechanism module workflow.
Sensors 25 03060 g002
Figure 3. Architecture scene section plan dataset production.
Figure 3. Architecture scene section plan dataset production.
Sensors 25 03060 g003
Figure 4. Comparison of five models. (a) Original image; (b) YOLOv5; (c) YOLOv8; (d) MMDetection; (e) YOLOv6; (f) Our model.
Figure 4. Comparison of five models. (a) Original image; (b) YOLOv5; (c) YOLOv8; (d) MMDetection; (e) YOLOv6; (f) Our model.
Sensors 25 03060 g004
Table 1. Compares the detection ability of different attention mechanisms based on YOLOv8.
Table 1. Compares the detection ability of different attention mechanisms based on YOLOv8.
AttentionPrecisionRecallmAP@0.5mAP@0.5:0.95
YOLOv8.yaml78.2%79.3%84.8%72.8%
SimAM77.1%82.3%83.6%71.5%
EMA86.8%78.6%88%74.8%
GAM80.6%77.4%84.2%72.6%
SEA84.8%77.4%85.7%73.9%
CA83.7%77.8%84.4%72.7%
Table 2. Comparison of detection capabilities of hybrid attention mechanisms.
Table 2. Comparison of detection capabilities of hybrid attention mechanisms.
AttentionPrecisionRecallmAP@0.5mAP@0.5:0.95
YOLOv8.yaml78.2%79.3%84.8%72.8%
SimAM + EMA (Ours)82.7%84.5%89.9%77.3%
Table 3. Comparison of different model detection capabilities.
Table 3. Comparison of different model detection capabilities.
ModelPrecisionRecallmAP@0.5mAP@0.5:0.95
YOLOv582.4%72%81%68.3%
YOLOv878.2%79.3%84.8%72.8%
MMDetection74.4%79.7%87.7%74.4%
YOLOv682.9%78%86%73.4%
SimAM + EMA (Ours)82.7%84.5%89.9%77.3%
Table 4. Ablation experiments of mixed attention mechanism.
Table 4. Ablation experiments of mixed attention mechanism.
SimAMEMAv1EMAv2PrecisionRecallmAP@0.5mAP@0.5:0.95
78.2%79.3%84.8%72.8%
77.1%82.3%83.6%71.5%
86.8%78.6%88%74.8%
73.6%86.3%83.1%72%
83.1%78.5%83.4%71.8%
82.7%84.5%89.9%77.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, J.; Shu, Z.; Zhou, W.; Hu, W.; Qiu, J.; Xu, F.; Wang, H.; Luo, G. YOLOv8 Architectural Scene Section Recognition Method Based on SimAM-EMA Hybrid Attention Mechanism. Sensors 2025, 25, 3060. https://doi.org/10.3390/s25103060

AMA Style

Ye J, Shu Z, Zhou W, Hu W, Qiu J, Xu F, Wang H, Luo G. YOLOv8 Architectural Scene Section Recognition Method Based on SimAM-EMA Hybrid Attention Mechanism. Sensors. 2025; 25(10):3060. https://doi.org/10.3390/s25103060

Chicago/Turabian Style

Ye, Jiangang, Zilong Shu, Wen Zhou, Weijun Hu, Jinwen Qiu, Feng Xu, Hui Wang, and Guoliang Luo. 2025. "YOLOv8 Architectural Scene Section Recognition Method Based on SimAM-EMA Hybrid Attention Mechanism" Sensors 25, no. 10: 3060. https://doi.org/10.3390/s25103060

APA Style

Ye, J., Shu, Z., Zhou, W., Hu, W., Qiu, J., Xu, F., Wang, H., & Luo, G. (2025). YOLOv8 Architectural Scene Section Recognition Method Based on SimAM-EMA Hybrid Attention Mechanism. Sensors, 25(10), 3060. https://doi.org/10.3390/s25103060

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop