Next Article in Journal
Online Centralized MPC for Lane Merging in Vehicle Platoons
Previous Article in Journal
DG-TTA: Out-of-Domain Medical Image Segmentation Through Augmentation, Descriptor-Driven Domain Generalization, and Test-Time Adaptation
Previous Article in Special Issue
Atmospheric Light Estimation Using Polarization Degree Gradient for Image Dehazing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Two-Stage Marker Detection–Localization Network for Bridge-Erecting Machine Hoisting Alignment

School of Electronic and Optical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(17), 5604; https://doi.org/10.3390/s25175604
Submission received: 20 July 2025 / Revised: 30 August 2025 / Accepted: 1 September 2025 / Published: 8 September 2025

Abstract

To tackle the challenges of complex construction environment interference (e.g., lighting variations, occlusion, and marker contamination) and the demand for high-precision alignment during the hoisting process of bridge-erecting machines, this paper presents a two-stage marker detection–localization network tailored to hoisting alignment. The proposed network adopts a “coarse detection–fine estimation” phased framework; the first stage employs a lightweight detection module, which integrates a dynamic hybrid backbone (DHB) and dynamic switching mechanism to efficiently filter background noise and generate coarse localization boxes of marker regions. Specifically, the DHB dynamically switches between convolutional and Transformer branches to handle features of varying complexity (using depthwise separable convolutions from MobileNetV3 for low-level geometric features and lightweight Transformer blocks for high-level semantic features). The second stage constructs a Transformer-based homography estimation module, which leverages multi-head self-attention to capture long-range dependencies between marker keypoints and the scene context. By integrating enhanced multi-scale feature interaction and position encoding (combining the absolute position and marker geometric priors), this module achieves the end-to-end learning of precise homography matrices between markers and hoisting equipment from the coarse localization boxes. To address data scarcity in construction scenes, a multi-dimensional data augmentation strategy is developed, including random homography transformation (simulating viewpoint changes), photometric augmentation (adjusting brightness, saturation, and contrast), and background blending with bounding box extraction. Experiments on a real bridge-erecting machine dataset demonstrate that the network achieves detection accuracy (mAP) of 97.8%, a homography estimation reprojection error of less than 1.2 mm, and a processing frame rate of 32 FPS. Compared with traditional single-stage CNN-based methods, it significantly improves the alignment precision and robustness in complex environments, offering reliable technical support for the precise control of automated hoisting in bridge-erecting machines.

1. Introduction

In modern bridge engineering, bridge-erecting machines serve as core equipment for the hoisting of precast girders/bridge panels, where the alignment during hoisting directly impacts the construction efficiency and operational safety. In automated construction scenarios, the high-precision detection and localization of markers (e.g., positioning targets, reference lines) is critical in achieving accurate docking between bridge-erecting machines and precast components. However, the complexity of real construction environments poses significant challenges for visual perception systems: dynamic lighting conditions (e.g., strong light reflection, shadow occlusion), random occlusion by dust/mechanical parts, and the surface contamination of markers after long-term use can all lead to false detection, missed detection, or localization errors in traditional vision methods.
Additionally, the dual requirements of real-time performance (to match the equipment movement speed) and alignment precision (millimeter-level error tolerance) make it difficult for traditional single-stage detection–localization methods based on convolutional neural networks (CNNs) to balance efficiency and robustness. Furthermore, the scarcity of annotated data in real construction scenarios (limited by construction cycles, safety regulations, etc.) further restricts the generalization capabilities of models.
To address the high precision requirements and data scarcity in marker detection–localization under complex construction environments, this paper proposes a two-stage marker detection–localization network for bridge-erecting machine hoisting alignment, with three key innovations. First, a “coarse detection–fine estimation” phased architecture is designed: the first stage uses a lightweight detection module with a dynamic hybrid backbone and adaptive feature switching to filter background interference and generate coarse marker region boxes; the second stage constructs a Transformer-based homography estimation module, which leverages self-attention to capture long-range dependencies between marker keypoints and the scene context, enhancing multi-scale feature interaction and positional encoding to achieve the end-to-end learning of precise homography matrices. Second, a multi-dimensional data augmentation strategy simulates lighting distortion, occlusion, and marker contamination to generate diverse training samples, improving the model’s adaptability to complex environments.
Experimental results show that the proposed method achieves detection accuracy (mAP) of 97.8%, a reprojection error of less than 1.2 mm in homography estimation, and a processing frame rate of 32 FPS on a real bridge-erecting machine construction dataset. Compared with traditional single-stage methods, it significantly enhances the alignment precision and robustness in complex environments, providing reliable technical support for the precise control of automated hoisting in bridge-erecting machines.

2. Related Works

Marker detection and localization under complex construction environments have been a long-standing challenge in computer vision and construction robotics. This section reviews key advancements in three interrelated areas, namely traditional feature-based methods, deep learning-driven detection–localization frameworks, and data augmentation strategies, while highlighting their limitations in addressing the specific demands of bridge-erecting machine hoisting alignment.
Traditional Feature-Based Methods. Early efforts in marker detection relied on handcrafted feature extraction and geometric matching. The Scale-Invariant Feature Transform (SIFT) [1], a classic local feature descriptor, has been widely used to detect and match keypoints for homography estimation. SIFT’s invariance to scale and rotation makes it robust to viewpoint changes, but it struggles with dynamic lighting variations (e.g., strong reflections or shadows) and partial occlusion—common in construction sites—due to its sensitivity to intensity changes. Similarly, Harris corner detectors [2] focus on local edge responses for corner detection, but they fail to capture the global context, leading to unstable performance when markers are contaminated (e.g., by dust or paint peeling). Other feature point extractors include Speeded-Up Robust Features (SURF) [3], Oriented FAST and Rotated BRIEF (ORB) [4] and the Boosted Efficient Binary Local Image Descriptor (BEBLID) [5]. ORB builds on two foundational techniques: the Features from Accelerated Segment Test (FAST) method, a rapid corner detector that identifies keypoints by comparing the pixel intensities in a circular region, and Binary Robust Independent Elementary Features (BRIEF), a descriptor that generates binary feature vectors via random pixel pair comparisons. ORB enhances these with rotation invariance, making it suitable for real-time scenarios. These methods are widely used for marker detection but share similar limitations in complex construction environments.
Deep Learning for Detection–Localization. The rise of deep learning has spurred the development of end-to-end detection–localization frameworks. Single-stage detectors like You Only Look Once (YOLO) [6] and the Single-Shot MultiBox Detector (SSD) [7] achieve real-time speeds by directly predicting bounding boxes and class scores in a single pass. However, their focus on efficiency often sacrifices precision, especially for small or occluded markers. For instance, YOLOv8 [8], a state-of-the-art single-stage model, struggles to distinguish markers from cluttered backgrounds (e.g., machinery or debris) in construction scenes, resulting in high false-positive rates.
Two-stage frameworks, such as the Faster Region-Based Convolutional Neural Network (Faster R-CNN) [9], improve the accuracy by first generating region proposals and then refining them. Faster R-CNN’s region proposal network (RPN) better handles object scale variations, but its computational complexity limits real-time performance—critical in aligning fast-moving hoisting equipment. Moreover, both single- and two-stage CNN-based methods rely on convolutional layers, which inherently model local spatial correlations but fail to capture long-range dependencies between marker keypoints and the scene context (e.g., the relationship between marker corners and nearby machinery). This limitation restricts their ability to estimate precise homography matrices under occluded or contaminated conditions.
Homography Estimation with Deep Learning. Homography estimation, a core component of marker localization, has been explored using CNNs. Recent advances in deep learning have revolutionized homography estimation by replacing traditional multi-step pipelines with end-to-end trainable frameworks. These methods can be broadly categorized into supervised, unsupervised, and hybrid approaches (e.g., self-supervised, semi-supervised), each addressing distinct challenges in geometric alignment [10].
Supervised methods leverage synthetic or labeled datasets to train networks for direct homography regression. Early work by DeTone et al. [11] introduced the four-point parameterization of the homography matrix, enabling efficient network training by regressing offsets of keypoints rather than the full 3 × 3 matrix. Subsequent studies optimized network architectures for efficiency, such as lightweight models based on ShuffleNet [12], which reduced the parameters to under 9 MB while maintaining accuracy, making them suitable for edge devices. Hybrid frameworks like HomoNetComb [13] combined deep learning with energy minimization, using CNNs to predict initial homographies and gradient descent to refine residuals, balancing speed and precision. Transformer-based models further improved the performance by capturing long-range feature dependencies, with attention mechanisms enhancing alignment in large-baseline scenarios [14]. However, supervised methods face limitations in generalizing to real-world scenes due to the scarcity of labeled data and synthetic-to-real domain gaps.
Unsupervised methods eliminate the need for labeled data by optimizing the photometric consistency between warped and target images. A key innovation was the introduction of homography flow [15], a low-rank representation of optical flow constrained by the homography subspace, enabling robust estimation by focusing on dominant planar motion. Generative adversarial networks (GANs) were also applied to enforce coplanarity constraints, where discriminators distinguish between original and warped images to guide homography prediction [16]. Content-aware masks [17] and contextual correlation layers [18] were integrated to suppress dynamic objects and occlusions, improving the robustness in low-texture or noisy environments. Multi-scale and cascaded network structures [19] further refined coarse-to-fine estimation, reducing the reprojection errors for large displacements. Despite progress, unsupervised methods struggle with training stability and may fail in scenes with significant depth variations.
Despite advancements, existing deep learning methods face trade-offs between accuracy, speed, and robustness in dynamic construction environments. Supervised models lack adaptability to unlabeled real-world data, while unsupervised approaches struggle with large parallax and occlusion. Single-stage CNNs often sacrifice precision for efficiency, failing to meet the millimeter-level alignment requirements of bridge-erecting machines. Thus, a two-stage framework combining lightweight detection and Transformer-based fine estimation is proposed to address these challenges.
Computer Vision in Civil Engineering and Construction. Computer vision is widely applied in civil engineering and construction, facilitating tasks like safety monitoring (e.g., detecting non-hardhat wearers [20]), tracking workers and equipment [21], and progress monitoring via 3D point clouds [22]. It enables activity recognition for earthmoving operations [23] and interactions between excavators and dump trucks [24]. Techniques include CNNs for object detection [25] and two-stream networks for worker activity recognition [26], using data from cameras, scanners, and unmanned aerial vehicles (UAVs) [27,28]. These applications address site complexity, enhancing safety, productivity, and decision making.
Data Augmentation for Construction Scenes. Data scarcity is a critical bottleneck in training robust models for construction environments, where annotated images are limited due to safety regulations and long construction cycles. Synthetic data generation, such as the SYNTHIA dataset [29] for urban scenes, uses 3D rendering to simulate diverse environments. However, these datasets focus on urban settings (e.g., streets and buildings) and fail to simulate construction-specific interferences like dust occlusion, marker contamination, or dynamic lighting from heavy machinery. Existing augmentation strategies for object detection (e.g., random cropping or flipping) [30] also do not address construction-specific challenges, leaving models poorly adapted to real-world scenarios.

3. Methodology

3.1. Overall Architecture

The proposed framework adopts a two-stage “coarse detection–fine estimation” architecture (Figure 1).
  • Stage 1: Lightweight Marker Detection Module for the rapid coarse localization of markers in complex environments.
  • Stage 2: Transformer-Based Homography Estimation Module for precise coordinate transformation between markers and hoisting equipment.

3.2. Stage 1: Lightweight Marker Detection Module

As shown in Figure 2, this stage focuses on efficient coarse localization through the dynamic hybrid block, which consists of two key components: a dynamic hybrid backbone (DHB) and a dynamic switching mechanism. Multi-scale feature maps are generated by the dynamic hybrid blocks, and these feature maps are further processed by the neck and detection head, which are consistent with YOLOv8 [8].

Dynamic Hybrid Block (DHB)

To balance geometric feature extraction and complex texture modeling, the DHB dynamically switches between convolutional and Transformer branches based on feature complexity, as shown in Figure 3. The input feature is processed by two branches of feature processing network and a dynamic switching block. The two-branch network is designed as follows:
  • The first branch uses MobileNetV3’s depthwise separable convolutions to extract geometric features (edges, corners). The depthwise convolution operation is defined as
    F d w = W d w X + b d w
    where W d w R k × k × C is the depthwise kernel, k is the kernel size, C is the input channel, ⊛ denotes depthwise convolution, and b d w is the bias.
  • The second branch introduces lightweight Transformer blocks with scene-aware attention. The window multi-head self-attention (W-MSA) is computed as
    W - MSA ( Z ) = Concat ( head 1 , , head h ) W O
    where head i = Attention ( Z W i Q , Z W i K , Z W i V ) , and a scene prior bias B (encoding marker aspect ratio/position distribution) is added to the attention scores:
    Attention ( Q , K , V ) = Softmax Q K d k + B V
For the dynamic switching module, a feature complexity discriminator calculates the entropy H ( F ) of feature map F:
H ( F ) = i = 1 H × W F i F 1 log F i F 1
The gating weight α [ 0 , 1 ] (via a sigmoid function) determines the branch contribution:
α = σ H ( F ) τ δ , F stage = α · F trans + ( 1 α ) · F conv
where τ = 3.5 (threshold), δ = 0.5 (scaling factor). The threshold τ acts as a criterion to judge the feature complexity. It determines whether the input feature maps have more high-level semantic information (when entropy H ( F ) > τ ) or low-level geometric information (when H ( F ) < τ ). The scaling factor ( δ = 0.5 ) controls the smoothness of weight switching via the sigmoid function. A smaller δ causes the weight change to be steeper. Here, it balances sensitivity and smoothness for stable adaptation to complex construction scenarios.
Threshold τ = 3.5 is derived from an entropy distribution analysis of 500 training images, where H ( F ) > 3.5 indicated high-level semantic features (e.g., cluttered backgrounds) and H ( F ) < 3.5 indicated low-level geometric features (e.g., marker edges). Testing τ { 2.5 ,   3.0 ,   3.5 ,   4.0 } showed that this value minimized feature type misclassification.
Scaling factor δ = 0.5 controls sigmoid smoothness. Evaluating δ { 0.3 ,   0.5 ,   0.7 } revealed that this value balanced sensitivity to feature changes and training stability, avoiding abrupt weight fluctuations in mixed-complexity scenarios (e.g., partially occluded markers).

3.3. Stage 2: Transformer-Based Homography Estimation Module

This stage refines marker localization using the coarse detection results from Stage 1, focusing on modeling long-range dependencies between marker keypoints and the scene context—critical for precise homography estimation under occlusion or contamination. The module architecture (Figure 4) consists of three core components: input feature processing, a Transformer encoder with fused positional encoding, and a homography decoding head.

3.3.1. Input Processing

Regions of interest (ROIs) cropped from the Stage 1 outputs (resized to 256 × 256) are fed into ResNet50’s convolutional layers (C3–C5), which extract multi-scale features with 256, 512, and 1024 channels, respectively. These features are concatenated and flattened into a sequence X R H × W × C , where H = W = 16 and C = 1024 for the final feature map, preserving the spatial and semantic information necessary for keypoint correlation.

3.3.2. Transformer Encoder

The encoder (6 layers) uses multi-head self-attention to model the relationships between all pairs of features in the sequence, enabling the capture of long-range dependencies (e.g., between marker corners and nearby structural edges). A critical design is the fused positional encoding, which combines the following:
  • Absolute positional encoding: Encodes pixel coordinates ( x , y ) in the ROI to preserve spatial layout.
  • Geometric prior encoding: Encodes marker-specific priors (center coordinates: ( center x , center y ) , width: w, height: h) derived from Stage 1’s coarse bounding box, anchoring attention to the marker geometry.
The positional encoding equation can be written as
X pos = X + PE ( center x , center y , w , h )
where PE ( · ) encodes the marker center coordinates and aspect ratio. The multi-head attention operation is
MultiHead ( Q , K , V ) = Concat ( head 1 , , head h ) W O .
This splits features into subspaces to model diverse correlations (e.g., local shape vs. global context).

3.3.3. Homography Decoding

After encoding, global average pooling (GAP) aggregates the sequence into a compact feature vector X ^ R C , capturing the global scene context. A 3-layer multi-layer perceptron (MLP) maps X ^ to a 9-dimensional vector (flattened 3 × 3 matrix), normalized by the Frobenius norm to ensure scale invariance:
H = MLP ( G A P ( X e n c ) ) H F
where H F = i , j H i , j 2 prevents scale ambiguity in homography regression.

3.4. Multi-Dimensional Data Augmentation

To address the challenge of limited annotated data for object detection tasks, we propose a synthetic sample generation pipeline that combines geometric transformations, photometric augmentations, and background integration. This approach generates diverse training samples with corresponding YOLO-formatted bounding box annotations, simulating real-world variations in object appearance and context.

3.4.1. Random Homography Transformation

First, we apply a small-magnitude random homography transformation to the target template to simulate viewpoint variations (e.g., rotation, scaling, shearing, and translation). The transformation matrix H R 3 × 3 is constructed with perturbations bounded by a parameter max_perturb (default: 0.1 ), which controls the maximum relative deviation from the identity matrix. Specifically,
  • Scaling factors for x- and y-axes are perturbed by ± max_perturb ;
  • Shearing terms H 0 , 1 and H 1 , 0 are perturbed by ± max_perturb / 2 ;
  • Translation offsets H 0 , 2 and H 1 , 2 are limited to ± max_perturb × template_width and ± max_perturb × template_width , respectively.
After transformation, the output size is dynamically adjusted to avoid cropping by aligning the minimum transformed coordinates to the top-left corner of the canvas.

3.4.2. Photometric Augmentation

To enhance the sample diversity, we apply photometric transformations to the transformed template. This includes the following:
  • Color space adjustments: Random brightness ( color_range [ 0.7 , 1.3 ] ) and saturation modifications in the HSV space, followed by contrast ( α [ 0.7 , 1.3 ] ) and brightness offset ( β [ 20 , 20 ] ) adjustments in the BGR space;
  • Gaussian noise injection: Additive Gaussian noise with variance noise_var [ 5 , 20 ] to simulate sensor noise or low-light conditions.

3.4.3. Background Integration and Bounding Box Extraction

The augmented template is then randomly pasted onto a background image (selected from a predefined dataset: VOC2007) while ensuring full containment within the background dimensions. The VOC2007 dataset is part of the PASCAL VOC project; it encompasses 9963 images and covers diverse scenes like urban streets, indoor rooms, and natural landscapes. A binary mask (thresholded at 1 pixel intensity) is used to identify the non-black (foreground) region of the transformed template, from which the minimum bounding box ( x 1 , y 1 , x 2 , y 2 ) is extracted. This bounding box is translated to absolute coordinates relative to the background image. Several examples of generated image samples are shown in Figure 5.

3.5. Loss Function

The first-stage detection loss follows the YOLOv11 formulation, which includes localization, confidence, and classification components, as defined in the original YOLOv11 architecture.
The second-stage loss for homography matrix H R 3 × 3 uses Smooth L1 to regress the predicted H ^ to the ground-truth H :
L homo = 1 9 k = 1 9 SmoothL 1 H ^ k H k

4. Experimental Setup and Result Analysis

4.1. Experimental Setup

The proposed network is trained on a PC with an NVIDIA GeForce RTX 2080Ti GPU (Santa Clara, CA, USA) and 64 GB RAM. The network is implemented with PyTorch 1.10.0 with CUDA 11.3. Training employs a learning rate of 0.001 and batch size of 16 for 100 epochs, with the Adam optimizer and cosine annealing for learning rate scheduling.

4.2. Experimental Datasets and Preprocessing

The Third Construction Co., Ltd. of China Construction Eighth Engineering Bureau, Nanjing, China.

4.2.1. Real Construction Dataset

The dataset was collected from 3 bridge construction sites (The Third Construction Co., Ltd. of China Construction Eighth Engineering Bureau, Nanjing, China) over 2 months (April–May 2024) using a Intel RealSense Depth Camera D455 (Santa Clara, CA, USA) (12 MP resolution, 30 fps) mounted on the bridge-erecting machine’s hoist arm. The dataset includes 2000 images with the following:
  • Markers: 5 types of rectangular markers (10–20cm in size) affixed to precast girders;
  • Annotations: Bounding boxes (labeled via LabelImg v1.8.5) and 4 corner coordinates (manually verified for sub-pixel accuracy using OpenCV4.0’s cornerSubPix);
  • Disturbances: 32% with lighting variations (morning/afternoon sun, overcast), 28% with partial occlusion (crane arms, worker bodies), 21% with contamination (dust, paint peeling)—a distribution aligned with field observations.

4.2.2. Augmented Dataset

Following the multi-dimensional augmentation strategy described in Section 3.4, 5000 synthetic images were generated by simulating interference factors (lighting distortion, random occlusion, marker contamination). These synthetic data were mixed with real data at a 1:1 ratio for training.

4.3. Evaluation Metrics

For the detection performance evaluation, the mean average precision ( m A P ) with an intersection over union threshold of 0.5 was used to evaluate the marker detection accuracy.
For the localization precision evaluation, the homography reprojection error is adopted, which is the average pixel difference between the predicted and ground-truth coordinates of marker corners after applying the estimated homography matrix.
For the real-time performance evaluation, the frames per second (FPS) was measured to evaluate the computational efficiency.

Robustness

Robustness was assessed by measuring the m A P and treprojection error on three challenging subsets:
  • Images with lighting variations;
  • Images with partial marker occlusion;
  • Images with contaminated markers (e.g., dust adhesion, paint peeling).

4.4. Comparative Experiments

Three comparative methods were tested to validate the proposed network’s superiority:
  • Single-stage detection + traditional homography: YOLOv8 (single-stage detector) combined with SIFT feature matching for homography estimation;
  • Two-stage CNN network: Faster R-CNN (two-stage detector) followed by a CNN-based homography regression module;
  • Proposed method without data augmentation: Identical to the proposed network but trained using only real data (no augmented samples).

4.5. Ablation Studies

Three key components were ablated to verify their contributions:
  • Transformer module: Performance comparison between the proposed network and a variant where the Transformer-based homography estimation module was replaced with a CNN regression network.
  • Data augmentation: Comparison of m A P and reprojection error with/without using augmented training data.
  • Positional encoding: Evaluation of three variants:
    • Using only absolute positional encoding;
    • Using only geometric prior encoding (marker center coordinates and aspect ratio);
    • Using the proposed fused positional encoding (absolute + geometric).

4.6. Result Analysis

4.6.1. Quantitative Results

Table 1 presents the performance metrics as mean values ± standard deviations from three independent training runs, confirming the result stability. The proposed method exhibits the lowest variability across all metrics (e.g., mAP std. dev. = 0.4%), indicating robust convergence and consistent performance.
During inference, the lightweight marker detection in Stage 1 has average memory usage of 4.2 GB and an inference time of 19 ms per image (1280 × 960). The homography estimation in Stage 2 has average memory usage of 5.8 GB and an inference time of 12 ms per image. The total end-to-end inference time is 31 ms per image.

4.6.2. Ablation Study Results

Transformer Module vs. CNN Regression. Table 2 shows that replacing the Transformer-based homography module with a CNN regression network reduces the mAP by 2.8% and increases the reprojection error by 0.8 mm. This gap stems from CNNs’ focus on local correlations; thus, they fail to model long-range dependencies between marker keypoints and distant scene structures—critical for handling occlusion or contamination in construction. The Transformer’s multi-head attention captures these global relationships, enabling robust estimation even when markers are partially obscured (e.g., by crane arms), whereas CNNs’ limited receptive fields lead to underfitting.
Impact of Data Augmentation. Without data augmentation, the mAP drops by 1.7% and the reprojection error rises by 0.6 mm, confirming that our multi-dimensional strategy (random homography, photometric changes, background blending) mitigates overfitting to limited real data. Synthetic samples simulate dust, variable lighting, and other disturbances, introducing diverse texture and illumination variations. This forces the model to learn invariant features, enhancing its generalization to the highly variable conditions of real construction sites.
Positional Encoding Variants. Fused positional encoding (absolute + geometric priors) outperforms single components, with a 1.3% higher mAP and 0.4 mm lower error than absolute encoding alone and a 1.5% higher mAP with a 0.5 mm lower error than geometric priors alone. Absolute encoding preserves fine-grained spatial details (e.g., marker edges), while geometric priors (center, dimensions from Stage 1) provide structural constraints. Their fusion balances local precision and the global context, preventing drift from noise or occlusion—vital for millimeter-level hoisting alignment.
Table 3 shows additional experiments on data ratios. The 2:5 ratio outperforms the others, confirming its rationality as it balances real-world representation and synthetic diversity. It is noted that the current disturbance definitions (e.g., marker contamination, occlusion) are simplified representations of real-world complexities. While we focused on dust and paint peeling for contamination and random shapes for occlusion, real scenarios may involve oil stains, rust, linear obstructions (e.g., mechanical arms), or diffuse obstructions (e.g., dust clouds). However, our data augmentation strategy prioritized capturing core disturbance features: texture degradation for contamination and spatial obstruction for occlusion. The proposed two-stage network, with its dynamic feature adaptation (via DHB) and long-range dependency modeling (via Transformer), learns generalized patterns beyond specific disturbance types. This is supported by its robust performance on the existing disturbed subsets, suggesting potential adaptability to unlisted disturbances. Future work will expand the disturbance library to cover more specific cases, further validating the model’s generalization ability.

4.6.3. Scalability Analysis

To assess the performance under varying operational conditions, we generated test data by randomly stitching two to four original images (640 × 480) to create higher resolutions (1280 × 960, 1920 × 1440, 2880 × 2160), with 2, 4–5, and 6–8 markers per frame, respectively. The results are shown in Table 4. The model maintains high accuracy (mAP > 96%) and low errors (<1.6 mm) across the stitched conditions.

4.6.4. Qualitative Analysis

Figure 6 presents the partial marker detection results. The detector can effectively detect the targets, with no false positives or false negatives observed. Figure 7 shows the precise estimation results for homography (the green bounding box in the figure). We can see that the output results of the detector cannot accurately locate the boundaries of the markers. However, after our homography estimation, the obtained bounding boxes precisely align with the markers, thus meeting the requirement for high-precision localization.

4.6.5. Robustness Verification

In extreme scenarios, the results are as follows:
  • Occluded markers: m A P = 95.6 % (as shown in Table 1);
  • Contaminated markers: reprojection error = 1.5 mm .
These results validate the network’s strong engineering applicability under complex construction conditions.

5. Conclusions and Discussion

The proposed two-stage network, which integrates lightweight detection with Transformer-driven homography estimation, successfully achieves high-precision and real-time marker detection and localization in complex construction environments. Additionally, the multi-dimensional data augmentation strategy effectively mitigates the challenge of limited real-world data availability, thereby enhancing the model’s generalization performance. Furthermore, the experimental results validate the robustness and practicality of the proposed method in engineering scenarios, providing critical technical support for automated hoisting operations by bridge-erecting machines.
To further advance this research, several key directions are identified. First, multi-modal fusion will be explored by incorporating laser point cloud or infrared sensor data to improve the alignment accuracy under low-light or dusty conditions. Second, model lightweighting will be pursued through techniques such as knowledge distillation or quantization compression, aiming to reduce the computational complexity and enable deployment on embedded devices (e.g., bridge-erecting machine controllers) while maintaining real-time performance. Third, dynamic scenario extension will be addressed by introducing temporal modeling (e.g., combining Transformer with LSTM) to handle dynamically moving markers during hoisting, thereby enhancing the stability of continuous frame alignment.

Author Contributions

Methodology, L.L. and Z.X.; software, L.L. and T.H.; validation, L.L. and T.H.;writing—original draft preparation, L.L.; writing—review and editing, L.L., T.H. and Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  2. Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988; British Machine Vision Association (BMVA): London, UK, 1988; Volume 15, pp. 10–5244. [Google Scholar]
  3. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  4. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2564–2571. [Google Scholar]
  5. Suárez, I.; Sfeir, G.; Buenaposada, J.M.; Baumela, L. BEBLID: Boosted efficient binary local image descriptor. Pattern Recognit. Lett. 2020, 133, 366–372. [Google Scholar] [CrossRef]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, IL, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  7. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  8. Varghese, R.; Sambath, M. Yolov8: A novel object detection algorithm with enhanced performance and robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–6. [Google Scholar]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  10. Luo, Y.; Wang, X.; Liao, Y.; Fu, Q.; Shu, C.; Wu, Y.; He, Y. A review of homography estimation: Advances and challenges. Electronics 2023, 12, 4977. [Google Scholar] [CrossRef]
  11. DeTone, D.; Malisiewicz, T.; Rabinovich, A. Deep image homography estimation. arXiv 2016, arXiv:1606.03798. [Google Scholar] [CrossRef]
  12. Wang, G.; You, Z.; An, P.; Yu, J.; Chen, Y. Efficient and robust homography estimation using compressed convolutional neural network. In Proceedings of the Digital TV and Multimedia Communication: 15th International Forum, IFTC 2018, Shanghai, China, 20–21 September 2018; Revised Selected Papers 15. Springer: Berlin/Heidelberg, Germany, 2019; pp. 156–168. [Google Scholar]
  13. Kang, L.; Wei, Y.; Xie, Y.; Jiang, J.; Guo, Y. Combining convolutional neural network and photometric refinement for accurate homography estimation. IEEE Access 2019, 7, 109460–109473. [Google Scholar] [CrossRef]
  14. Zhou, H.; Hu, W.; Li, Y.; He, C.; Chen, X. Deep Homography Estimation With Feature Correlation Transformer. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1397–1402. [Google Scholar]
  15. Ye, N.; Wang, C.; Fan, H.; Liu, S. Motion basis learning for unsupervised deep homography estimation with subspace projection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 13117–13125. [Google Scholar]
  16. Hong, M.; Lu, Y.; Ye, N.; Lin, C.; Zhao, Q.; Liu, S. Unsupervised homography estimation with coplanarity-aware gan. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 17663–17672. [Google Scholar]
  17. Zhang, J.; Wang, C.; Liu, S.; Jia, L.; Ye, N.; Wang, J.; Zhou, J.; Sun, J. Content-aware unsupervised deep homography estimation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part I 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 653–669. [Google Scholar]
  18. Nie, L.; Lin, C.; Liao, K.; Liu, S.; Zhao, Y. Depth-aware multi-grid deep homography estimation with contextual correlation. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4460–4472. [Google Scholar] [CrossRef]
  19. Hou, B.; Ren, J.; Yan, W. Unsupervised Multi-Scale-Stage Content-Aware Homography Estimation. Electronics 2023, 12, 1976. [Google Scholar] [CrossRef]
  20. Fang, Q.; Li, H.; Luo, X.; Ding, L.; Luo, H.; Rose, T.M.; An, W. Detecting non-hardhat-use by a deep learning method from far-field surveillance videos. Autom. Constr. 2018, 85, 1–9. [Google Scholar] [CrossRef]
  21. Zhu, Z.; Ren, X.; Chen, Z. Visual tracking of construction jobsite workforce and equipment with particle filtering. J. Comput. Civ. Eng. 2016, 30, 04016023. [Google Scholar] [CrossRef]
  22. Sharif, M.M.; Nahangi, M.; Haas, C.; West, J. Automated model-based finding of 3D objects in cluttered construction point cloud models. Comput.-Aided Civ. Infrastruct. Eng. 2017, 32, 893–908. [Google Scholar] [CrossRef]
  23. Golparvar-Fard, M.; Heydarian, A.; Niebles, J.C. Vision-based action recognition of earthmoving equipment using spatio-temporal features and support vector machine classifiers. Advanced Engineering Informatics 2013, 27, 652–663. [Google Scholar] [CrossRef]
  24. Kim, J.; Chi, S.; Seo, J. Interaction analysis for vision-based activity identification of earthmoving excavators and dump trucks. Autom. Constr. 2018, 87, 297–308. [Google Scholar] [CrossRef]
  25. Fang, W.; Ding, L.; Zhong, B.; Love, P.E.; Luo, H. Automated detection of workers and heavy equipment on construction sites: A convolutional neural network approach. Adv. Eng. Inform. 2018, 37, 139–149. [Google Scholar] [CrossRef]
  26. Luo, X.; Li, H.; Cao, D.; Yu, Y.; Yang, X.; Huang, T. Towards efficient and objective work sampling: Recognizing workers’ activities in site surveillance videos with two-stream convolutional networks. Autom. Constr. 2018, 94, 360–370. [Google Scholar] [CrossRef]
  27. Jeelani, I.; Han, K.; Albert, A. Automating and scaling personalized safety training using eye-tracking data. Autom. Constr. 2018, 93, 63–77. [Google Scholar] [CrossRef]
  28. Irizarry, J.; Costa, D.B. Exploratory study of potential applications of unmanned aerial systems for construction management tasks. J. Manag. Eng. 2016, 32, 05016001. [Google Scholar] [CrossRef]
  29. Ros, G.; Sellart, L.; Materzynska, J.; Vazquez, D.; Lopez, A.M. The synthia dataset: A large collection of synthetic images for semantic segmentation of urban scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3234–3243. [Google Scholar]
  30. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed method. Red boxes denotes the results of marker detection, green boxes denotes the results of homography estimation.
Figure 1. The overall framework of the proposed method. Red boxes denotes the results of marker detection, green boxes denotes the results of homography estimation.
Sensors 25 05604 g001
Figure 2. The structure of the marker detection network.
Figure 2. The structure of the marker detection network.
Sensors 25 05604 g002
Figure 3. The detailed structure of the dynamic hybrid block.
Figure 3. The detailed structure of the dynamic hybrid block.
Sensors 25 05604 g003
Figure 4. Framework of the Transformer-based homography estimation module. Input ROIs are processed by ResNet50 to extract multi-scale features, which are fed into a Transformer encoder with fused positional encoding. Global average pooling (GAP) and a multi-layer perceptron (MLP) then regress the 3 × 3 homography matrix.
Figure 4. Framework of the Transformer-based homography estimation module. Input ROIs are processed by ResNet50 to extract multi-scale features, which are fed into a Transformer encoder with fused positional encoding. Global average pooling (GAP) and a multi-layer perceptron (MLP) then regress the 3 × 3 homography matrix.
Sensors 25 05604 g004
Figure 5. Examples of generated image samples.
Figure 5. Examples of generated image samples.
Sensors 25 05604 g005
Figure 6. Examples of detection results.
Figure 6. Examples of detection results.
Sensors 25 05604 g006
Figure 7. Examples of homography estimation results.
Figure 7. Examples of homography estimation results.
Sensors 25 05604 g007
Table 1. Comparison of Detection, Localization and Real-Time Performance (mean ± std. dev. over 3 runs).
Table 1. Comparison of Detection, Localization and Real-Time Performance (mean ± std. dev. over 3 runs).
MethodDetectionLocalizationReal-TimeRobustness
mAP (%) Reprojection Error (mm) FPS mAP occlusion (%)
YOLOv8+SIFT90.2 ± 0.83.5 ± 0.345 ± 182.1 ± 1.2
Faster R-CNN+CNN93.4 ± 0.62.3 ± 0.238 ± 188.7 ± 0.9
Ours (no augmentation)96.1 ± 0.51.8 ± 0.232 ± 192.4 ± 0.7
Proposed Method97.8 ± 0.41.2 ± 0.132 ± 195.6 ± 0.5
Table 2. Ablation study on key components.
Table 2. Ablation study on key components.
Variant mAP (%)Reprojection Error (mm)FPS
Full model (proposed)97.81.232
Without Transformer (CNN regression)95.02.033
Without data augmentation96.11.832
Absolute pos. encoding only96.51.632
Geometric prior encoding only96.31.732
Table 3. Performance under different real-to-synthetic data ratios.
Table 3. Performance under different real-to-synthetic data ratios.
Ratio (Real–Synthetic) mAP (%)Reprojection Error (mm)FPS
1:1 (2000:2000)96.91.432
2:5 (2000:5000)97.81.232
3:5 (3000:5000)97.21.332
Table 4. Performance under stitched images (varying resolutions/marker counts).
Table 4. Performance under stitched images (varying resolutions/marker counts).
Stitched Condition mAP (%) Reprojection Error (mm)FPS
1280 × 960, 2 markers97.31.330
1920 × 1440, 4–5 markers96.51.512
2880 × 2160, 6–8 markers96.11.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, L.; Xiao, Z.; Hu, T. Two-Stage Marker Detection–Localization Network for Bridge-Erecting Machine Hoisting Alignment. Sensors 2025, 25, 5604. https://doi.org/10.3390/s25175604

AMA Style

Li L, Xiao Z, Hu T. Two-Stage Marker Detection–Localization Network for Bridge-Erecting Machine Hoisting Alignment. Sensors. 2025; 25(17):5604. https://doi.org/10.3390/s25175604

Chicago/Turabian Style

Li, Lei, Zelong Xiao, and Taiyang Hu. 2025. "Two-Stage Marker Detection–Localization Network for Bridge-Erecting Machine Hoisting Alignment" Sensors 25, no. 17: 5604. https://doi.org/10.3390/s25175604

APA Style

Li, L., Xiao, Z., & Hu, T. (2025). Two-Stage Marker Detection–Localization Network for Bridge-Erecting Machine Hoisting Alignment. Sensors, 25(17), 5604. https://doi.org/10.3390/s25175604

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop