Next Article in Journal
AC-Diagnostics of Transport Phenomena in Dilute Suspensions
Next Article in Special Issue
Uncertainty Evaluation of CMM and Optical 3D Scanning in Centrifugal Rotor Inspection
Previous Article in Journal
Recent Advances in Digital Fringe Projection Profilometry (2022–2025): Techniques, Applications, and Metrological Challenges—A Review
Previous Article in Special Issue
The Critical Role of International Comparisons in Global Metrology System: An Overview
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Flexible Wheel Alignment Measurement Method via APCS-SwinUnet and Point Cloud Registration

1
School of Artificial Intelligence, Changsha University of Science and Technology, Changsha 410114, China
2
Department of Mechanical Engineering, Politecnico di Milano, 20156 Milan, Italy
3
College of Electrical and Information Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
Metrology 2026, 6(1), 4; https://doi.org/10.3390/metrology6010004
Submission received: 21 November 2025 / Revised: 23 December 2025 / Accepted: 5 January 2026 / Published: 12 January 2026
(This article belongs to the Special Issue Applied Industrial Metrology: Methods, Uncertainties, and Challenges)

Abstract

To achieve low-cost and flexible wheel angles measurement, we propose a novel strategy that integrates wheel segmentation network with 3D vision. In this framework, a semantic segmentation network is first employed to extract the wheel rim, followed by angle estimation through ICP-based point cloud registration. Since wheel rim extraction is closely tied to angle computation accuracy, we introduce APCS-SwinUnet, a segmentation network built on the SwinUnet architecture and enhanced with ASPP, CBAM, and a hybrid loss function. Compared with traditional image processing methods in wheel alignment, APCS-SwinUnet delivers more accurate and refined segmentation, especially at wheel boundaries. Moreover, it demonstrates strong adaptability across diverse tire types and lighting conditions. Based on the segmented mask, the wheel rim point cloud is extracted, and an iterative closest point algorithm is then employed to register the target point cloud with a reference one. Taking the zero-angle condition as the reference, the rotation and translation matrices are obtained through point cloud registration. These matrices are subsequently converted into toe and camber angles via matrix-to-angle transformation. Experimental results verify that the proposed solution enables accurate angle measurement in a cost-effective, simple, and flexible manner. Furthermore, repeated experiments further validate its robustness and stability.

Graphical Abstract

1. Introduction

Wheel alignment is a routine yet essential maintenance procedure that ensures the wheels are properly angled relative to the vehicle’s frame and the road surface [1]. Manufacturers set the wheel alignment within a specified range before the vehicle leaves the factory [2]. However, factors such as suspension wear, heavy loads, poor road conditions, collisions, or chassis damage can cause wheel misalignment, which refers to a deviation from the manufacturer’s specifications [3]. Wheel misalignment may involve numerous issues, including uneven tire wear, wheel shimmy, suspension damage, poor handing, increased fuel consumption, and safety risk. Therefore, regular wheel alignment inspections and adjustments are crucial. During maintenance, three primary angles, namely camber, caster and toe angles, are usually measured and adjusted. Toe and camber angles of the vehicle wheels are the two alignment parameters most commonly inspected and adjusted during routine maintenance and post-repair operations, and they are also the primary focus of most existing research on vision-based wheel alignment. Therefore, at the current stage, this work primarily focuses on the estimation of toe and camber angles.

1.1. Challenges of 3D Vision-Based Methods

Compared to methods based on inertial measurement unit and 2D vision, the 3D vision-based measurement methods offer advantages such as low cost, high precision, and greater flexibility in application. However, 3D vision-based wheel alignment still faces several challenges, which are summarized below.
(1)
Complicated calibration and limited flexibility: Most passive and active 3D measurement systems require a complex and rigorous calibration procedure to reconstruct the wheel shape. Once calibrated, cameras and light sources need to remain fixed, which limits its flexibility and application range. Moreover, the reflective surface of the wheel hub leads to low texture and feature information, which affects stereo matching accuracy and ultimately reduces the accuracy of 3D reconstruction.
(2)
Additional target board and potential damage: Commercial scanner-based methods often require mounting a clamped target board with a special reflective film onto the wheel. This process is time-consuming, may cause secondary damage to the wheel, and incurs additional costs due to the consumable film, thereby reducing overall efficiency.
(3)
Inefficient and noise-sensitive full-cloud registration: Directly using the entire 3D point cloud acquired by the sensor to estimate wheel angles not only increases computational cost and reduces measurement efficiency, but also introduces interference from background, noise, and points from the vehicle body during point-cloud registration. These irrelevant points can significantly degrade registration accuracy. Therefore, it is more effective to perform 3D registration and angle estimation using only a stable, wheel-related subset of the point cloud.
(4)
Challenges in precise wheel segmentation: The reconstructed point cloud includes not only the wheel but also background elements, making accurate region-of-interest (ROI) extraction essential. Existing methods often rely on traditional image processing algorithms (e.g., Canny edge detection and Hough transforms), which are sensitive to background clutter and environmental changes. These methods often require manual parameter tuning and lack the precision needed for accurate wheel boundary and detail extraction.

1.2. Outline of Our Work

To address the challenges outlined above, this paper presents a low-cost and practical wheel alignment method that integrates semantic segmentation with point cloud registration. The proposed approach overcomes the limitations of using 2D or 3D single-modality data alone and fully leverages the rich texture information in 2D images together with the depth and pose information encoded in 3D point clouds, thereby improving both the efficiency and the accuracy of wheel angle measurement. The primary application context of this work is post-production wheel alignment in small and medium-sized garages and workshops, where vehicles are serviced after leaving the factory. The main contributions are summarized as follows.
(1)
As a relatively stable region of the wheel, the wheel rim provides reliable geometric cues, so using only the rim point cloud to estimate wheel angles helps improve both measurement efficiency and angle accuracy. Building on this idea, the proposed method reformulates the wheel alignment task as a pipeline of wheel rim mask extraction, corresponding point cloud registration, and wheel angle calculation. Compared with existing approaches, it offers greater flexibility, lower cost, and higher efficiency, as it removes the need for complex calibration procedures, target boards, and additional auxiliary equipment or materials.
(2)
Wheel rim extraction is critical for accurate angle estimation. To enhance the precision of wheel rim segmentation, we propose APCS-SwinUnet, a task-driven adaptation of Swin-Unet for wheel rim extraction. Atrous spatial pyramid pooling is embedded into the encoder to capture multi-scale contextual information for wheel rims of different sizes and viewpoints, while a channel-spatial attention mechanism in the decoder selectively enhances rim features and suppress background clutter. This design jointly improves the representation of global wheel contours and fine-grained rim structures, both essential for reliable angle estimation. Compared with traditional image processing pipelines and baseline deep networks, APCS-SwinUnet achieves higher accuracy and robustness in wheel rim segmentation.
(3)
The segmented mask is used to isolate the wheel rim point cloud, effectively filtering out irrelevant background data. The iterative closest point algorithm is then employed to register the initial and target wheel rim point clouds. After registration, the corresponding rotation and translation matrices are obtained, and the rotation matrix is subsequently used to compute the wheel’s toe and camber angles.
The rest of this paper is organized as follows. Section 2 reviews related work on wheel alignment. Section 3 presents the overview framework of this paper. Section 4 details the motivation for segmentation and the proposed network structure. Section 5 explains the principle of wheel angles calculated using the iterative closest point method. Section 6 describes the experimental setup and configuration. In Section 7, the segmentation and angle measurement results are presented. Section 8 summarizes the paper and discusses future development.

2. Related Work

In this section, we review related work from two complementary perspectives. First, we summarize existing wheel alignment techniques in terms of sensing and measurement strategies. Most conventional approaches can be broadly grouped into inertial-based methods and vision-based inspection methods, where the latter can be further divided into 2D and 3D vision-based measurement schemes. Second, since the proposed method formulates wheel alignment as a thin structure segmentation problem, we also review recent deep learning-based thin structure segmentation methods.

2.1. Wheel Alignment Sensing and Measurement Methods

2.1.1. Inertial-Based Wheel Alignment Methods

Inertial-based wheel alignment methods typically involve contact measurement. Specific sensors such as inertial measurement units (IMUs), accelerometers, gyroscopes, and angle sensors are mounted on the wheel or vehicle to collect acceleration and angular velocity data [4]. This data is preprocessed to reduce noise, correct sensors bias, and compensate for drift. Afterward, the data is transmitted to a processor (e.g., mobile phone, computer, or embedded device) via Bluetooth, WiFi, or RS232 for further analysis and computation of wheel alignment angles.
Numerous studies have explored the inertial-based methods. Young et al. [4] employed a microcontroller and triaxial accelerometer to measure the camber angle, introducing a calibration technique to correct axis misalignment between the sensor and the measurement system. Chatur et al. [5] proposed a wireless alignment method using accelerometers. D’Mello et al. [6] developed an internet of things (IoT) device consisting of a 6-axis microelectromechanical system (MEMS) sensor (MPU6050), an ESP32 microcontroller, and battery to measure toe and camber angles. Similarly, Bohari et al. [7] designed a Bluetooth-enabled alignment system using the MPU6050 sensor. Lee et al. [8] presented an online toe angle measurement approach based on lateral tire force, which was validated through simulations in CarSim and Simulink. To reduce energy consumption, Tang et al. [9] introduced a dual wake-up and self-calibration strategy for long-term monitoring. Inertia-based offer methods offer advantages such as high sampling frequency, low latency, low power consumption, and resistance to environmental conditions. However, they face several challenges. Physical contact is often required, risking damage to the wheel hub. Furthermore, sensor drift may cause cumulative errors, and measurement accuracy is affected by electromagnetic interference, temperature, and sensor location.

2.1.2. Vision-Based Wheel Alignment Methods

Compared with inertial-based methods, vision-based wheel alignment methods offer the advantages of convenience, highly automation, and efficiency. Typically, vision-based measurement techniques are divided into 2D and 3D vision measurement methods.
  • 2D Vision-Based Measurement Methods
A 2D vision inspection platform usually consists of target boards, infrared light emitting diodes (IR LEDs), image acquisition modules such as charged coupled device (CCD) or video cameras, and a computer [10]. Each target board, made of reflective material and featuring special patterns (e.g., chessboard or concentric circles), is mounted on the wheel rim using a four-legged clamp to represent the wheel attitude. During the measurement, IR LEDs illuminate the target boards, which are then captured by the CCDs. The technician turns the steering wheel in different directions and moves the vehicle back and forth to determine the axes of rotation and roll for each wheel. Finally, the wheel angles are calculated by analyzing the captured images based on projective geometry. Numerous 2D vision-based methods have been proposed. For instance, Xu et al. [11] proposed a toe and camber angles measurement method using a relevance vector machine and spatial clustering to classify the concentric circles on the target board. Xu et al. [12] proposed a global registration system combining multiple sensors with articulated arms and target boards, although this system is costly and complex. Ge et al. [13] introduced a multi-camera calibration technique using a transparent target board and image correlation. Jiang et al. [14] designed a calibration method for large cabin assembly with an improved homography matrix and combined large targets. Roshan et al. [15] developed a checkboard with augmented markers made of high-emissivity material and aluminum to expand its applicability. Xu et al. [16] proposed a kinpin inclination and caster measurement method using a 1D target with concentric circles.
Two-dimensional vision-based approaches are known for their high accuracy and efficiency, making them widely adopted in vehicle maintenance. However, they also have the following limitations. A complete 2D vision wheel alignment system is expensive, and installing clamps and the target boards is time-consuming and cumbersome, reducing measurement efficiency and increasing costs. Moreover, direct contact between the clamps and the wheels may cause potential.
  • 3D Vision-Based Measurement Methods
Three-dimensional vision-based measurement methods are generally categorized into passive and active 3D vision methods. Passive vision does not project any light source onto the wheel but relies on two or more cameras to recover depth maps or point clouds using triangulation between the camera centers and the wheel. In contrast, active 3D vision involves projecting a structured light source, such as point, line or surface, onto the wheel to reconstruct the object’s point cloud. Padegaonakar et al. [17] proposed a binocular passive vision method to measure wheel angles. Canny edge detection was applied to extract circular contours of the wheel, from which the outermost contour was selected as the region of interest (ROI). Wheel angles were then calculated using the depth map and trigonometry principles. While this method can reconstruct wheel point cloud, it has the following limitations. Passive stereo vision requires stereo rectification and stereo matching. Stereo rectification, based on intrinsic and extrinsic camera parameters, aligns corresponding points onto a common horizontal line. Stereo matching is then performed by exploiting surface textures and feature information. However, wheel hubs are highly reflective and often exhibit homogeneous textures, resulting in limited distinguishable features. This leads to low stereo matching accuracy, which in turn degrades the overall precision of 3D reconstruction. To address these issues, Xu et al. [2] proposed a method using a line laser and 1D target for large-scale onsite measurement. Furferi et al. [18] attached non-structured markers to the tire sidewall, combing image processing and stereo triangulation to associate 3D marker positions with the wheel plane. Senjalia et al. [19] developed a system using laser triangulation and camera calibration, where wheel regions were identified using OpenCV’s blob detection. However, this approach requires frequent threshold adjustment due to varying wheel shape, which reduces flexibility and efficiency. Furthermore, lighting conditions, reflections, and surface contamination can significantly affect accuracy. Kim et al. [20] employed two orthogonally mounted laser modules on the wheel knuckle to measure toe and camber angles. Baek et al. [21] developed a surface light-based depth-sensing system that acquires the wheel’s point cloud for geometric and posture analysis. Like 2D vision methods, this method requires a target board to be fixed on the wheel. Although no pattern is used, the board is used to represent the attitude of the wheel. To facilitate the ROI extraction, a thin (0.075 mm) adhesive film was applied. Wheel condition was assessed by comparing the current point cloud with a reference plane. While this method enables fast and stable point cloud acquisition, it still requires mounting a target board with a fixed relative position to the wheel. If the relative position changes, an error will be introduced. Moreover, additional matte adhesive film increases costs and reduce efficiency.

2.2. Deep Learning-Based Thin Structure Segmentation Methods

Current research on deep learning-based thin structure segmentation is still relatively limited, and most existing methods are built on convolutional neural networks (CNN) or Transformer-based architectures. Wang et al. [22] proposed a dual-path pavement crack segmentation network that combines a lightweight CNN encoder with a Transformer encoder incorporating attention and an efficient feedforward design to fuse local and global features for crack delineation. Gao et al. [23] proposed a multi-scale global attention network that integrates a densely connected attention U-Net backbone with a global context attention module for hierarchical context aggregation and top-down guidance, achieving consistently competitive retinal vessel segmentation and noticeably improved thin-vessel delineation. Geetha et al. [24] proposed an image-processing-assisted 1D frequency-domain CNN framework that extracts crack candidate regions via adaptive local binarization and then iteratively tracks missing single-pixel-wide propagating cracks. Siriborvornratanakul [25] evaluated Segment Anything Model (SAM) as a zero-shot solution for thin crack segmentation and reported that, although the SAM-Canny pipeline can delineate crack patterns reasonably well on multiple benchmark datasets, its numerical scores remain inferior to task-specific deep learning architectures, e.g., modified U-Net, when assessed under stringent 1-pixel crack annotation criteria. Li et al. [26] developed a dual-path progressive fusion network for retinal vessel segmentation, where a CNN branch is responsible for capturing fine local vessel details and a recurrent convolutional branch models richer contextual information. In addition, they introduced a progressive fusion mechanism composed of interactive fusion, cross-layer fusion, and scale feature fusion modules to integrate multi-scale features and mitigate the semantic gap between encoder and decoder.
Although these studies have advanced thin structure segmentation for targets such as vessels and pavement cracks, they do not fully address the wheel rim segmentation problem in our wheel alignment pipeline. The wheel rim in this work is not only a thin structure but also exhibits inner and outer boundaries with approximate symmetry, which differs significantly from tree-like or irregular thin patterns, and there is very limited work specifically targeting wheel-like thin structures. Moreover, rim segmentation here is only an intermediate step: our core objective is to estimate wheel angles from the segmented rim point cloud, so background and tire points can affect the robustness of subsequent point cloud matching, making accurate boundary segmentation particularly critical. In addition, practical wheel alignment requires a favorable trade-off between segmentation accuracy and efficiency, whereas many existing thin structure models are too computationally demanding for such scenarios.
In summary, inertial-based methods require physical sensor installation, which may damage the wheel and are sensitive to environmental factors such as electromagnetic interference and temperature. Although vision-based methods improve measurement efficiency, most of them still rely on target boards to estimate the wheel angle. To address these limitations from a technical perspective, this paper proposes a low-cost, non-contact, and practically feasible wheel alignment approach that integrates 3D vision and semantic segmentation techniques. Furthermore, to better handle the wheel rim, which is a thin, wheel-like structure with both inner and outer boundaries and approximate symmetry, we develop an APCS-SwinUNet architecture tailored for accurate rim segmentation.

3. Framework of Proposed Solution

The overall framework of the proposed method is shown in Figure 1. In the first stage, a wheel rim segmentation network based on APCS-SwinUnet is used to extract the wheel rim region from the RGB image. To achieve accurate and fine-grained segmentation, the network incorporates an ASPP module, an attention fusion module, and a hybrid Dice–Focal loss. The resulting segmentation mask is then mapped to the point cloud to obtain the wheel rim point cloud. In the second stage, the initial and target wheel rim point clouds are registered using the ICP algorithm, yielding the rotation and translation matrices that describe the wheel pose. The toe and camber angles are subsequently derived from the rotation matrix and converted to degrees, providing the final wheel angle measurements. Thus, the segmentation network and the point cloud registration are the two core components of the framework, bridging image-level processing and 3D geometric estimation.

4. Wheel Rim Segmentation Based on APCS-SwinUnet

This section presents the network architecture for wheel rim segmentation. We first introduce the motivation and objective of employing a segmentation network for wheel rim extraction. Next, we outline the overall structure of the proposed APCS-SwinUnet, followed by a detailed description of the key components, including atrous spatial pyramid pooling, convolutional block attention module, and hybrid loss function.

4.1. Motivation for Using Segmentation Network

As mentioned in the introduction, most existing wheel alignment methods require mounting a target board on the wheel and involve complex calibration procedures. Although some emerging approaches offer simplified setups, they still rely on target boards or matte adhesive films, which reduce measurement efficiency and increase system cost. To address these limitations, we propose a target-free alignment method by leveraging the point cloud of the wheel rim, eliminating the need for additional auxiliary materials. Traditionally, the target board is used to capture changes in the wheel’s attitude and angle due to the non-planar surface of the wheel hub. However, a closer examination reveals that the area between the wheel and the tire contains a circular smooth surface, which naturally spans a range of angles and can effectively reflect the wheel’s attitude changes, thereby making it a viable alternative for alignment purposes. The 3D scanner can capture the RGB image and point cloud of the entire scene, whereas our task focuses specifically on extracting the wheel rim’s point cloud.
Wheel rim extraction is critical for accurate angle estimation. As shown in Figure 2, the wheel rim is a narrow annular structure that occupies only a small portion of the wheel, making its extraction particularly challenging for the following reasons: (1) The wheel rim has both inner and outer boundaries, and its thin, curved geometry complicates accurate localization. (2) Although there is a noticeable reflectance difference between the wheel rim and the tire, the wheel rim and hub often share similar reflectivity, leading to weak grayscale contrast and ambiguous edges. (3) Since the point cloud of the wheel rim is used to estimate the overall wheel posture, background and tire regions need to be excluded, requiring precise segmentation of the rim contours. Traditional image processing methods can, in principle, perform this task, but they typically involve multiple steps and require manual tuning for different wheel types, resulting in limited flexibility and poor generalization. Deep learning-based segmentation offers a more robust and adaptive alternative.
Prior studies on thin or narrow structures (such as vessels, roads, or cracks) often leverage multi-scale context and attention mechanisms to improve boundary delineation, and are predominantly implemented within U-Net-like or transformer-based encoder–decoder architectures. However, these methods are generally designed for tree-like or line-like topologies, rather than for closed, narrow annular rims with dual inner-outer boundaries and weak contrast between the rim and the hub, and they are not specifically optimized for downstream wheel alignment tasks. As a result, existing architectures still struggle to accurately delineate the thin boundary regions of the wheel rim and to provide sufficiently stable contours for subsequent 3D point cloud registration and angle estimation. To address these issues, we design APCS-SwinUnet as a task-specific adaptation of established segmentation modules, tailored to the geometric and photometric characteristics of the wheel rim and its role in wheel alignment measurement.

4.2. Network Architecture of APCS-SwinUnet

The architecture of APCS-SwinUnet is illustrated in Figure 3. It is built upon the Swin Transformer and U-Net frameworks and achieves high segmentation accuracy while maintaining efficiency. First, an atrous spatial pyramid pooling (ASPP) module is added after the third encoder layer to capture multi-scale features relevant to wheel rims. Second, a convolutional block attention module (CBAM) is integrated into the third decoder layer to strengthen the network’s focus on the rim contour and suppress background clutter. Finally, a hybrid loss function combining Dice and Focal losses is employed to better handle class imbalance and stabilize training. In this way, APCS-SwinUnet is not a completely novel architecture, but a carefully engineered combination and adaptation of known modules tailored to the challenges of wheel rim segmentation.
These design choices are particularly effective for wheel rim segmentation for several reasons. The ASPP module provides annulus-aware multi-scale context, enabling the network to capture both the thin rim band and the full circular structure, which helps maintain a closed, consistent rim contour while suppressing spurious responses from the tire and background. CBAM is especially beneficial under weak rim-hub contrast: by refining channel and spatial attention at decoder stages where rim and hub features are mixed, it enhances subtle structural cues along the rim boundary and reduces confusion between these two regions. The Swin-Transformer backbone contributes by modeling long-range dependencies along the rim circumference and across the whole wheel, which promotes global shape consistency while preserving local boundary details that are crucial for producing stable contours for subsequent 3D point cloud registration and angle estimation.

4.2.1. Atrous Spatial Pyramid Pooling

The segmentation task in this study aims to accurately localize the global contour of the wheel rim while capturing fine-grained boundary details. To enhance segmentation performance, we incorporate the ASPP module, as illustrated in Figure 4. ASPP strengthens the network’s ability to capture multi-scale contextual information by combing atrous convolution with image-level features at varying sampling rates. The ASPP structure [27] consists of multiple convolution layers, atrous convolution layers with different dilation rates, pooling layers, ReLU activations, and an upsampling layer. Its core component is a set of parallel atrous convolutions, which enable the network to extract features across multiple receptive fields without increasing the size or computational cost. Specifically, we employ three parallel atrous convolutions with dilation rates of 4, 6 and 8. Since the proposed segmentation network is based on a Transformer architecture, the input sequence needs to be converted into a 2D convolutional format before entering the ASPP. After processing, the ASPP output is then reshaped back into sequence form for subsequent Transformer-based operations.

4.2.2. Attention Fusion Module

To enhance the network’s ability to recognize wheel contour, we add a convolutional block attention module after the last Swin-transformer block of the decoder. The structure of the CBAM is shown in Figure 5. The CBAM [28] consists of a channel attention module and a spatial attention module. The channel attention focuses on identifying what features are important, and comprises an average pooling layer, a max pooling layer, a multilayer perceptron (MLP), and a Sigmoid activation. This process is formulated in Equation (1). In contrast, the spatial attention module emphasizes where the important features are located. It includes average and max pooling operations followed by a convolution layer, as described in Equation (2). The final feature map is obtained by sequentially applying both attention mechanisms, as shown in Equation (3).
F c = σ W 1 W 0 F a v e g _ p o o l + W 1 W 0 F max _ p o o l
F s = σ f 7 × 7 F a v e g _ p o o l , F max _ p o o l , F s 1 × H × W
F = F c F , F = F s F
where F a v g _ p o o l and F max _ p o o l denote the average pooling and maximum pooling, respectively. W 0 and W 1 denote the weights of the MLP. f 7 × 7 denotes the convolutional operation with a convolutional kernel size 7 × 7 . σ denotes the sigmoid function. represents the element-wise multiplication operation. F and F denote the temporary output and the refined output, respectively. F , F c and F s denote the original feature map, output of channel attention module and output of spatial attention module, respectively.

4.2.3. Hybrid Loss Function

To enhance the network’s ability to capture fine details of wheel boundaries and improve training stability, we employed a hybrid loss function that combines Dice loss and Focal loss, as defined in Equations (4)–(7).
L l o s s = 1 τ L d i c e + τ L f o c a l
L d i c e = 1 N D t = 1 N D 1 2 × k p t k g t k k p t k + k g t k
L f o c a l = 1 N D t = 1 N D k = 1 N P α t 1 c t k β t log c t k
c t k =         p t k ,         i f   g t k = 1 1 p t k ,   i f   g t k = 0
where L , L d i c e and L f o c a l denote the total Dice, Dice loss and Focal loss, respectively. τ in Equation (4) is used to balance the contributions of L d i c e and L f o c a l . p t k and g t k denote the predicted and ground-truth probabilities for pixel k in category t , respectively. N D and N P denote the number of classes and the total number of pixels, respectively. α t and β t are introduced to adjust the weighting for each class t , helping to mitigate the class imbalance between foreground and background.

5. Toe and Camber Angles Calculation Based on Iterative Closest Point

This section is about how to combine the segmented mask and the acquired point cloud to calculate the wheel angles. It is organized in three main parts: point cloud extraction of wheel rim, point cloud registration of wheel rim based on iterative closest point and wheel angle calculation.

5.1. Point Cloud Extraction of Wheel Rim

Based on the mask generated by the APCS-SwinUnet, the wheel rim needs to be extracted from the 3D point cloud. As illustrated in Figure 6, the 3D scanner stores point cloud data in a row-wise sequence. Let the point cloud sequence be denoted by P i   ( X i ,   Y i ,   Z i ) , where i denotes the index of the point cloud, and X , Y and Z have the same index. The point cloud corresponding to the point ( x w , y h ) on the mask can be calculated using the following equation.
P i = P i ,   i f   m a s k ( x w , y h ) > 0 P i = 0 ,   i f   m a s k ( x w , y h ) 0 , i = ( y h 1 ) × W + x w
where P and P denote the original point cloud and the extracted point cloud, respectively. ( x w , y h ) denotes a point on the segmented mask. W and H refer to the width and height of the image, respectively.

5.2. Point Cloud Registration of Wheel Rim Based on Iterative Closest Point

This section is about how to combine the segmented mask and the acquired point cloud to calculate the wheel angles. It is organized in three main parts: point cloud extraction of wheel rim, point cloud registration of wheel rim based on iterative closest point and wheel angle calculation. Suppose that the initial and target point cloud sequences are P p 1 , p 2 , p 3 , p i ,   i N and Q q 1 , q 2 , q 3 , q j ,   j N , respectively, where N denotes the total number of points. Since P and Q are acquired from different viewpoints by the same 3D scanner, each point j in the set P can find a corresponding point in Q by the rigid transformation, including rotation matrix R and translation matrix t , such that
j ,   q j = R p j + t
Then, we use the iterative closest point (ICP) algorithm [29] to find R and t that minimize the alignment error defined by the following objective function.
F = j = 1 N q j ( R p j + t ) 2
To ensure robustness, a least-squares optimization is performed based on all point correspondences and the singular value decomposition (SVD) algorithm is used to solve for R and t . This approach guarantees a stable and optimal solution under the rigid transformation assumption, while ensuring that the resulting R and t satisfy orthogonality constraints. Such a formulation not only minimizes the geometric alignment error but also ensures the reliability of rotation matrix R and translation matrix t for subsequent angular measurements or geometric analyses.

5.3. Toe and Camber Angles Calculation

After applying the ICP algorithm, the rotation matrix R is obtained, which describes the rotational relationship between the initial and target point clouds with respect to each axis. The rotation matrix can be expressed as follows:
R = r 11 r 12 r 13 r 21 r 21 r 23 r 31 r 32 r 33 = cos ψ cos θ cos ψ sin θ sin ϕ sin ψ cos ϕ cos ψ sin θ cos ϕ + sin ψ sin ϕ sin ψ cos θ sin ψ sin θ sin ϕ + cos ψ cos ϕ sin ψ sin θ cos ϕ cos ψ sin ϕ sin θ cos θ sin ϕ cos θ cos ϕ
where r 11 ~ r 33 denote the elements of the rotation matrix R . ψ , θ and ϕ represent the Euler angles around the z , y and x axes, respectively, which are commonly referred to as the yaw, pitch, and roll angles. Since the rotation matrix R is obtained through the ICP algorithm, the Euler angle can be calculated using Equations (12)–(14).
ψ = tan 1 r 21 r 11
θ = sin 1 r 31
ϕ = tan 1 r 32 r 33
Then, the Euler angle can be converted from radians to degrees using Equation (15).
ψ = ψ × 180 π ,   θ = θ × 180 π ,   ϕ = ϕ × 180 π
where ψ , θ and ϕ denote the Euler angles in degrees corresponding to ψ , θ and ϕ respectively. In this paper, ϕ and θ represent the toe angle and camber angle, respectively.

6. Experimental Setup and Configuration

This section presents the detailed parameters of the measurement system, including the overall system architecture, detailed parameters of the 3-D scanner, clinometer and server.

6.1. Measurement System

As shown in Figure 7, the measurement system consists of an Azure Kinect, a clinometer, a tripod, and a computer. The Azure Kinect was positioned approximately 70 cm from the measured wheel and 29.3 cm above the ground. The intrinsic parameters (focal length, principal point, and lens distortion) and the extrinsic transformation between the RGB and depth cameras are taken from the factory calibration provided by the Azure Kinect SDK. Since the wheel-angle estimation is performed directly in the depth coordinate frame, the relevant extrinsic relationship is the pose of the Azure Kinect with respect to the wheel, defined by the measured distance, height, and the rotation angle given by the clinometer.
Since a commercial wheel alignment system is very expensive, we do not yet have sufficient budget to purchase such equipment. As a result, we adopt a feasibility-oriented experimental setup in which the Azure Kinect sensor is mounted on a tripod with a lockable pan-tilt head, and its pose is adjusted manually via the handle. In this setup, the camber and toe angles are simulated by rotating the Azure Kinect vertically and horizontally, respectively, while keeping the wheel itself stationary. Because the sensor is adjusted manually, it is only approximately, but not perfectly, parallel to the wheel plane. The corresponding camber and toe angles are measured using a clinometer, and these readings are treated as ground-truth values in our experiments. In the implemented system, the initial scan is taken as the geometric reference configuration of the wheel rim, and all subsequent measurements are obtained by registering the new rim point clouds to this reference using ICP. Consequently, the proposed method relies on relative pose estimation between the reference and current rim configurations, rather than on any absolute pose of the sensor. This design makes the approach inherently robust to non-parallelism between the camera plane and the wheel surface, as long as the relative motion between the sensor and the wheel is accurately captured.
The main uncertainty sources in this feasibility setup include depth-measurement noise, the limited resolution and alignment accuracy of the clinometer, manual sensor positioning, and segmentation errors in the wheel-rim boundary that propagate to the estimated angles. A complete uncertainty budget and a more detailed repeatability analysis based on higher-precision fixtures and encoders are left for future work.

6.2. Three-Dimensional Scanner Description

The 3D scanner used in this study is the Azure Kinect, as shown in Figure 8. The Kinect is a low-cost 3D scanner that integrates three main imaging components: an RGB camera, a depth camera, and infrared (IR) emitters. It recovers the depth map and 3D point cloud based on the time-of-flight (ToF) principle. Specifically, the device emits IR light pulses onto the object, and the depth is calculated based on the time taken for the reflected light to return to the sensor. The key parameters used in this study are as follows: the RGB camera acquires images at a resolution of 1920 × 1080 pixels, with a frame rate of 30 fps and a 16:9 aspect ratio. The depth camera operates in Wide Field of View (WFOV) mode. The IR settings include 9 pulses, 125 µs pulse width, 8 idle periods, 1450 µs idle time, and 12.8 ms exposure time. Data acquisition was implemented using the Azure Kinect SDK in Visual Studio 2019. A custom program was developed to capture both RGB images and the corresponding 3D point clouds of the measured wheel.

6.3. Description of Clinometer

The clinometer is used to measure the rotation angles of the Azure Kinect. As shown in Figure 7, it is fixed to the Kinect and rotates together with it either horizontally or vertically. The clinometer used in this study has a measurement range from 0° to 360° degrees, with a resolution of 0.05°.

6.4. Server Configuration

The APCS-SwinUnet network was trained on a server equipped with an NVIDIA GeForce RTX 3090Ti graphics card (24 GB; NVIDIA Corporation, Santa Clara, CA, USA), an Intel Core i9 CPU (4.3 GHz; Intel Corporation, Santa Clara, CA, USA), and 64 GB of RAM (Samsung Electronics Co., Ltd., Suwon, Republic of Korea). The system runs on Ubuntu 20.04. The software environment includes Python 3.8, Pytorch 1.13, and CUDA 11.6.

7. Experimental Results and Analysis

This section aims to verify and evaluate the feasibility of the proposed solution, focusing on the segmentation results and the calculated wheel angles. The experimental details and corresponding results are presented below.

7.1. Wheel Segmentation Experiments

We first provide details of the network training process, including the dataset composition and parameter settings of the proposed APCS-SwinUnet. Subsequently, the segmentation performance is evaluated under normal conditions through both quantitative analysis based on defined evaluation metrics and quantitative assessment of the segmentation results. Next, we evaluate the segmentation performance of the proposed method on our wheel dataset under varying illumination conditions. Then, feature maps from both the encoder and decoder are visualized. Finally, the performance of different segmentation methods is compared on a public wheel dataset.

7.1.1. Datasets for Training and Testing

In this study, 535 different wheel types were collected. These wheel types come from different vehicles, most of which are passenger vehicles, and were captured primarily under sunny and overcast outdoor conditions. Among these 535 wheel types, 423 types were assigned to the training set and the remaining 112 types to the test set. This split is performed at the wheel-type level, so that images (original or augmented) belonging to the same wheel type never appear in both the training and test sets. To enhance the diversity and robustness of the training data, we applied a series of data augmentation techniques to the initial training set, including brightness and contrast adjustment, noise injection, rotation, and other geometric and photometric transformations. As a result, the training set contains 5499 images in total. The test set comprises 112 original, unaltered images and is not augmented, in order to preserve the integrity and validity of the evaluation.

7.1.2. Parameter Settings of APCS-SwinUnet

The network was trained for 30 epochs with a batch size of 6. Input images were resized to 512 × 512 pixels after being fed into the network. The Adam optimizer was used, with an initial learning rate 0.2 and a drop path rate of 0.001, respectively. The window size and embedding dimension were set to 32 and 96, respectively. The numbers of multi-head attention modules in each stage were 3, 6, 12, and 24. The dilation rates in the ASPP module were set to 4, 6 and 8, respectively.

7.1.3. Evaluation Criteria for Segmentation Network

We first use the Dice similarity coefficient (Dice) to evaluate the overlap and similarity between the predicted mask (P) and the ground truth mask (G), as illustrated in the schematic diagram in Figure 9a. The Dice coefficient can be calculated by Equation (16).
D i c e = 2 P G P G
where P G and P G denote the number of intersecting pixels and the total number of pixels of the predicted and ground truth regions, respectively. In addition, we introduce the Hausdorff distance (HD) metric, whose schematic diagram is shown in Figure 9b. While Dice coefficient is more sensitive to the interior of the mask, HD focuses on boundary differences. The HD is defined by the following equation.
D G , P = max d ( G , P ) , d ( G , P ) d ( G , P ) = max g G min p P g p d ( P , G ) = max p P min g G p g
where represents a distance paradigm operation (e.g., Euclidean distance). The Hausdorff Distance is a bidirectional metric composed of two one-way distances, d ( G , P ) and d ( P , G ) . Specifically, D G , P computes the distance g p between each point g i in the set of G to the point p i that is closest to this point g i . Then, the maximum value of these distances is the value of this distance as the value of d ( G , P ) . Similarly, d ( P , G ) is computed by the same way. Finally, the bidirectional HD is the largest of the two unidirectional distances d ( G , P ) and d ( P , G ) , which denotes the maximum degree of mismatch between G and P . In this study, we adopt HD95, which represents the 95% of all point-to-set distances between H and G . HD95 can reduce the effect of outliers and enhance the stability of the evaluation.

7.1.4. Comparisons with Different Segmentation Networks

To evaluate the performance of the proposed APCS-SwinUnet network, we conducted comparisons with several representative segmentation networks, including FCN [30], DeepLabV3 [27], OcUnet [31], U-Net [32], Att-Unet [33], U-Net 2+ [34], U-Net 3+ [35], TransUnet [36], and SwinUnet [37] networks. Both quantitative and qualitative segmentation results are provided for comprehensive analysis.
(1)
Quantitative Analysis:
The Dice scores achieved by different methods are summarized in Table 1. As shown in Table 1, there are significant differences among the methods. The proposed method achieves a relatively high Dice score and a comparatively low HD95 value. A higher Dice score indicates better overall contour of the wheel, while a lower HD95 indicates improved accuracy in delineating fine boundary details. Empirically, APCS-SwinUnet outperforms strong baselines, including SwinUnet and TransUnet, in both Dice and HD95. In particular, SwinUnet achieves a Dice of 90.23% and an HD95 of 2.47, whereas APCS-SwinUnet achieves a Dice of 90.66% and an HD95 of 2.11. The 0.43% improvement in Dice reflects better overall rim contour segmentation, while the reduction in HD95 directly indicates more accurate localization of fine boundary regions, which is precisely the aspect targeted in thin-structure segmentation. In practice, these improvements translate into measurable gains in toe and camber angle estimation, confirming that the adapted architecture is effective for the real-world wheel alignment scenario.
These improvements mainly stem from three factors: the strong feature extraction of Swin Transformer blocks, the ASPP module’s ability to capture multi-scale rim boundaries, and the CBAM attention mechanism, which focuses on rim regions while suppressing background clutter.
(2)
Qualitative Analysis:
To more intuitively compare the strengths and weakness of each method, we conducted a qualitative analysis on typical segmentation results. Specifically, five types of wheels (T1~T5), shown in the first row of Figure 10, were segmented using different network in Table 1. For wheel type T1, all listed methods achieve satisfactory segmented results. However, when the wheel rotates, as in the T2 column, the segmentation becomes more challenging. As shown in the T2 column of Figure 10, U-Net and Att-Unet networks not only miss part of the wheel region but also mistakenly classify background regions as wheel areas. In contrast, the other methods demonstrate better performance on T2 type of wheel. For the T3 type, methods including FCN, DeepLabV3, U-Net and U-Net2+ tend to misclassify the wheel fender as part of the wheel region, likely due to the fender’s circular shape resembling the upper portion of the wheel. In type T4, the wheel’s lower edge is very close to the ground, making it difficult to distinguish from the background. As a result, methods such as U-Net, Att-UNet, U-Net2+, U-Net3+, and SwinUnet misclassify parts of the ground as the wheel. Type T5 presents a different challenge: the wheel edges are particularly narrow, which leads to partial omission of the wheel region by U-Net, Att-UNet, U-Net2+, U-Net3+, TransUnet, and SwinUnet. In contrast, the proposed method consistently delivers accurate segmentation across all wheel types (T1~T5), producing results that are visually closer to the ground truth (GT), as seen in Figure 10. It is worth noting that although some methods may not show obvious omissions or misclassifications, their segmentation masks may still be slightly larger or smaller than the ground truth, particularly around the boundaries. This discrepancy is reflected in the quantitative results presented in Table 1. In summary, from both quantitative and qualitative perspectives, the proposed method demonstrates superior and more stable segmentation performance compared to other mainstream networks. Although the Dice score of our method is 0.43% higher than that of SwinUnet, it achieves better boundary precision and segmentation stability, which is beneficial for subsequent point cloud registration and angle computation.

7.1.5. Comparison Results Under Inconsistent Illuminations

To evaluate the impact of illumination changes, we collected 106 wheels under diverse lighting conditions, including cloudy weather, direct sunlight, and shadow occlusion, with examples shown in Figure 11. The previously trained models were applied to this dataset, and the segmentation results are summarized in Table 2. A comparison between Table 1 and Table 2 reveals that most methods suffer from performance degradation to varying degrees under inconsistent illumination, particularly in accurately delineating the edges of the wheel rims, as reflected by the HD95 metrics. In contrast, the proposed method exhibits a smaller performance drop, indicating greater stability and robustness under different lighting conditions.

7.1.6. Feature Visualization of Encoder and Decoder

To provide a clearer and more intuitive understanding of the proposed network’s internal representations, we use the wheel image in Figure 11b as an example and visualize the feature maps of each stage of the encoder and decoder. The feature maps of the four transformer modules in the encoder are shown in Figure 12a–d, while those in the decode are shown in Figure 12e–h, respectively. For consistency, all feature maps in Figure 12a–h are displayed at the same size. Figure 12a–d illustrate the progressive abstraction from low-level to high-level semantic features, whereas Figure 12e–h demonstrate the gradual restoration of spatial resolution via upsampling. As observed in Figure 12, the APCS-SwinUnet network consistently focuses on regions near the wheel rim in both encoder and decoder stages, while background and vehicle body regions are significantly suppressed. This facilitates accurate segmentation of the wheel rim region, particularly its contour.

7.1.7. Comparison Results on the Public Wheel Dataset

To evaluate the performance of the proposed method on public wheel datasets, we selected a publicly available dataset named CAWDEC [38]. Representative sample images are shown in Figure 13. As this dataset does not provide ground-truth masks for wheel hub contours, we manually annotated 101 images to support evaluation. Given the time-consuming labeling process, we annotated only 101 images from the CAWDEC dataset. These annotated images were then used to compare the proposed method with several mainstream segmentation approaches. Thus, this experiment mainly serves as a complementary 2D rim-segmentation benchmark and is inherently constrained by the limited number of manually annotated samples and the practical cost of dense wheel-rim labeling. To better assess the generalization ability of the proposed method, we did not retrain the model on the CAWDEC dataset. Instead, the model trained on our own dataset was directly applied to perform segmentation. The segmentation results of different methods are summarized in Table 3.
As observed, most methods exhibit a notable drop in accuracy, likely due to inconsistencies in acquisition devices, shooting angles, and environmental conditions. Despite these challenges, the proposed method achieves comparatively better segmentation performance. It is worth noting that segmentation serves only as an intermediate step in the proposed framework, with the resulting segmentation masks used to extract the point cloud of the wheel rim. Therefore, all subsequent evaluations are conducted on images captured by the developed platform. As shown in Table 1, the proposed method also achieves strong segmentation performance on the developed platform. Accurate wheel-rim annotation requires fine-grained, pixel-level masks and is therefore extremely time-consuming. Nevertheless, since this research is still in the early stages, we plan to further expand the dataset in both size and diversity in future work.

7.2. Point Cloud Extraction Result of Wheel Rim

After obtaining the wheel rim mask using APCS-SwinUnet, the corresponding point cloud can be extracted from the full point cloud. To illustrate this process, we take the wheel in Figure 14a as an example. Its point cloud, acquired by the Azure Kinect, is shown in Figure 14b. The wheel image is then fed into the APCS-SwinUnet for inference, producing a mask containing only the wheel rim, as shown in Figure 14c. Finally, the point cloud of the wheel rim is extracted using the segmented mask and the method described in Section 5.1, which is shown in Figure 14d.

7.3. Wheel Angle Measurement and Evaluation

7.3.1. Point Cloud Registration Result

The initial point cloud and registered point cloud are shown in Figure 15a and Figure 15d, respectively. Figure 15b,c displays the initial point clouds from two additional positions, with their corresponding registration results presented in Figure 15e,f. This study uses the reference and target point clouds shown in Figure 15a as an example to illustrate the process of point cloud registration and wheel angle calculation. The registration between the reference and target point clouds, as shown in Figure 15b, is performed using the ICP algorithm described in Section 5.2. The corresponding rotation matrix is provided in Equation (18).
R = r 11 r 12 r 13 r 21 r 21 r 23 r 31 r 32 r 33 =   0 . 9988 0 . 0001 0 . 0495 0 . 0002 1 . 0000 0 . 0034 0 . 0495 0 . 0034 0 . 9988
The rotation matrix R is an orthogonal matrix used to align the reference point cloud with the target point cloud. The rotation matrix R contains nine elements, denoted as r 11 ~ r 33 . The specific definition of r 11 ~ r 33 is provided in Equation (11). X ^ 0.9988 , 0.0002 , 0.0495 T , Y ^ 0.0001 ,   1.0000 , - 0.0034 T and Z ^ 0.00495 ,   0 . 0034 ,   0 . 9988 T in matrix R represent the projections of X , Y and Z axes of the target coordinate system onto the reference coordinate system, respectively. Next, the rotation matrix can be decomposed into radian angles ( ψ ,   θ ,   ϕ ) representing rotation rotated around Z , Y , X axes. ( ψ ,   θ ,   ϕ ) in radian values are then converted into their corresponding degrees angles ( ψ ,   θ ,   ϕ ) . Both the radian and degree values are summarized in Table 4. Since the Kinect is currently moving along the horizontal direction, this simulated a camber angle, with the resulting angle θ y representing the target camber angle.
To evaluate the reliability and robustness of point cloud registration process, Gaussian noise with a standard deviation σ ranging from 0.01 to 0.10 was added to both the original reference point cloud and target point cloud shown in Figure 15a. The ICP algorithm was then applied to perform point cloud registration and compute the corresponding rotation matrices, which were subsequently used to calculate the wheel angles. To more intuitively assess the impact of segmentation on angle estimation accuracy, a relative error metric is introduced, as defined in Equation (19).
R e = y i y ^ i y i
where y i and y ^ i denote the real angle and the measured angle of i , respectively.
The measured angles, absolute errors and relative errors under different noise levels are summarized in Table 5. As shown in Table 5, the estimated wheel angles remain relatively stable despite increasing noise levels, with both absolute and relative errors being small. These results demonstrate the obtained rotation matrices and the point cloud registration process are stable and robust.

7.3.2. Measurement Results of Toe and Camber Angles

In this section, we give the results of angle measurements at different positions and directions. In addition, in order to evaluate the effect of the measurement distance on angle measurement, the measurement platform was placed at different distances from the wheel. The measured results are shown in Table 6 and Table 7, respectively. D s and D e in Table 6 and Table 7 denote the distance of the innermost and outermost sides of the wheel from the platform, respectively. We can see from Table 6 and Table 7 that the maximum measurement errors of the toe and camber angles are and −0.105° and 0.158°, respectively. Most of the measured angles are very close to the real angle, which proves that the proposed method has a strong stability with the change in the measurement distance.

7.4. Extended Experiments

7.4.1. Angle Measurement Results of the Raw Point Couds

To verify the necessity of the proposed segmentation-guided registration scheme, we first attempted direct registration using the raw wheel point clouds, each containing up 2,073,000 points. However, due to the high point density, point cloud registration was extremely time-consuming and often exhibited poor convergence. To address this issue, we applied downsampling to the raw wheel point clouds at various rates (ranging from 10% to 80%) and performed point cloud registration on the sampled data. The experimental results shown in Table 8 indicate that increasing the sampling rate leads to longer registration times, while the angle estimation accuracy remains limited and does not improve with more point data. This is primarily due to the presence of unstable points from the car body, tire, and background in the raw data, which negatively affect the registration process and ultimately introduce angle estimation errors. In addition to poor accuracy, when the sampling rate reaches 80%, the point cloud registration takes nearly seven hours, rendering the registration process time-consuming and less viable. In contrast, the proposed solution first segments the wheel rim using a dedicated network before point cloud registration, enabling faster and more accurate angle estimation. This proposed solution not only improves the measurement efficiency but also enhances the angle estimation accuracy. The experimental results in Table 8 validate the above conclusions.

7.4.2. Impact of Pixel Shifts on Angle Measurement

To evaluate the impact of wheel rim segmentation on wheel angle estimation, this section uses the wheel mask shown in Figure 16a as an example. The masks are then horizontally shifted by several pixels to the left and right, as shown in Figure 16b,c, respectively. These shifted masks are subsequently used for point cloud registration, and the corresponding wheel angles are calculated. The experimental results are presented in Table 9, where S L , M L , S R , M R , A e , and R e denote the leftward pixel shift, leftward angular measurement, rightward pixel shift, rightward angular measurement, absolute angular error, and relative angular error, respectively. As observed from Table 9, the angle estimation error increases with the magnitude of the pixel shift, regardless of whether the masks are shifted to the left or right. This growing error is primarily attributed to the inclusion of non-wheel-rim point cloud regions in the registration process as the shift increases. These extraneous regions lie outside the actual wheel rim and do not form a consistent planar surface. Their varying distances from the measurement platform introduce inaccuracies in point cloud registration, ultimately leading to deviations from the true wheel angle. These findings highlight the importance of accurate wheel rim segmentation for precise wheel angle estimation.

7.4.3. Impact of Segmentation Methods on Angle Measurement

To evaluate the impact of different segmentation methods on wheel angle measurement, a wheel with a ground-truth angle of 3.200° is selected as an example for analysis. Segmentation masks generated by various methods are used to estimate the wheel angle, and the results are summarized in Table 10. As shown in Table 10, the accuracy of angle estimation varies significantly across different segmentation methods, primarily due to significant differences in the Dice coefficients and HD95 values. Compared with the other methods, the proposed approach achieves more accurate segmentation of the wheel rim, particularly around the boundaries, which results in higher precision in the final angle estimation. The experimental results presented in this section demonstrate a strong correlation between the accuracy of wheel rim segmentation and the accuracy of angle estimation. Moreover, compared with other methods, the proposed approach yields smaller angle estimation errors, further validate its effectiveness.

7.4.4. Runtime and Computational Efficiency

To quantify the computational efficiency of the proposed solution, we evaluated the end-to-end runtime of the complete pipeline, which consists of three main stages: wheel rim edge segmentation, point cloud registration, and angle estimation. On our test platform, the segmentation network requires approximately 0.463 s per image to generate the wheel rim mask. The subsequent point cloud registration and angle estimation take about 4.476 s per sample, resulting in a total processing time of roughly 4.939 s per wheel. For comparison, when point cloud matching is performed directly on the raw point cloud without prior segmentation, the computational burden increases drastically. At a sampling rate of 80%, the point cloud registration alone requires nearly 7 h to complete, and the resulting accuracy is significantly lower. These results demonstrate that the proposed method offers a favorable trade-off between computational efficiency and practical applicability in wheel-alignment scenarios.

7.4.5. Repeatability Experiments at Different Distances and Illuminations

To evaluate the stability and robustness of the proposed method under varying distances and illuminations, a series of repeated experiments were conducted. Four alignment cases were considered in this section: toe-in, toe-out, negative camber, and positive camber. The experiments were carried out on a sunny and breezy day, with the ambient temperature around 31 degrees Celsius. To account for illumination variations, wheel measurements were taken at 11:39 a.m., 12:45 p.m., 18:01 p.m., and 19:08 p.m. The corresponding wheel images captured at these different times are shown in Figure 17.
Taking the toe-in case as an example, five RGB images and corresponding point clouds were acquired at the initial wheel position using the Azure Kinect, with a two-minute interval between each capture. After repositioning the sensor, another five sets of RGB images and point clouds were collected at the target position using the same procedure. Including the time required for data storage, each angular repeatability experiments took approximately twenty minutes. Given the extended acquisition period, several random factors were considered, including sensor noise, illumination variation, wind disturbances, and the inherent measurement uncertainty of the Azure Kinect. The repeatability results for toe-in, toe-out, negative camber, and positive camber angles are summarized in Table 11. For each alignment case, five data groups were collected at both the initial and target positions. These were paired to compute five corresponding angular values per case. As shown in Table 11, the angular measurement errors under varying distances and lighting conditions remain within a narrow range. To quantitatively assess the accuracy, mean absolute error (MAE) and root mean square error (RMSE) were calculated according to Equation (20).
M A E = 1 n i = 1 n y i y ^ i R M S E = 1 n i = 1 n y i y ^ i 2
where M A E and R M S E denote the mean absolute error and root mean square error. y i and y ^ i denote the real value and the measured value of i , while N denotes the total number of samples. The MAE and RMSE for the toe angle are −0.045° and 0.078°, respectively, while those for the camber angle are 0.010° and 0.066°, respectively. These repeated experiments show that the proposed method exhibits good stability and robustness against environmental variations. It is worth noting that direct sunlight should be avoided during measurement. A distance ranging from 670 mm to 850 mm between the wheel and the measurement platform is considered acceptable. Furthermore, since wheel alignment tasks are typically performed indoors in vehicle maintenance facilities, the measurement environment is generally well-controlled and stable.

8. Conclusions and Future Development

To enable low-cost and flexible wheel alignment, this paper proposes a novel strategy that integrates semantic segmentation and 3D vision. We develop an APCS-SwinUnet segmentation network that incorporates ASPP, CBAM, and a hybrid loss function to achieve accurate wheel rim extraction. Using the segmented rim as the region of interest, we apply the ICP algorithm to register the reference and target point clouds, and then convert the resulting rotation matrix into toe and camber angles. Experimental results show that high-quality segmentation is crucial for reliable point-cloud registration and precise angle estimation. Overall, the proposed solution provides accurate, stable, and robust wheel-angle measurements while remaining cost-effective and flexible compared with conventional wheel alignment methods.
Quantitatively, APCS-SwinUnet achieves a Dice score of 90.66% and an HD95 of 2.11 for rim segmentation, representing a 0.43% increase in Dice and a 0.36 reduction in HD95 compared with the baseline model. Repeated experiments further show that the MAE and RMSE are −0.045° and 0.078° for the toe angle, and 0.010° and 0.066° for the camber angle, respectively, confirming that the proposed system can deliver accurate and reliable wheel-angle measurements. In the revised manuscript, we also relate these errors to the accuracy requirements specified in standards for four-wheel alignment instruments. Specifically, the allowable indication errors for total toe and single-wheel toe are within ±4′, and that for camber is within ±2′, corresponding to approximately 0.03–0.07°. Although some of our current measurements yield absolute toe and camber errors that are slightly larger, the errors remain well below 0.1° and are close to the target accuracy range.
The primary purpose of this study is to validate the feasibility of combining semantic segmentation with point-cloud registration for wheel-angle measurement. Accordingly, the current experiments focus on a single wheel in a controlled environment, and the proposed system should be regarded as a feasibility prototype rather than an industrial-grade product. For applications requiring higher angle-measurement accuracy, the proposed algorithmic framework can, in principle, be integrated with industrial-grade 3D sensors. We also explicitly acknowledge that the RGB images and associated 3D point-cloud data in our self-collected dataset have limited diversity in wheel types and acquisition conditions. This limitation constrains the current assessment of generalization and motivates future extensions to larger and more diverse datasets. In particular, we have outlined a future plan to enlarge the 3D dataset with additional wheel designs, environmental conditions, and viewpoints so as to more comprehensively assess and improve the generalization ability of the proposed model. Due to current resource constraints, a direct quantitative comparison with a commercial wheel-alignment system is not yet feasible. However, we recognize that such a benchmark would be technically sound and highly valuable, and we plan to allocate future resources to acquire a commercial system for more rigorous, industry-relevant validation. Future work will first extend the system to simultaneously measure the angles of all four wheels by integrating four such subsystems. We also plan to further narrow the gap to industrial standards by (i) adopting industrial-grade 3D scanners with higher stability and smaller systematic errors, (ii) performing rigorous comparisons against certified commercial alignment systems, and (iii) conducting a quantitative uncertainty analysis in accordance with metrological standards (e.g., ISO-GUM [39]), including a complete measurement model, an uncertainty budget, and traceability to SI units. Additional research will address robustness under more challenging conditions, such as dirty or partially occluded wheels and diverse rim geometries, and will investigate dedicated mechanical fixtures to improve stability and repeatability. Finally, we envision leveraging advanced machine-learning techniques not only for segmentation but also for data-driven uncertainty estimation, thereby strengthening the connection between this work and broader topics in three-dimensional metrology and machine learning for metrology.

Author Contributions

Conceptualization, B.S. and E.Z.; methodology, B.S. and E.Z.; software, B.S.; validation, B.S., E.Z. and H.L.; formal analysis, B.S., E.Z. and H.L.; investigation, B.S. and E.Z.; resources, E.Z.; data curation, B.S.; writing—original draft preparation, B.S.; writing—review and editing, B.S., E.Z. and H.L.; visualization, B.S.; supervision, E.Z.; project administration, E.Z.; funding acquisition, E.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Nature Science Foundation of China under Grant 62173133, in part by the Postgraduate Study Abroad Program by China Scholarship Council under Grant 202306130094, and in part by the Postgraduate Scientific Research Innovation Project of Hunan Province CX20220398.

Data Availability Statement

The datasets used for the examples described outside Section 7.1.7 are not readily available, because the data are part of an ongoing study. The publicly available dataset [38] was used for the example described in Section 7.1.7.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yunta, J.; Garcia-Pozuelo, D.; Diaz, V.; Olatunbosun, O. Influence of camber angle on tire tread behavior by an on-board strain-based system for intelligent tires. Measurement 2019, 145, 631–639. [Google Scholar] [CrossRef]
  2. Xu, G.; Wang, Y.; He, W.; Shen, H.; Chen, F.; Li, X.T.; Zhao, X.X. Large-scale all-wheel alignment globally registered by laser line family and verified by global benchmark. IEEE Trans. Instrum. Meas. 2024, 73, 5013510. [Google Scholar] [CrossRef]
  3. Young, J.S.; Hsu, H.Y.; Chuang, C.Y. Camber angle inspection for vehicle wheel alignments. Sensors 2017, 17, 285. [Google Scholar] [CrossRef]
  4. Feng, L.H.; Chen, W.; Cheng, M.; Zhang, W.G. The gravity-based approach for online recalibration of wheel force sensors. IEEE/ASME Trans. Mechatron. 2019, 24, 1686–1697. [Google Scholar] [CrossRef]
  5. Chatur, S. Computer based wireless automobile wheel alignment system using accelerometer. Inter. J. Eng. Sci. 2015, 4, 62–69. [Google Scholar]
  6. D’Mello, G.; Gomes, R.; Mascarenhas, R.; Ballal, S.; Kamath, V.S.; Lobo, V.J. Wheel alignment detection with IoT embedded system. Mater. Today Proc. 2022, 52, 1924–1929. [Google Scholar] [CrossRef]
  7. Bohari, A.A.; Hafiz, M.F.H.M.; Yi, S.S.; Jamal, N.; Talib, M.N.M.; Safuan, S.N.M. Development of automobile wheel smart alignment monitoring system. PaperASIA 2024, 40, 28–35. [Google Scholar] [CrossRef]
  8. Lee, H.; Choi, S.B. Online detection of toe angle misalignment based on lateral tire force and tire aligning moment. Int. J. Automot. Technol. 2023, 24, 623–632. [Google Scholar] [CrossRef]
  9. Tang, X.L.; Shi, Y.; Chen, B.; Longden, M.; Farooq, R.; Lees, H.; Jia, Y. A miniature and intelligent low-power in situ wireless monitoring system for automotive wheel alignment. Measurement 2023, 211, 112578. [Google Scholar] [CrossRef]
  10. Song, L.M.; Wang, R.H.; Chen, E.Z.; Yang, Y.G.; Zhu, X.J.; Liu, M.Y. Research on global calibration method of large-scene multi-vision sensors in wheel alignment. Meas. Sci. Technol. 2022, 32, 105023. [Google Scholar] [CrossRef]
  11. Xu, G.; Wei, H.; Fang, C.; Hui, S.; Li, X.T. Automatic and accurate vision-based measurement of camber and toe-in alignment of vehicle wheel. IEEE Trans. Instrum. Meas. 2022, 71, 5024613. [Google Scholar] [CrossRef]
  12. Xu, G.; Shen, H.; Li, X.T.; Chen, F.; He, W. Large-range reconstruction with non-pre-calibrated camera via MBM of laser referenced by radial-collinear-features. IEEE Trans. Instrum. Meas. 2022, 71, 5013510. [Google Scholar] [CrossRef]
  13. Ge, P.X.; Wang, H.Q.; Wang, Y.H.; Wang, B. Calibration of ring multicamera system with transparent target for panoramic measurement. IEEE Sens. J. 2022, 22, 23154–23164. [Google Scholar] [CrossRef]
  14. Jiang, T.; Cui, H.; Cheng, X.S. A calibration strategy for vision-guided robot assembly system of large cabin. Measurement 2020, 163, 107991. [Google Scholar] [CrossRef]
  15. Roshan, M.C.; Isaksson, M.; Pranata, A. A geometric calibration method for thermal cameras using a ChArUco board. Infrared Phys. Technol. 2024, 138, 105219. [Google Scholar] [CrossRef]
  16. Xu, G.; He, W.; Chen, F.; Shen, H.; Li, X.T. One-dimension orientation method of caster and kingpin inclination of vehicle wheel alignment. Measurement 2022, 198, 111371. [Google Scholar] [CrossRef]
  17. Padegaonkar, A.; Brahme, M.; Bangale, M.; Raj, A.N.J. Implementation of machine vision system for finding defects in wheel alignment. Int. J. Comput. Inf. Technol. 2014, 1, 339–344. [Google Scholar]
  18. Furferi, R.; Governi, L.; Volpe, Y.; Carfagni, M. Design and assessment of a machine vision system for automatic vehicle wheel alignment. Int. J. Adv. Robot. Syst. 2013, 10, 242. [Google Scholar] [CrossRef]
  19. Senjalia, J.; Pandya, P.; Kapadia, H. Measurement of wheel alignment using camera calibration and laser triangulation. In Proceedings of the 2013 Nirma University International Conference on Engineering (NUiCONE), Ahmedabad, India, 28–30 November 2013; pp. 1–5. [Google Scholar]
  20. Kim, S.H.; Lee, K.I. Wheel alignment of a suspension module unit using a laser module. Sensors 2020, 20, 1648. [Google Scholar] [CrossRef] [PubMed]
  21. Baek, D.; Cho, S.; Bang, H. Wheel alignment inspection by 3D point cloud monitoring. Mech. Sci. Technol. 2014, 28, 1465–1471. [Google Scholar] [CrossRef]
  22. Wang, J.; Zeng, Z.; Sharma, P.K.; Alfarraj, O.; Tolba, A.; Zhang, J.; Wang, L. Dual-path network combining CNN and transformer for pavement crack segmentation. Autom. Constr. 2024, 158, 105217. [Google Scholar] [CrossRef]
  23. Gao, G.; Li, J.Y.; Yang, L.; Liu, Y.H. A multi-scale global attention network for blood vessel segmentation from fundus images. Measurement 2023, 222, 113553. [Google Scholar] [CrossRef]
  24. Geetha, G.K.; Yang, H.J.; Sim, S.H. Fast detection of missing thin propagating cracks during deep-learning-based concrete crack/non-crack classification. Sensors 2023, 23, 1419. [Google Scholar]
  25. Siriborvornratanakul1, T. Image segmentation for thin structures using a zero-shot learner. Int. J. Inf. Technol. 2025, 17, 721–726. [Google Scholar] [CrossRef]
  26. Li, J.Y.; Gao, G.; Yang, L.; Bian, G.B.; Liu, Y.H. DPF-Net: A dual-path progressive fusion network for retinal vessel segmentation. IEEE Trans. Instrum. Meas. 2023, 72, 2517817. [Google Scholar] [CrossRef]
  27. Chen, L.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  28. Woo, S.; Park, J.; Lee, J.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  29. Xu, M.B.; Han, Y.M.; Zhong, X.T.; Sang, F.Y.; Zhang, Y.A. A precise registration method for large-scale urban point clouds based on phased and spatial geometric features. Meas. Sci. Technol. 2025, 36, 015202. [Google Scholar] [CrossRef]
  30. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  31. Yuan, Y.H.; Huang, L.; Guo, J.Y.; Zhang, C.; Chen, X.L.; Wang, J.D. OCNet: Object context for semantic segmentation. Int. J. Comput. Vis. 2021, 129, 2375–2398. [Google Scholar] [CrossRef]
  32. Zunair, H.; Hamza, A.B. Sharp U-Net: Depthwise convolutional network for biomedical image segmentation. Comput. Biol. Med. 2021, 136, 104699. [Google Scholar] [CrossRef] [PubMed]
  33. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.0399. [Google Scholar] [CrossRef]
  34. Zhou, Z.W.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A nested U-Net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  35. Huang, H.M.; Lin, L.F.; Tong, R.F.; Hu, H.J.; Zhang, Q.W.; Iwamoto, Y.; Han, X.; Chen, Y.-W.; Wu, J. Unet 3+: A full-scale connected unet for medical image segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar]
  36. Chen, J.N.; Lu, Y.Y.; Yu, Q.H.; Luo, X.D.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  37. Cao, H.; Wang, Y.Y.; Chen, J.; Jiang, D.S.; Zhang, X.P.; Tian, Q.; Wang, M.N. Swin-Unet: Unet-like pure Transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 205–218. [Google Scholar]
  38. Available online: https://www.kaggle.com/datasets/adamnovozmsk/cawdec?resource=download (accessed on 31 May 2019).
  39. ISO/IEC. ISO/IEC Guide 98-6:2021(en), Uncertainty of Measurement—Part 6: Developing and Using Measurement Models; International Organization for Standardization: Geneva, Switzerland, 2021. [Google Scholar]
Figure 1. Overall framework of the proposed solution.
Figure 1. Overall framework of the proposed solution.
Metrology 06 00004 g001
Figure 2. Schematic illustration of the wheel rim.
Figure 2. Schematic illustration of the wheel rim.
Metrology 06 00004 g002
Figure 3. The network architecture of APCS-SwinUnet.
Figure 3. The network architecture of APCS-SwinUnet.
Metrology 06 00004 g003
Figure 4. The network architecture of ASPP.
Figure 4. The network architecture of ASPP.
Metrology 06 00004 g004
Figure 5. The architecture of the CBAM.
Figure 5. The architecture of the CBAM.
Metrology 06 00004 g005
Figure 6. Schematic diagram of point cloud storage.
Figure 6. Schematic diagram of point cloud storage.
Metrology 06 00004 g006
Figure 7. Schematic diagram of the measurement platform.
Figure 7. Schematic diagram of the measurement platform.
Metrology 06 00004 g007
Figure 8. Schematic diagram of the Azure Kinect.
Figure 8. Schematic diagram of the Azure Kinect.
Metrology 06 00004 g008
Figure 9. Schematic diagram of metrics. (a) Schematic diagram of Dice. (b) Schematic diagram of HD95.
Figure 9. Schematic diagram of metrics. (a) Schematic diagram of Dice. (b) Schematic diagram of HD95.
Metrology 06 00004 g009
Figure 10. The segmented results produced by different networks.
Figure 10. The segmented results produced by different networks.
Metrology 06 00004 g010
Figure 11. Wheel images under various illumination conditions. (a) Sunlight at noon. (b) Afternoon. sunlight. (c) Backlit with shadow. (d) Partial shadow.
Figure 11. Wheel images under various illumination conditions. (a) Sunlight at noon. (b) Afternoon. sunlight. (c) Backlit with shadow. (d) Partial shadow.
Metrology 06 00004 g011
Figure 12. Visualization of feature maps at different stages of the network. (ad) Encoder stages 1~4 (left side of Figure 3). (eh) Decoder stage 5~8 (right side of Figure 3).
Figure 12. Visualization of feature maps at different stages of the network. (ad) Encoder stages 1~4 (left side of Figure 3). (eh) Decoder stage 5~8 (right side of Figure 3).
Metrology 06 00004 g012
Figure 13. Examples from the public wheel dataset. (af) Sample illustrations from various, publicly available wheel datasets.
Figure 13. Examples from the public wheel dataset. (af) Sample illustrations from various, publicly available wheel datasets.
Metrology 06 00004 g013
Figure 14. Point cloud extraction results of the wheel rim. (a) RGB image of the wheel. (b) Segmented mask of the wheel rim. (c) Original point cloud. (d) Extracted point cloud corresponding to the wheel rim.
Figure 14. Point cloud extraction results of the wheel rim. (a) RGB image of the wheel. (b) Segmented mask of the wheel rim. (c) Original point cloud. (d) Extracted point cloud corresponding to the wheel rim.
Metrology 06 00004 g014
Figure 15. Point cloud registration results between the initial point clouds and target point clouds. (ac) Initial and target point clouds. (df) Corresponding point cloud registration results.
Figure 15. Point cloud registration results between the initial point clouds and target point clouds. (ac) Initial and target point clouds. (df) Corresponding point cloud registration results.
Metrology 06 00004 g015
Figure 16. Illustration of pixel-wise mask shifting. (a) Original mask. (b) Left shift in the mask. (c) Right shift in the mask.
Figure 16. Illustration of pixel-wise mask shifting. (a) Original mask. (b) Left shift in the mask. (c) Right shift in the mask.
Metrology 06 00004 g016
Figure 17. Wheel images under different illumination conditions and distances. (ad) Wheel images captured at 12:45 p.m., 11:39 a.m., 18:01 p.m. and 19:08 p.m., respectively.
Figure 17. Wheel images under different illumination conditions and distances. (ad) Wheel images captured at 12:45 p.m., 11:39 a.m., 18:01 p.m. and 19:08 p.m., respectively.
Metrology 06 00004 g017
Table 1. Comparisons with different networks.
Table 1. Comparisons with different networks.
MethodsDice (%)HD95
FCN84.128.62
DeepLabV386.678.18
OcNet88.118.06
U-Net85.7611.90
Att-Unet88.659.65
U-Net 2+89.103.00
U-Net 3+89.493.43
TransUnet89.852.56
SwinUnet90.232.47
APCS-SwinUnet90.662.11
Table 2. Comparisons under inconsistent illumination.
Table 2. Comparisons under inconsistent illumination.
MethodsDice (%)HD95
FCN84.3516.95
DeepLabV386.5710.51
OcNet87.7417.48
U-Net84.1021.48
Att-Unet86.9731.55
U-Net 2+86.8925.86
U-Net 3+88.6012.46
TransUnet90.192.80
SwinUnet90.162.71
APCS-SwinUnet90.462.51
Table 3. Comparisons of wheel public dataset.
Table 3. Comparisons of wheel public dataset.
MethodsDice (%)HD95
FCN70.3828.79
DeepLabV372.474.37
OcNet72.244.58
U-Net72.8716.68
Att-Unet74.316.13
U-Net 2+74.495.53
U-Net 3+75.509.51
TransUnet75.354.00
SwinUnet75.279.92
APCS-SwinUnet75.883.84
Table 4. Angle measurement results.
Table 4. Angle measurement results.
ψ (rad) θ (rad) ϕ (rad) ψ (°) θ (°) ϕ (°)
−0.00020 0.04952 −0.00340 −0.011462.83729−0.19481
Table 5. Angle measurement results under noisy point clouds.
Table 5. Angle measurement results under noisy point clouds.
Noise LevelReal (°)Meas (°)Error (°) R e (%)
0.012.85002.8373−0.01270.45
0.022.85002.8341−0.01590.56
0.032.85002.8333−0.01670.59
0.042.85002.8334−0.01660.58
0.052.85002.8339−0.01610.56
0.062.85002.8348−0.01520.53
0.072.85002.8340−0.01600.56
0.082.85002.8333−0.01670.59
0.092.85002.8339−0.01610.56
0.102.85002.8336−0.01640.58
Table 6. Measurement results of toe angles.
Table 6. Measurement results of toe angles.
No D s (mm) D e (mm)Real (°)Meas (°)Error (°) R e (%)
1687743−1.450−1.457−0.0070.48
26777542.8502.837−0.0130.46
3691733−1.650−1.755−0.1056.36
Table 7. Measurement results of camber angles.
Table 7. Measurement results of camber angles.
No D s (mm) D e (mm)Real (°)Meas (°)Error (°) R e (%)
1751802−0.750−0.807−0.0577.60
2778821−0.650−0.685−0.0355.38
37508083.2003.3580.1584.94
Table 8. Registration time and angle estimation across sampling rates.
Table 8. Registration time and angle estimation across sampling rates.
Sampling RateTime (s)Real (°)Meas (°) R e (%)
10%260.1051.766−1.08438.04
20%1006.0791.747−1.10338.70
30%2461.5751.732−1.11839.23
40%5875.7711.728−1.12239.37
50%8530.9511.728−1.12239.37
60%10,223.7561.721−1.12939.61
70%16,907.1511.720−1.13039.65
80%24,226.5771.723−1.12739.54
Table 9. Angle measurement results under different mask pixel shifts.
Table 9. Angle measurement results under different mask pixel shifts.
S L M L (°) A e (°) R e (%) S R M R (°) A e (°) R e (%)
0−1.4570.0110.480−1.4570.0110.48
−1−1.4390.0110.761−1.469−0.0191.31
−2−1.4240.0261.792−1.481−0.0312.14
−3−1.4160.0342.343−1.498−0.0483.31
−4−1.4050.0453.104−1.510−0.0604.14
−5−1.3940.0563.865−1.528−0.0785.38
−6−1.3570.0936.416−1.543−0.0936.41
−7−1.3210.1298.907−1.560−0.1107.59
−8−1.2910.15910.978−1.594−0.1449.93
−9−1.2570.19313.319−1.612−0.16211.17
−10−1.2020.24817.1010−1.666−0.21614.90
−15−0.8400.61042.0715−2.019−0.56939.24
−20−0.4800.97066.9020−2.431−0.98167.66
−300.2071.657114.2830−3.101−1.651113.86
−400.9632.413166.4140−3.833−2.383164.34
−501.8503.300227.5950−4.892−3.442237.38
Table 10. Angle measurement results of different segmentation methods.
Table 10. Angle measurement results of different segmentation methods.
MethodsDice (%)HD95Meas (°)Error (°) R e (%)
FCN83.292.243.4250.2257.03
DeepLabV385.123.003.5300.33010.31
OcNet86.362.243.5150.3159.84
U-Net87.944.473.7690.56917.78
Att-Unet89.673.613.6460.44613.94
U-Net 2+89.134.002.741−0.45914.34
U-Net 3+89.272.002.980−0.2206.88
TransUnet89.722.832.874−0.32610.19
SwinUnet90.092.243.4040.2046.38
APCS-SwinUnet90.552.003.3580.1584.94
Table 11. Repeated experiments under different distances and illuminations.
Table 11. Repeated experiments under different distances and illuminations.
NoToe−InToe−OutNegative−CamberPositive−Camber
12:45 p.m.~12.59 p.m.11:39 a.m.~11:57 a.m.18:01 p.m.~18:19 p.m.19:08 p.m.~19:25 p.m.
D s = 698 mm D e = 744 mm D s = 704 mm D e = 742 mm D s = 769 mm D e = 831 mm D s = 698 mm D e = 745 mm
Real (°)Meas (°)Error (°)Real (°)Meas (°)Error (°)Real (°)Meas (°)Real (°)Meas (°)Error (°)Real (°)
1−0.900−0.902−0.0021.4001.307−0.093−0.900−1.014−0.1141.2501.3230.073
2−0.900−0.946−0.0461.4001.292−0.108−0.900−0.932−0.0321.2501.2600.010
3−0.900−0.933−0.0331.4001.227−0.173−0.900−0.954−0.0541.2501.224−0.026
4−0.900−0.8770.0231.4001.283−0.117−0.900−1.000−0.1001.2501.228−0.022
5−0.900−0.912−0.0121.4001.281−0.119−0.900−0.997−0.0971.2501.3150.065
6−0.900−0.8730.0271.4001.348−0.052−0.900−0.935−0.0351.2501.4010.151
7−0.900−0.904−0.0041.4001.311−0.089−0.900−0.8870.0131.2501.3250.075
8−0.900−0.8730.0271.4001.259−0.141−0.900−0.8880.0121.2501.2890.039
9−0.900−0.8380.0621.4001.314−0.086−0.900−0.924−0.0241.2501.2870.037
10−0.900−0.8630.0371.4001.318−0.082−0.900−0.938−0.0381.2501.3890.139
11−0.900−0.906−0.0061.4001.380−0.020−0.900−0.997−0.0971.2501.3460.096
12−0.900−0.959−0.0591.4001.355−0.045−0.900−0.925−0.0251.2501.2740.024
13−0.900−0.927−0.0271.4001.313−0.087−0.900−0.926−0.0261.2501.2510.001
14−0.900−0.8700.0301.4001.358−0.042−0.900−0.958−0.0581.2501.235−0.015
15−0.900−0.906−0.0061.4001.368−0.032−0.900−0.974−0.0741.2501.3420.092
16−0.900−0.8730.0271.4001.258−0.142−0.900−0.958−0.0581.2501.3920.142
17−0.900−0.924−0.0241.4001.256−0.144−0.900−0.8910.0091.2501.3160.066
18−0.900−0.909−0.0091.4001.214−0.186−0.900−0.907−0.0071.2501.3040.054
19−0.900−0.8500.0501.4001.266−0.134−0.900−0.944−0.0441.2501.2830.033
20−0.900−0.8990.0011.4001.279−0.121−0.900−0.948−0.0481.2501.3810.131
21−0.900−0.8890.0111.4001.342−0.058−0.900−0.922−0.0221.2501.3430.093
22−0.900−0.918−0.0181.4001.326−0.074−0.900−0.8510.0491.2501.2750.025
23−0.900−0.902−0.0021.4001.272−0.128−0.900−0.8680.0321.2501.248−0.002
24−0.900−0.8420.0581.4001.331−0.069−0.900−0.907−0.0071.2501.236−0.014
25−0.900−0.8770.0231.4001.342−0.058−0.900−0.909−0.0091.2501.3430.093
MAE−0.0450.010
RMSE0.0780.066
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shi, B.; Liu, H.; Zappa, E. A Flexible Wheel Alignment Measurement Method via APCS-SwinUnet and Point Cloud Registration. Metrology 2026, 6, 4. https://doi.org/10.3390/metrology6010004

AMA Style

Shi B, Liu H, Zappa E. A Flexible Wheel Alignment Measurement Method via APCS-SwinUnet and Point Cloud Registration. Metrology. 2026; 6(1):4. https://doi.org/10.3390/metrology6010004

Chicago/Turabian Style

Shi, Bo, Hongli Liu, and Emanuele Zappa. 2026. "A Flexible Wheel Alignment Measurement Method via APCS-SwinUnet and Point Cloud Registration" Metrology 6, no. 1: 4. https://doi.org/10.3390/metrology6010004

APA Style

Shi, B., Liu, H., & Zappa, E. (2026). A Flexible Wheel Alignment Measurement Method via APCS-SwinUnet and Point Cloud Registration. Metrology, 6(1), 4. https://doi.org/10.3390/metrology6010004

Article Metrics

Back to TopTop