You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

8 December 2025

Low-Cost Optical Displacement Measurement for SHM Applications Supported by CNN Object Detection

,
and
Structural Concrete Institute, Leipzig University of Applied Sciences (HTWK Leipzig), 04277 Leipzig, Germany
*
Author to whom correspondence should be addressed.
This article belongs to the Section Civil Engineering

Abstract

This paper presents a cost-effective approach to structural health monitoring (SHM) based on standard image processing and convolutional neural networks (CNNs) for object detection. The proposed algorithm accurately identifies and tracks geometric measurement motives across image sequences, enabling precise position and two-dimensional displacement determination. Experimental investigations using a minimal implementation with open-source Python (3.12) libraries demonstrate close agreement with reference measurements and low measurement noise. The study also highlights how geometric shape selection, motive arrangement, and preprocessing techniques influence measurement accuracy. This robust, scalable, minimally invasive method offers a low-cost alternative to traditional SHM systems. Its flexibility allows it to be adapted to various infrastructures. Potential future enhancements include strain measurements from multiple motives, multi-plane monitoring, and machine learning–based error correction. These features suggest that the approach is a promising solution for reliable, affordable, and adaptable SHM.

1. Introduction

Structural health monitoring (SHM) of infrastructure has become increasingly important and represents an indispensable tool for ensuring the integrity and safety of buildings and civil structures [1,2,3]. In view of recent structural failures such as bridge collapses, continuous monitoring and analysis of structural conditions have gained growing attention in both research and practical applications. Modern infrastructure is exposed to numerous stressors, including excessive use beyond the intended service life, increasing heavy traffic, progressive ageing, and extreme weather events associated with climate change. Consequently, innovative SHM solutions are essential to guarantee the long-term safety and reliability of critical infrastructure. In addition to established SHM methods, new approaches are becoming increasingly relevant. The present study demonstrates the advantages of image-based measurement systems employing standard cameras. These systems are cost-effective, require only simple hardware components, and can operate under varying environmental conditions. They can also be retrofitted into existing monitoring setups, are minimally invasive, and can be implemented using open-source software.
Systematic monitoring of structures, combined with targeted maintenance actions derived from measurement data, contributes significantly to resource conservation and the prevention of potential structural failures [3,4]. Within this context, methods capable of detecting deformations and displacements are of particular scientific and practical relevance. Beyond real-time monitoring, SHM enables data-driven condition forecasting, allowing maintenance intervals and preventive measures to be optimized [5,6]. As modern construction techniques and materials become more widespread, the scalability of monitoring systems—achieved through the integration of distributed measurement points—has become crucial for ensuring adequate surveillance of complex and large-scale structures [7].
A wide range of established techniques exist for monitoring deformations and displacements, each differing in measurement principle, accuracy, and field applicability [8,9,10,11]. Classical methods include strain gauges, which provide direct and highly accurate strain measurements [12]. However, they require direct access to measurement points and are limited to local areas [13]. Fiber optic sensors, particularly spatially resolved systems, have become increasingly popular owing to their ability to measure strain over long distances and their resistance to electromagnetic interference. Fiber Bragg grating sensors allow discrete measurements at defined fiber locations, whereas distributed fiber optic sensors provide continuous strain profiles, enabling both local and global deformation analysis in complex structures [14]. Additional methods include displacement transducers such as inductive, capacitive, and potentiometric sensors [15], which offer high precision at discrete points and are suitable for critical regions. Laser-based sensors provide another contactless alternative, capable of detecting positional changes with high accuracy by measuring the distance to a fixed reflector point [10]. However, they require an unobstructed line of sight. Camera-based techniques have also gained prominence as flexible and versatile tools for displacement detection [10,16,17]. These systems can monitor the movement of entire structural segments, while digital image correlation (DIC) enables high-resolution surface deformation analysis [18]. Such methods capture both local and global displacements, though they rely on adequate image quality and stable environmental conditions. Each monitoring principle offers specific advantages and limitations; therefore, selection of an appropriate method depends on structural characteristics, environmental influences, measurement accuracy, range, and installation effort [3].
Despite the diversity of available techniques, current SHM systems still face technical and practical challenges that restrict their widespread implementation [3,4]. A key limitation lies in the high cost of complex measurement systems [4,19,20], largely due to expensive sensors, specialized hardware components such as interrogators and amplifiers, and maintenance-intensive installations. The integration of sensors and cables—such as fiber optic splicing—requires technical expertise and may introduce installation errors that compromise measurement accuracy. Furthermore, many systems provide single-point measurements, which limits their applicability for large-scale monitoring. Environmental influences such as temperature and humidity variations necessitate compensation procedures, increasing the complexity of data acquisition. Dependence on proprietary software further reduces system flexibility, hinders interoperability between different sensor types, and results in long-term costs due to vendor lock-in [21]. In addition, many sensing systems are mechanically sensitive and unsuitable for continuous operation in harsh environments [22]. Power supply and data transmission present additional challenges for wired systems, as cabling across large structures is logistically demanding and costly [19,21].
To overcome these limitations, cost-effective image-based measurement systems using standard cameras have emerged as a promising alternative [23]. Single-camera setups offer low acquisition costs and eliminate the need for proprietary measuring hardware by employing commercially available industrial or modified digital single-lens reflex (DSLR) cameras [23,24]. Their modular system architecture allows flexible adaptation to various monitoring scenarios, while data processing can rely on open-source algorithms and image analysis methods [25]. A single camera can monitor multiple measurement points simultaneously [26], and specially designed measurement patterns enable sub-micrometre displacement detection [24]. This technique can capture a wide range of structural changes, from simple tilting to complex deformation patterns [11,27]. The use of open-source data analysis tools enhances transparency, adaptability, and independence from proprietary software [28,29]. Furthermore, machine learning algorithms have recently been proposed to improve the accuracy of displacement measurements by automating feature detection and motive tracking [30]. This approach is minimally invasive, as measurement devices can be temporarily or permanently affixed to structural surfaces without requiring complex integration. Additionally, this method can be incorporated into existing SHM systems, enabling gradual expansion of these systems.
Previous vision-based SHM studies have demonstrated the feasibility of displacement measurement using single-target tracking, full-field optical methods, or simplified background scenarios [31,32,33,34]. However, the proposed approach combines convolutional neural network (CNN)-based object detection, motive-specific geometric contour analysis, and centroid-based sub-pixel displacement estimation. This combination allows for the detection and differentiation of multiple visually similar measurement motives within larger image regions, including those with complex or busy backgrounds. The detected motives can be linked to a database containing geometric and dimensional metadata, facilitating automated Region of Interest (ROI) assignment and downstream processing. The resulting system is modular and scalable, supporting controlled laboratory validation and potential adaptation to field monitoring applications.

2. General Considerations

The following section presents the key considerations for determining the position and displacement of measurement motives (MMs) using a basic single-camera system. The proposed method can be readily scaled, optimized, and implemented through open-source algorithms. The approach is structured around four fundamental components:
  • Pre-selection of regions of interest (ROIs) based on object detection.
  • ROI segmentation.
  • Relative and absolute position determination.
  • Displacement calculation.
Figure 1 illustrates the basic workflow of the proposed method, and the key considerations for each component are described below.
Figure 1. Basic workflow.
The measurement sequence depends on the design of the MMs, which fulfills two primary functions. First, it determines the center of gravity, thereby defining the MM’s position within the image. Second, it incorporates known geometric shapes to establish the image scale in the measurement plane. This step is essential for converting pixel-based displacement values into real-world distances.
The motive should be designed so that its components can be reliably segmented and identified using standard image-processing techniques. Although no strict design rules exist, simple geometric shapes—such as circles, triangles, or rectangles—are advantageous, as they can be readily distinguished from one another. Double-symmetrical arrangements are generally preferred to enhance robustness.
The required motive size depends on the measurement task and imaging parameters, including sensor size, lens characteristics, and measurement distance. The relative size of the motive within the image frame strongly influences the accuracy of the evaluation algorithm.

2.1. Pre-Select ROIs Based on Object Detection

The design of suitable MMs provides the foundation for reliable image-based displacement determination. Once an appropriate MM has been defined, the next step involves its detection and localization within recorded images.
Detecting MMs and geometric patterns in large image formats can be challenging, particularly when complex background configurations are present [35]. This difficulty increases when multiple, visually similar MMs must be distinguished. In SHM applications, however, reliable detection and classification of several MMs within large image regions containing busy backgrounds are essential.
The proposed approach uses object detection with a convolutional neural network (CNN) to address these challenges. This allows for the simultaneous detection and identification of multiple objects. Defining pre-selected ROIs restricts processing to specific areas, reducing potential interference and optimizing computational requirements.
A CNN is a feedforward neural network that automatically extracts and recognizes image features through kernel optimization. It is a core technique in modern machine learning applications. Compared to classical methods, such as template matching or handcrafted feature detection, CNN-based object detection demonstrates advantages in robustness, scalability, and adaptability across varying imaging conditions [36]. Classical detection pipelines, which rely on sliding windows, manual template matching, or handcrafted descriptors, may exhibit limited generalization and require extensive tuning, particularly in cluttered or complex visual scenes. In contrast, CNN architectures can discriminate between visually similar objects and maintain detection reliability under background clutter, illumination changes, and minor geometric variations [37,38,39]. Comparative evaluations indicate that one-stage CNN detectors (e.g., SSD and YOLO) offer high detection accuracy, efficient inference, and moderate resource consumption [40]. These characteristics are relevant for real-time or resource-constrained applications such as structural health monitoring (SHM).
Several CNN-based object recognition algorithms are available as open-source libraries for common programming languages, including C and Python. Examples of these libraries include Faster R-CNN, SSD, and YOLO [41]. YOLO currently provides the best performance in terms of inference speed and reliability for real-time applications. However, the choice of algorithm should be adapted to the application’s specific requirements. In the present approach, object detection primarily supports displacement measurement, for which processing efficiency and resource optimization are important. While object detection alone lacks the precision necessary to determine the MM center for robust displacement estimation, it generates essential ROIs and supports plausibility validation.
Furthermore, recognized MMs can be linked to a database containing metadata such as the number and dimensions of geometric elements. This metadata is subsequently utilized in downstream processing. The following parameters are required as outputs of the detection step:
  • IDs of detected objects,
  • Associated confidence scores,
  • Coordinates of bounding boxes, and
  • Coordinates of centroids.

2.2. ROI Segmentation

The identified ROIs must be pre-processed before further analysis. Edge detection prior to segmentation improves both accuracy and efficiency. Widely used edge detection algorithms include those developed by Prewitt, Canny, and Sobel [42]. A critical step in edge detection is image smoothing, which reduces noise and softens edge transitions. While smoothing minimizes false edge detection and discontinuities in pixel intensity, it can also result in the loss of important information. Therefore, the choice of smoothing method should be tailored to the application, with median, Gaussian, and bilateral filters being commonly employed. Detailed guidance on algorithm selection, smoothing operations, and input parameters (e.g., kernel size) can be found in the literature [43,44,45,46,47]. The output of this step consists of the preselected ROIs represented as logical matrices.
To assign geometric elements to the MM, closed contours must first be identified within the edge-detected logical matrices. Various contour extraction approaches have been described in the literature [48,49]. Most algorithms trace the outer boundaries of closed regions, allowing the separation of shapes within the ROIs into individual elements. These elements can then be classified based on their geometric properties. Differentiation according to the number of corners is particularly useful, with common detectors including the Harris-Stephens [50,51] and Shi-Tomasi algorithms [52]. For the purposes of this study, determining the number of corners is sufficient, making these approaches adequate. Exact corner localization, as provided by the Förstner algorithm [53], is not required. Elements without corners, such as circles or ellipses, can be categorized based on metrics such as the circumference-to-area ratio or through the Hough transform [54]. Once contours are detected and classified, additional properties, including surface area and geometric centers, can be computed.
A simultaneous plausibility check is performed based on characteristics such as the number of matched geometric shapes, allowing incorrect segmentations to be detected early.
The following parameters are transferred to the next processing step:
  • Recognized contours,
  • Classification of contours into geometric elements, and
  • Geometric properties of the elements (e.g., coordinates of geometric centers, surface area).

2.3. Relative/Absolute Position Determination

The segmentation results are subsequently used to determine the position of each MM relative to its corresponding ROI. Several approaches can be applied for this purpose. The most straightforward method involves calculating the overall center of gravity of all geometric shapes constituting the MM. To enhance accuracy or tailor the position determination to specific applications, the calculation can be restricted to selected shapes only.
Alternative methods include computing the MM center based on the relative distances between individual centroids. The choice of methodology largely depends on the MM design and the symmetry characteristics of its geometric arrangement. Since these approaches rely on averaged values derived from multiple centroids or inter-element distances, position determination can achieve subpixel-level precision.
Finally, the absolute position of each MM center within the overall image is obtained by referencing the ROI coordinates.

2.4. Displacement Calculation

Displacement is determined by comparing the positions of the MMs across multiple input images. The final step involves converting these relative displacements into physical units of length. To achieve this, the relationship between the optical image in the measurement plane and the actual size of the geometric elements on the MM must be established.
This requires precise knowledge of the dimensions of the geometric shapes. One approach is to calculate the ratio of the areas of selected or all elements. Alternatively, distances between the centers or edges of individual elements can be used. Several strategies can yield satisfactory results, allowing optimization for specific measurement tasks.
This calibration step is critical for ensuring measurement accuracy, as even small deviations can cause significant proportional errors. Once calibrated, the MM positions determined at different time points can serve as reference values for a sequence of displacement measurements across multiple images.

3. Experimental Investigations

To evaluate the accuracy and scalability of the proposed measurement algorithm, a minimal experimental setup was implemented. The objective was to design a basic MM and verify the algorithm’s performance using the simplest possible configuration.

3.1. Experimental Design

3.1.1. Measurement Motive

As part of the experimental investigations, a simple MM was developed incorporating various geometric shapes. The inclusion of different geometries enabled the assessment of how parameters such as the number of corners and corner angles influence the accuracy of the evaluation algorithm.
The selected configuration consisted of a symmetrical arrangement of right-angled isosceles triangles and rectangles with an aspect ratio of 2:1, combined with a centrally positioned circle (see Figure 2). For testing, the MM was printed on a 4 mm-thick aluminum Dibond panel and mounted onto a linear stage using an aluminum angle bracket. The MM had overall dimensions of 50 mm · 50 mm. Each rectangle measured 10 mm · 5 mm, while the triangles had side lengths of 10 mm. The central circle had a radius of 10 mm.
Figure 2. (a) Measurement motive (MM); (b) dimensioned representation of the MM.

3.1.2. Test Setup

The experimental setup consisted of a rigid frame constructed from aluminum system profiles. The camera was mounted on one side using a bracket made of 15 mm-thick steel plates. On the opposite side, a linear stage was installed to hold the MM via an aluminum angle bracket. The distance between the camera and the MM was 2.35 m.
Two LED spotlights were symmetrically positioned at a distance of 0.87 m from the test object. The illumination was adjusted to achieve an illuminance of 800 lx at a color temperature of 4400 K. The entire frame was vibration-isolated using six steel isolator springs combined with elastomeric bearings. The experiments were conducted in a controlled environment with a constant temperature of 20 °C and a relative humidity of 65%. A schematic of the test setup is shown in Figure 3.
Figure 3. Schematic test setup.
A monochrome industrial camera with a 2/3-inch sensor (A2A4200-12gmBAS, Basler AG, Ahrensburg, Germany) was used for image acquisition. The camera was equipped with a fixed focal length lens of 50 mm and an adjustable aperture range of F2.8–F16.0 (C23-5028-5M-P, Basler AG, Ahrensburg, Germany). The sensor (GMAX2509, Gpixel Microelectronics Inc., Chengchun, China) provides a maximum resolution of 4200 pixels  ·  2160 pixels, with a 12-bit pixel depth and a pixel size of 2.5 μ m   ·  2.5 μ m .
This minimal experimental setup did not include explicit camera calibration. The measurement motive was positioned close to the optical axis and within the central sensor region, where radial and tangential distortions are minimal. According to optical theory and empirical analyses, radial and tangential lens distortions are strongest at the sensor’s periphery and diminish near the optical axis [55]. Under the present conditions, residual distortions are expected to be small compared to the spatial variations of interest.
The linear stage (L-836.501212, PI SE & Co. KG, Karlsruhe, Germany) featured a bidirectional repeatability of 0.8 μ m and was equipped with a linear encoder offering a resolution of 80 nm. Owing to its high precision, the encoder readings were used as the position reference for all measurements. A stepped motion profile was applied, consisting of five displacement steps of 1.5 mm in both the positive and negative directions. Each position was held for 170 s, resulting in a total test duration of 60 min. The corresponding displacement path recorded by the stage’s linear encoder is shown in Figure 4. A separate data recorder was used to synchronously capture both the camera images and the encoder readings at intervals of 5 s.
Figure 4. Displacement path.

3.2. Minimal Implementation of the Algorithm

To verify the functionality of the proposed approach, a minimal implementation of the measurement algorithm described in Section 2 was developed. The algorithm was fully implemented in Python and leverages the following libraries:
  • ultralytics (8.3.99)
  • torch (2.6.0+cu126)
  • opencv-python (4.10.0.84)
  • numpy (1.26.4)
The subsequent paragraphs describe the individual components and functions of the algorithm, illustrating how object detection, ROI segmentation, geometric analysis, and displacement computation are implemented in this minimal example.

3.2.1. CNN Object Detection

For object detection, the real-time YOLO series detector (Generation 11) was employed. This CNN from Ultralytics Inc. is licensed under AGPL 3.0, permitting free non-commercial use. A custom dataset of 253 images was created to train the model, containing various scenes both with and without the MM present. The dataset was split into 222 training images and 31 validation images.
The model for this minimal implementation was trained using the pretrained YOLO11l network, which is based on the COCO dataset. Training was performed over 200 epochs with a target image size of 1024 pixels.

3.2.2. ROI Segmentation

Only the ROIs identified by the object detection step were used for further processing. This significantly reduces computational load by limiting the analysis to a small section of the image. Given the MM’s illumination and the high resolution of the camera sensor, a median filter (cv2.medianBlur()) with a 5 · 5 pixel window was applied to reduce noise. Edge detection was then performed on the filtered ROI using the Canny algorithm (cv2.Canny()) with gradient thresholds of 100 and 200.
Contour detection was performed using OpenCV’s findContours() function, based on [56]. The retrieval mode was set to RETR_EXTERNAL to consider only outer contour points, and the approximation method was set to CHAIN_APPROX_NONE to preserve all contour points. This function returns nested arrays of closed contours, which were iterated to classify individual shapes:
General    
Contours with an area smaller than 100 pixels were ignored to reduce noise-related false detections. This threshold should be adjusted according to the expected minimum size of the geometric shapes.
Circles     
Circularity was calculated as
C = 4 · π · A P 2
where A is the area and P is the perimeter of the contour [57]. A perfect circle has a circularity of 1. Shapes with circularity above 0.85 were classified as circles, based on empirical tuning for the test setup.
Polygons 
Triangles and rectangles were identified using the Ramer-Douglas-Peucker algorithm, implemented via approxPolyDP() [58,59], which approximates a curve by reducing the number of points. The approximation tolerance was set to 10% of the contour perimeter. The number of corners in each polygon was determined from the resulting array of approximated points. This method is robust for the current setup, though alternative corner detection algorithms may be preferred for different applications.
Finally, centroids were computed for all classified contours using image moments, and a plausibility check was performed by verifying the expected number of shapes in each category (one circle, four triangles, and four rectangles).

3.2.3. Relative/Absolute Position Determination

Two approaches were used to determine the overall center of gravity of the MM. First, the arithmetic mean of all shape centroids was calculated. Second, the center of gravity was calculated separately for each type of geometric shape. The nomenclature for these approaches is summarized in Table 1.
Table 1. Nomenclature for position determination approaches.
The relative coordinates of the center of gravity within the ROI can then be converted to absolute positions in the overall image by incorporating the ROI coordinates.

3.2.4. Image Scale Determination

Three approaches were used to determine the image scale:
Area         
The simplest method is to compare the actual size of the geometric shapes on the MM with the size of the enclosed area of the classified contours:
β = A A
where A is the contour area in pixels and A is the actual area of the corresponding shape.
Circles     
Circular shapes can be used by comparing parameters such as radius, circumference, or area. Here, the radius was determined using the OpenCV function minEnclosingCircle() on contours classified as circles. The actual circle radius is 10 mm.
Polygons 
In this approach, the image scale is determined by comparing the distances between centroids of the triangles and rectangles. The distances between the triangle centroids are 26.667 mm for adjacent elements and 37.712 mm for opposite elements. For rectangles, the distances are 24.749 mm for adjacent elements and 35 mm for opposite elements. The overall scale factor β is calculated using the mean of all respective shape ratios.
Table 2 summarizes the nomenclature for the individual image scale determination approaches.
Table 2. Nomenclature for image scale determination approaches.

3.2.5. Displacement Measurement

Finally, the displacements were determined by comparing the MM positions across multiple images and converting them into physical units using the corresponding image scale factor. Different combinations of position determination and image scale calculation were considered, as summarized in Table 3.
Table 3. Nomenclature for displacement measurement approaches.

4. Results

4.1. Object Detection

Object detection yielded satisfactory results across the entire dataset, achieving an average confidence score of 0.899. The detection quality, however, varied slightly depending on the position of the measurement motive, as illustrated in Figure 5.
Figure 5. Confidence scores of YOLO object detection across the dataset.

4.2. Filtering and Edge Detection

Figure 6 shows the ROI of the detected MM, on which edge detection was applied to the initial measurement image.
Figure 6. ROI of the detected MM with applied edge detection.
The application of a median filter results in smooth contours and effectively reduces noise. However, some sharp corners are slightly rounded, and minor rounding of edges can be observed.

4.3. Segmentation

The segmentation results were consistently reliable across the entire dataset. For each image, the correct number of closed contours was detected and successfully classified into geometric shapes. The plausibility checks confirmed that the expected number of circles, triangles, and rectangles were identified in all cases. This indicates that the chosen threshold value for corner detection is appropriate for the applied setup.
Figure 7 illustrates an example of the segmented geometric elements within the MM in the overall image.
Figure 7. Segmentation of geometric elements within the MM.

4.4. Determination of Motive Position

The centroids of the geometric shapes in the MM were determined using image moments. Figure 8 shows the categorized shapes and their centroids for the first image.
Figure 8. Centroids of the geometric shapes in the cropped ROI.
Figure 9 compares the absolute centroid positions for each shape category over the measurement period before the first displacement step. Noise remains in the subpixel range. Circle and rectangle centroids are closely aligned, while triangle centroids exhibit larger deviations, influencing the overall arithmetic mean of all elements.
Figure 9. Comparison of absolute centroid positions for each shape category.

4.5. Determination of the Image Scale

The image scale was calculated from the geometric elements in the MM to convert pixel displacements into physical units. Figure 10 shows the resulting image scale values for the different approaches across the dataset. Mean values and standard deviations are summarized in Table 4.
Figure 10. Image scale values obtained from different geometric elements.
Table 4. Mean values and standard deviations of the image scale determination methods.
Several observations can be made: the methods based on the area of all shapes (pxsza) and the circle diameter (pxszc) exhibit slightly lower values and higher variability compared to the centroid-distance approaches. In contrast, the scales determined from the distances between rectangle (pxszr) and triangle (pxszt) centroids are closely aligned and show minimal scatter. Overall, the centroid-distance methods provide the most consistent and reliable image scale for subsequent displacement measurements.

4.6. Displacement Measurement

Table 3 summarizes the combinations of position determination and image scale calculation methods used for displacement measurements. For consistency, combinations were chosen so that the same geometric shapes are used for position and scale determination.
First, displacements were calculated solely using circles (dispc), then using triangles (dispt) or rectangles (dispr). Finally, mixed combinations of position and scale from triangles and rectangles were used (disptr and disprt). Figure 11 compares the X-direction displacements for the first five steps.
Figure 11. Comparison of displacement calculation using different methods.
Displacements calculated solely from rectangles (dispr) show the best agreement with reference values, with deviations typically below ±10 μm even at large displacement steps. Measurement uncertainty remains low across all ranges, with an extended uncertainty below 0.6 μm (Figure 12 and Figure 13).
Figure 12. Absolute deviations of the measurements determined using rectangles.
Figure 13. Measurement uncertainty of the values determined using rectangles.
Displacements determined using triangles (dispt) exhibit moderate absolute deviations that vary with the displacement magnitude. While individual steps show low scatter, the overall accuracy is limited by a scaling-factor effect (Figure 14 and Figure 15). Uncertainty increases with displacement.
Figure 14. Absolute deviations of the measurements determined using triangles.
Figure 15. Measurement uncertainty of the values determined using triangles.
Mixed methods (disptr and disprt) produce intermediate deviations. Absolute deviations are moderate, but disptr shows asymmetrical deviations: negative X shifts produce larger errors than positive shifts (Figure 16 and Figure 17). In contrast, disprt exhibits a more balanced distribution of deviations (Figure 18 and Figure 19).
Figure 16. Absolute deviation of measurements determined using triangles for position and rectangles for image scale (disptr).
Figure 17. Measurement uncertainty of measurements determined using triangles for position and rectangles for image scale (disptr).
Figure 18. Absolute deviation of measurements determined using rectangles for position and triangles for image scale (disprt).
Figure 19. Measurement uncertainty of measurements determined using rectangles for position and triangles for image scale (disprt).
Measurements based solely on circles (dispc) and on the areas of all elements (dispa) showed large deviations from the reference values and high variability across all displacement steps; consequently, they were not investigated further, as they do not provide reliable displacement information. Figure 20 provides an overview of the absolute deviations for all examined methods based on polygons. Overall, rectangle-based methods provide the highest accuracy and lowest measurement uncertainty, triangle-based methods reveal scaling-dependent deviations, and circle-based measurements are least reliable.
Figure 20. Absolute deviation of measurements: Approaches based on polygons.

5. Discussion

The experimental investigations demonstrate that the developed image-based approach enables accurate and stable displacement measurements, even when implemented with a minimal setup. The CNN-based object detection provided high confidence values and reliable identification of MMs, despite a limited training dataset. The minor fluctuations observed in confidence levels likely result from positional dependencies and limited training diversity, but they do not significantly influence subsequent processing or measurement quality.
Edge detection within the extracted ROIs yielded robust contour recognition for all geometric elements. Nevertheless, information loss occurred at edges and corners due to preprocessing with a median filter and the Gaussian filter embedded in the cv.Canny() function. These filters effectively reduced noise and improved contour smoothness but caused rounded corners and slight edge shifts. As a result, the detected edges deviate marginally from their actual course, introducing minor inaccuracies in size and centroid estimation. However, since this effect is consistent across all edges, its influence on centroid determination is minimal—particularly for double-symmetrical shapes such as rectangles. For asymmetrical geometries, such as triangles, this loss of information has a stronger effect on the calculated centroid and, consequently, on the derived distances and scale factors.
To quantify this effect, both the root mean squared error (RMSE) and the mean absolute error (MAE) were calculated for the shape area of contours determined from unfiltered and filtered images across the entire dataset. As shown in Table 5, the computed values reveal that triangles exhibit greater information loss, presumably due to their acute corners. Unlike double-symmetrical rectangles, the isosceles triangles are more susceptible to deviations in centroid determination due to edge loss because of their shape. This phenomenon is evident in the inaccurate determination of the image scale.
Table 5. Error metrics for information loss in triangles and rectangles.
To further investigate shape-dependent deviations in determining the center of gravity, a quantitative edge width analysis was performed on rectangles and triangles. The left and right edges of each shape were evaluated separately. The edge profiles were analyzed using a one-dimensional plateau and transition detection algorithm. This algorithm extracts grayscale intensity profiles along selected lines in an image and identifies regions of plateaus, which represent the interior of a geometric shape, and regions of transitions, which represent shape boundaries. This method combines median filtering, Otsu’s thresholding [60], and local variance analysis inspired by derivative edge detection principles [61]. Stable plateaus were identified as regions of low local variance, and transitions were defined as contiguous segments with high variance exceeding a minimum width. Transitions were identified using run length analysis [62]. For each transition, the start and end positions, width, and direction of the grayscale gradient were calculated. Initial smoothing was performed using a median filter with a window size of three pixels. Then, the Otsu thresholding method was applied to the smoothed profile to enable binary segmentation into plateau and transition areas. Local variance was calculated over a sliding window of five pixels to identify stable plateaus (low variance) and transitions (high variance exceeding five grayscale values). This methodology was applied row-wise to all vertical edge areas of the triangles and rectangles during the measurement period before the first displacement step. This allowed the width or length of the transition areas to be determined and evaluated separately for the left and right edges. Figure 21 illustrates the procedure at a position that intersects two triangles and a rectangle.
Figure 21. Edge width analysis procedure for triangles and rectangles.
Table 6 shows the mean values and standard deviations of edge width, separated by edge side and geometric figure.
Table 6. Mean edge widths and standard deviations for triangles and rectangles.
The analysis reveals that triangles exhibit pronounced asymmetry in edge width, with left edges being wider than right edges. This asymmetry likely contributes to the observed centroid deviations because the broader left edges can shift the centroid position more substantially during image processing. In contrast, rectangles have relatively symmetric edge widths, which enables them to maintain a stable centroid position despite edge-induced distortions. The asymmetry in triangle edge widths may result from a combination of lighting conditions, print quality, angle of incidence, and the triangles’ geometric shape.
As shown in the results, the determination of the MM’s absolute position within the image is highly stable, exhibiting deviations of approximately 0.3 pixels (34 μ m ) in the X direction and 0.2 pixels (23 μ m ) in the Y direction between the applied position determination methods. This confirms that, in bilaterally symmetrical configurations, geometric variations have only a negligible influence on positional accuracy. Notably, these deviations correspond to absolute offsets rather than multiplicative (factorial) errors and therefore remain nearly constant across successive displacement steps. Consequently, their impact on displacement determination is minimal, as the relative positional relationships are preserved throughout the measurement sequence.
In contrast, deviations in image scale determination have a considerably stronger effect, since they introduce multiplicative errors that scale proportionally with the measured displacements. Variations in scale calibration of up to 4.5 μ m per pixel were observed, substantially influencing the computed displacement values and representing the dominant source of measurement uncertainty. The influence of these distinct error types becomes particularly apparent in the displacement evaluation, where the stability of positional determination and the accuracy of scale calibration jointly define the achievable measurement precision. This effect becomes evident in the displacement results. Rectangle-based methods (dispr) show the best correlation with the reference system, achieving deviations within ±10 μ m and extended uncertainties below 1 μ m . Given a maximum reference displacement of 7.5 mm, this corresponds to a relative deviation of less than ±0.15%, which is highly promising for such a minimal implementation. The high accuracy is primarily attributed to the geometric stability of double-symmetrical shapes and the reduced centroid sensitivity to filtering effects. In contrast, triangle- and circle-based measurements exhibit scale-dependent deviations arising from slight inconsistencies in image scale determination.
The observed symmetry of the deviation curves indicates that most systematic errors are linked to the scaling factor rather than to random noise. The sign reversal of deviations between positive and negative displacements reflects a slight underestimation or overestimation of the scale factor depending on displacement direction. Asymmetrical effects—such as in disptr—are likely caused by lighting variations, partial shading on the printed MMs, or minimal angular misalignments in the setup.
Although the camera was not calibrated, the results were largely consistent with the expected trends. However, the absence of calibration could introduce small errors in determining image scale and position. Accurately determining intrinsic and extrinsic camera parameters is critical for making reliable spatial measurements in scenarios involving multiple motives across the measurement plane or more complex geometries [63,64]. Lens distortions, including radial and tangential components, can generate systematic deviations in measured displacements, which can be mitigated through proper calibration [65].
It must be emphasized that all results were obtained under controlled laboratory conditions. Under real-world conditions, numerous environmental and optical influences—such as variable lighting, direct sunlight reflections, shadows, air shimmer due to local temperature gradients, fog, increased humidity, and airborne particles—can significantly affect optical measurements. These influences are not specific to this system but affect all optical SHM approaches [4,10]. Future research should focus on detecting and compensating for these effects, possibly through adaptive preprocessing, calibration routines, or machine learning–based correction algorithms.
Beyond field applications, the approach also offers high potential for laboratory use. Due to its precision, modularity, and open-source implementation, it is particularly suited for experimental investigations and validation tasks. The simplicity of the setup combined with the low cost of standard cameras and Python-based processing makes it ideal for cost-sensitive research environments where high spatial accuracy and scalability are required.
The accuracy of the displacement results is also influenced by the distance between the camera and the measurement motives. A larger distance generally leads to a reduction in precision due to the lower resolution of the captured image and the effects of angular distortion [11,25]. Furthermore, the size of the measurement motives also plays a crucial role in the overall measurement accuracy. Larger measurement motives typically provide higher precision since they offer more visual information to detect and track, thus reducing relative errors in position determination [18,28].
The camera fixture and its stability during measurements are another critical factor. Small shifts or rotations of the camera during the measurement process can lead to substantial measurement errors due to the geometric distortions that arise from even slight changes in angle or distance [17]. Additionally, vibrations and other mechanical disturbances can induce temporary errors, which would compromise measurement accuracy. Therefore, ensuring the stability of the camera setup is essential for continuous, reliable measurements. This highlights the need for secure mounting and vibration dampening when conducting measurements in real-world conditions, where environmental factors such as building movements or traffic can induce micro-movements in the measurement system [4]. Thus, it would be important to consider the setup’s stability when planning a measurement campaign.
In addition to mechanical stability, the geometric alignment between the camera and the measurement plane is a crucial factor for ensuring measurement accuracy. In the present minimal implementation, the measurement plane must be positioned approximately perpendicular to the camera’s optical axis to avoid perspective-induced distortions. Any deviation from this orthogonal alignment leads to non-uniform image scaling and apparent shape deformation, which can affect both position and scale estimations. To compensate for such misalignments, corrective procedures can be applied, for instance, by comparing the apparent sizes of multiple MMs of identical geometry across the field of view. From these relative differences, an alignment correction factor can be derived. More advanced strategies may include reciprocal weighting of the individual elements or geometric calibration based on known reference distances, which would enable accurate displacement estimation even under non-ideal viewing conditions.
The experimental results further confirm that the accuracy of the proposed measurement approach is primarily determined by the robustness of image scale estimation, the symmetry of the measurement motive, and the consistency of preprocessing filters. However, for long-term SHM applications, it is essential to address potential sources of error and investigate solutions for compensating for environmental influences, camera fixture instability, and other real-world challenges. These findings form the basis for further improvements and extensions, such as the integration of strain measurements and multi-plane displacement measurements, which could significantly enhance the approach’s applicability and versatility.

6. Conclusions

This study presented the development and validation of a simple yet precise image-based algorithm for two-dimensional displacement measurement using standard cameras and open-source software. The algorithm combines CNN-based object detection with classical image processing methods for contour and centroid determination, enabling automated, cost-efficient, and scalable displacement monitoring.
Even in its minimal implementation, the system achieved high accuracy. The configuration using double-symmetrical rectangular motives (dispr) produced the most accurate and stable results, with maximum deviations below ±10 μ m at a reference displacement of 7.5 mm, corresponding to a relative deviation below ±0.15%. This demonstrates the method’s high potential for precise and repeatable displacement detection without specialized hardware or proprietary software.
The investigations further showed that the geometric design of the measurement motives significantly affects accuracy. Double-symmetrical geometries such as rectangles proved particularly robust against edge-related information losses during filtering, whereas triangular and mixed configurations exhibited systematic scaling effects. The use of open-source Python-based libraries enables full transparency, adaptability, and scalability of the algorithm for diverse monitoring applications.
The cost-efficiency of the presented approach stems from its use of generic imaging hardware. Displacement measurements can be performed with consumer-grade or low-cost industrial cameras, and existing camera installations can often be repurposed without additional investment. Commercial displacement sensors, such as capacitive, eddy-current, and LVDT transducers, typically range from several hundred to several thousand dollars per unit [66], while proprietary vision-based systems remain similarly expensive due to specialized optics, controlled illumination, and closed processing hardware. Instead, the proposed method relies on widely available camera-lens combinations and software-based post-processing to tailor the optical configuration to measurement requirements. Its open-source, programmable software environment reduces development effort and enhances adaptability. This is particularly advantageous for multi-point or area-based measurements, as expanding the monitored region mostly requires adding more visual measurement motives within the camera’s field of view, rather than installing additional sensors.
Overall, the findings highlight that low-cost, image-based measurement systems can provide accuracy and reliability comparable to more complex sensor systems. Such methods represent a cost-effective, minimally invasive, and scalable alternative for SHM [4,10,21], while also offering value for laboratory applications requiring flexible experimental setups and reproducible results.
Although the results demonstrate significant potential, the study remains limited to controlled laboratory conditions and a single planar measurement configuration. The quantified effects of contour smoothing and centroid deviations are specific to this imaging setup and may differ with other optical systems or motive designs. Additionally, only translational displacements were considered; rotations, perspective distortions, and long-term stability were not addressed. While these limitations do not undermine the validity of the approach, they highlight key considerations for broader structural health monitoring applications. Future work should therefore focus on field implementation, including robust compensation for environmental and optical effects such as variable illumination, air turbulence, and humidity fluctuations. Furthermore, the algorithm should be expanded to facilitate multi-point strain analysis and displacement evaluation across multiple planes to enhance its practical applicability and versatility.

Author Contributions

Conceptualization, D.M. and M.W.; methodology, D.M. and M.W.; software, D.M.; validation, D.M. and M.W.; formal analysis, D.M.; investigation, D.M.; data curation, D.M.; writing—original draft preparation, D.M. and M.W.; writing—review and editing, D.M. and M.W.; visualization, D.M.; supervision, M.W. and K.H.; project administration, K.H. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge the support of the Open Access Publication Fund at Leipzig University of Applied Sciences (HTWK Leipzig).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Nithin, T.A. Real-time structural health monitoring: An innovative approach to ensuring the durability and safety of structures. I-Manag. J. Struct. Eng. 2022, 11, 43. [Google Scholar] [CrossRef]
  2. Brownjohn, J. Structural health monitoring of civil infrastructure. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2007, 365, 589–622. [Google Scholar] [CrossRef]
  3. Aktan, E.; Bartoli, I.; Glišić, B.; Rainieri, C. Lessons from Bridge Structural Health Monitoring (SHM) and Their Implications for the Development of Cyber-Physical Systems. Infrastructures 2024, 9, 30. [Google Scholar] [CrossRef]
  4. Farrar, C.R.; Worden, K. An introduction to structural health monitoring. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2007, 365, 303–315. [Google Scholar] [CrossRef]
  5. Rabi, R.R.; Vailati, M.; Monti, G. Effectiveness of Vibration-Based Techniques for Damage Localization and Lifetime Prediction in Structural Health Monitoring of Bridges: A Comprehensive Review. Buildings 2024, 14, 1183. [Google Scholar] [CrossRef]
  6. Anjum, A.; Hrairi, M.; Aabid, A.; Yatim, N.; Ali, M. Civil Structural Health Monitoring and Machine Learning: A Comprehensive Review. Fract. Struct. Integr. 2024, 18, 43–59. [Google Scholar] [CrossRef]
  7. Laflamme, S.; Ubertini, F.; Di Matteo, A.; Pirrotta, A.; Perry, M.; Fu, Y.; Li, J.; Wang, H.; Hoang, T.; Glisic, B.; et al. Roadmap on measurement technologies for next generation structural health monitoring systems. Meas. Sci. Technol. 2023, 34, 093001. [Google Scholar] [CrossRef]
  8. Gharehbaghi, V.R.; Noroozinejad Farsangi, E.; Noori, M.; Yang, T.Y.; Li, S.; Nguyen, A.; Málaga-Chuquitaype, C.; Gardoni, P.; Mirjalili, S. A Critical Review on Structural Health Monitoring: Definitions, Methods, and Perspectives. Arch. Comput. Methods Eng. 2022, 29, 2209–2235. [Google Scholar] [CrossRef]
  9. Hassani, S.; Dackermann, U. A Systematic Review of Optimization Algorithms for Structural Health Monitoring and Optimal Sensor Placement. Sensors 2023, 23, 3293. [Google Scholar] [CrossRef]
  10. Kot, P.; Muradov, M.; Gkantou, M.; Kamaris, G.S.; Hashim, K.; Yeboah, D. Recent Advancements in Non-Destructive Testing Techniques for Structural Health Monitoring. Appl. Sci. 2021, 11, 2750. [Google Scholar] [CrossRef]
  11. Palma, P.; Steiger, R. Structural health monitoring of timber structures–Review of available methods and case studies. Constr. Build. Mater. 2020, 248, 118528. [Google Scholar] [CrossRef]
  12. Dos Reis, J.; Oliveira Costa, C.; Sá da Costa, J. Strain gauges debonding fault detection for structural health monitoring. Struct. Control Health Monit. 2018, 25, e2264. [Google Scholar] [CrossRef]
  13. Bolandi, H.; Lajnef, N.; Jiao, P.; Barri, K.; Hasni, H.; Alavi, A.H. A Novel Data Reduction Approach for Structural Health Monitoring Systems. Sensors 2019, 19, 4823. [Google Scholar] [CrossRef] [PubMed]
  14. Weisbrich, M.; Holschemacher, K.; Bier, T. Comparison of different fiber coatings for distributed strain measurement in cementitious matrices. J. Sens. Sens. Syst. 2020, 9, 189–197. [Google Scholar] [CrossRef]
  15. Scuro, C.; Lamonaca, F.; Porzio, S.; Milani, G.; Olivito, R. Internet of Things (IoT) for masonry structural health monitoring (SHM): Overview and examples of innovative systems. Constr. Build. Mater. 2021, 290, 123092. [Google Scholar] [CrossRef]
  16. Kim, J.W.; Choi, H.W.; Kim, S.K.; Na, W.S. Review of Image-Processing-Based Technology for Structural Health Monitoring of Civil Infrastructures. J. Imaging 2024, 10, 93. [Google Scholar] [CrossRef]
  17. Dong, C.Z.; Catbas, F.N. A review of computer vision–based structural health monitoring at local and global levels. Struct. Health Monit. 2021, 20, 692–743. [Google Scholar] [CrossRef]
  18. Boursier Niutta, C.; Tridello, A.; Ciardiello, R.; Paolino, D.S. Strain Measurement with Optic Fibers for Structural Health Monitoring of Woven Composites: Comparison with Strain Gauges and Digital Image Correlation Measurements. Sensors 2023, 23, 9794. [Google Scholar] [CrossRef]
  19. Morgenthal, G.; Eick, J.F.; Rau, S.; Taraben, J. Wireless Sensor Networks Composed of Standard Microcomputers and Smartphones for Applications in Structural Health Monitoring. Sensors 2019, 19, 2070. [Google Scholar] [CrossRef] [PubMed]
  20. Mustapha, S.; Lu, Y.; Ng, C.T.; Malinowski, P. Sensor Networks for Structures Health Monitoring: Placement, Implementations, and Challenges—A Review. Vibration 2021, 4, 551–585. [Google Scholar] [CrossRef]
  21. Malik, H.; Khattak, K.S.; Wiqar, T.; Khan, Z.H.; Altamimi, A.B. Low Cost Internet of Things Platform for Structural Health Monitoring. In Proceedings of the 2019 22nd International Multitopic Conference (INMIC), Islamabad, Pakistan, 29–30 November 2019. [Google Scholar] [CrossRef]
  22. Kralovec, C.; Schagerl, M. Review of Structural Health Monitoring Methods Regarding a Multi-Sensor Approach for Damage Assessment of Metal and Composite Structures. Sensors 2020, 20, 826. [Google Scholar] [CrossRef] [PubMed]
  23. Sony, S.; Laventure, S.; Sadhu, A. A literature review of next-generation smart sensing technology in structural health monitoring. Struct. Control Health Monit. 2019, 26, e2321. [Google Scholar] [CrossRef]
  24. Feng, D.; Feng, M.; Ozer, E.; Fukuda, Y. A Vision-Based Sensor for Noncontact Structural Displacement Measurement. Sensors 2015, 15, 16557–16575. [Google Scholar] [CrossRef]
  25. Henke, K.; Pawlowski, R.; Schregle, P.; Winter, S. Use of digital image processing in the monitoring of deformations in building structures. J. Civ. Struct. Health Monit. 2015, 5, 141–152. [Google Scholar] [CrossRef]
  26. Ye, X.W.; Dong, C.Z.; Liu, T. A Review of Machine Vision-Based Structural Health Monitoring: Methodologies and Applications. J. Sens. 2016, 2016, 7103039. [Google Scholar] [CrossRef]
  27. Feng, D.; Feng, M.Q. Experimental validation of cost-effective vision-based structural health monitoring. Mech. Syst. Signal Process. 2017, 88, 199–211. [Google Scholar] [CrossRef]
  28. Chen, B.; Tomizuka, M. OpenSHM: Open Architecture Design of Structural Health Monitoring Software in Wireless Sensor Nodes. In Proceedings of the 2008 IEEE/ASME International Conference on Mechtronic and Embedded Systems and Applications, Beijing, China, 12–15 October 2008; pp. 19–24. [Google Scholar] [CrossRef]
  29. Basto, C.; Pelà, L.; Chacón, R. Open-source digital technologies for low-cost monitoring of historical constructions. J. Cult. Herit. 2017, 25, 31–40. [Google Scholar] [CrossRef]
  30. Xu, Q.; Wang, X.; Yan, F.; Zeng, Z. Non-Contact Extensometer Deformation Detection via Deep Learning and Edge Feature Analysis. In Design Studies and Intelligence Engineering; IOS Press: Amsterdam, The Netherlands, 2023. [Google Scholar] [CrossRef]
  31. Xu, Y.; Brownjohn, J.M.W. Review of machine-vision based methodologies for displacement measurement in civil structures. J. Civ. Struct. Health Monit. 2018, 8, 91–110. [Google Scholar] [CrossRef]
  32. Zhuang, Y.; Chen, W.; Jin, T.; Chen, B.; Zhang, H.; Zhang, W. A Review of Computer Vision-Based Structural Deformation Monitoring in Field Environments. Sensors 2022, 22, 3789. [Google Scholar] [CrossRef] [PubMed]
  33. Lee, J.; Lee, K.C.; Cho, S.; Sim, S.H. Computer Vision-Based Structural Displacement Measurement Robust to Light-Induced Image Degradation for In-Service Bridges. Sensors 2017, 17, 2317. [Google Scholar] [CrossRef] [PubMed]
  34. Ferraris, C.; Amprimo, G.; Pettiti, G. Computer Vision and Image Processing in Structural Health Monitoring: Overview of Recent Applications. Signals 2023, 4, 539–574. [Google Scholar] [CrossRef]
  35. Peters, J.F. Foundations of Computer Vision; Springer: Cham, Switzerland, 2017. [Google Scholar] [CrossRef]
  36. Pagire, V.; Chavali, M.; Kale, A. A comprehensive review of object detection with traditional and deep learning methods. Signal Process. 2025, 237, 110075. [Google Scholar] [CrossRef]
  37. Edozie, E.; Shuaibu, A.N.; John, U.K.; Sadiq, B.O. Comprehensive review of recent developments in visual object detection based on deep learning. Artif. Intell. Rev. 2025, 58, 277. [Google Scholar] [CrossRef]
  38. Sun, Y.; Sun, Z.; Chen, W. The evolution of object detection methods. Eng. Appl. Artif. Intell. 2024, 133, 108458. [Google Scholar] [CrossRef]
  39. Ahmed, M.; Hashmi, K.A.; Pagani, A.; Liwicki, M.; Stricker, D.; Afzal, M.Z. Survey and Performance Analysis of Deep Learning Based Object Detection in Challenging Environments. Sensors 2021, 21, 5116. [Google Scholar] [CrossRef]
  40. Bilous, N.; Malko, V.; Frohme, M.; Nechyporenko, A. Comparison of CNN-Based Architectures for Detection of Different Object Classes. AI 2024, 5, 2300–2320. [Google Scholar] [CrossRef]
  41. Aboyomi, D.D.; Daniel, C. A Comparative Analysis of Modern Object Detection Algorithms: YOLO vs. SSD vs. Faster R-CNN. ITEJ (Inf. Technol. Eng. J.) 2023, 8, 96–106. [Google Scholar] [CrossRef]
  42. Kumari, R.; Chandra, D. Real-time Comparison of Performance Analysis of Various Edge Detection Techniques Based on Imagery Data. Curr. J. Appl. Sci. Technol. 2023, 42, 22–31. [Google Scholar] [CrossRef]
  43. Li, P.; Wang, H.; Yu, M.; Li, Y. Overview of Image Smoothing Algorithms. J. Phys. Conf. Ser. 2021, 1883, 012024. [Google Scholar] [CrossRef]
  44. Chen, B.H.; Tseng, Y.S.; Yin, J.L. Gaussian-Adaptive Bilateral Filter. IEEE Signal Process. Lett. 2020, 27, 1670–1674. [Google Scholar] [CrossRef]
  45. Shreyamsha Kumar, B.K. Image denoising based on gaussian/bilateral filter and its method noise thresholding. Signal Image Video Process. 2013, 7, 1159–1172. [Google Scholar] [CrossRef]
  46. Das, S. Comparison of Various Edge Detection Technique. Int. J. Signal Process. Image Process. Pattern Recognit. 2016, 9, 143–158. [Google Scholar] [CrossRef]
  47. Magnier, B.; Abdulrahman, H.; Montesinos, P. A Review of Supervised Edge Detection Evaluation Methods and an Objective Comparison of Filtering Gradient Computations Using Hysteresis Thresholds. J. Imaging 2018, 4, 74. [Google Scholar] [CrossRef]
  48. Yang, D.; Peng, B.; Al-Huda, Z.; Malik, A.; Zhai, D. An overview of edge and object contour detection. Neurocomputing 2022, 488, 470–493. [Google Scholar] [CrossRef]
  49. Papari, G.; Petkov, N. Edge and line oriented contour detection: State of the art. Image Vis. Comput. 2011, 29, 79–103. [Google Scholar] [CrossRef]
  50. Harris, C.; Stephens, M. A Combined Corner and Edge Detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988; pp. 23.1–23.6. [Google Scholar] [CrossRef]
  51. Sánchez, J.; Monzón, N.; Salgado, A. An Analysis and Implementation of the Harris Corner Detector. Image Process. Line 2018, 8, 305–328. [Google Scholar] [CrossRef]
  52. Shi, J.; Tomasi. Good features to track. In Proceedings of the 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar] [CrossRef]
  53. Förstner, W.; Gülch, E. A Fast Operator for Detection and Precise Location of Distict Point, Corners and Centres of Circular Features. In Proceedings of the ISPRS Intercommission Conference on Fast Processing of Photogrammetric Data, Interlaken, Switzerland, 2–4 June 1987. [Google Scholar]
  54. Illingworth, J.; Kittler, J. The Adaptive Hough Transform. IEEE Trans. Pattern Anal. Mach. Intell. 1987, PAMI-9, 690–698. [Google Scholar] [CrossRef]
  55. Lava, P.; Van Paepegem, W.; Coppieters, S.; De Baere, I.; Wang, Y.; Debruyne, D. Impact of lens distortions on strain measurements obtained with 2D digital image correlation. Opt. Lasers Eng. 2013, 51, 576–584. [Google Scholar] [CrossRef]
  56. Suzuki, S.; be, K. Topological structural analysis of digitized binary images by border following. Comput. Vis. Graph. Image Process. 1985, 30, 32–46. [Google Scholar] [CrossRef]
  57. Montero, R.S.; Bribiesca, E. State of the Art of Compactness and Circularity Measures. Int. Math. Forum 2009, 4, 1305–1335. [Google Scholar]
  58. Ramer, U. An iterative procedure for the polygonal approximation of plane curves. Comput. Graph. Image Process. 1972, 1, 244–256. [Google Scholar] [CrossRef]
  59. Douglas, D.H.; Peucker, T.K. Algorithms for the Reduction of the Number of Points Required to Represent a Digitized Line or its Caricature. In Classics in Cartography: Reflections on Influential Articles from Cartographica; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2011; pp. 15–28. [Google Scholar] [CrossRef]
  60. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  61. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2009, PAMI-8, 679–698. [Google Scholar] [CrossRef]
  62. Rosenfeld, A.; Kak, A.C. Digital Picture Processing, 2nd ed.; Elsevier Science: Saint Louis, MO, USA, 2014; Volume 1. [Google Scholar]
  63. Xiong, P.; Wang, S.; Wang, W.; Ye, Q.; Ye, S. Model-Independent Lens Distortion Correction Based on Sub-Pixel Phase Encoding. Sensors 2021, 21, 7465. [Google Scholar] [CrossRef]
  64. Huai, J.; Shao, Y.; Jozkow, G.; Wang, B.; Chen, D.; He, Y.; Yilmaz, A. Geometric Wide-Angle Camera Calibration: A Review and Comparative Study. Sensors 2024, 24, 6595. [Google Scholar] [CrossRef]
  65. Pan, B.; Yu, L.; Wu, D.; Tang, L. Systematic errors in two-dimensional digital image correlation due to lens distortion. Opt. Lasers Eng. 2013, 51, 140–147. [Google Scholar] [CrossRef]
  66. Tsao, C.C.; Tseng, Y.C.; Chen, Y.S.; Chang, W.H.; Huo, L.T. Precision Low-Cost Compact Micro-Displacement Sensors Based on a New Arrangement of Cascaded Levers with Flexural Joints. Sensors 2022, 23, 326. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.