Next Article in Journal
Versatile Multimodality Imaging System Based on Detectorless and Scanless Optical Feedback Interferometry—A Retrospective Overview for A Prospective Vision
Previous Article in Journal
Overview of Time Synchronization for IoT Deployments: Clock Discipline Algorithms and Protocols
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Vision Method for Robust Monitoring of Multi-Point Dynamic Displacements with Smartphones in an Interference Environment

1
School of Civil Engineering and Transportation, South China University of Technology, Guangzhou 510641, China
2
State Key Laboratory of Subtropical Building Science, South China University of Technology, Guangzhou 510641, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(20), 5929; https://doi.org/10.3390/s20205929
Submission received: 4 August 2020 / Revised: 10 October 2020 / Accepted: 16 October 2020 / Published: 20 October 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Current research on dynamic displacement measurement based on computer vision mostly requires professional high-speed cameras and an ideal shooting environment to ensure the performance and accuracy of the analysis. However, the high cost of the camera and strict requirements of sharp image contrast and stable environment during the shooting process limit the broad application of the technology. This paper proposes an improved vision method to implement multi-point dynamic displacement measurements with smartphones in an interference environment. A motion-enhanced spatio-temporal context (MSTC) algorithm is developed and applied together with the optical flow (OF) algorithm to realize a simultaneous tracking and dynamic displacement extraction of multiple points on a vibrating structure in the interference environment. Finally, a sine-sweep vibration experiment on a cantilever sphere model is presented to validate the feasibility of the proposed method in a wide-band frequency range. In the test, a smartphone was used to shoot the vibration process of the sine-sweep-excited sphere, and illumination change, fog interference, and camera jitter were artificially simulated to represent the interference environment. The results of the proposed method are compared to conventional displacement sensor data and current vision method results. It is demonstrated that, in an interference environment, (1) the OF method is prone to mismatch the feature points and leads to data deviated or lost; (2) the conventional STC method is sensitive to target selection and can effectively track those targets having a large proportion of pixels in the context with motion tendency similar to the target center; (3) the proposed MSTC method, however, can ease the sensitivity to target selection through in-depth processing of the information in the context and finally enhance the robustness of the target tracking. In addition, the MSTC method takes less than one second to track each target between adjacent frame images, implying a potential for online measurement.

1. Introduction

Structural health monitoring, as a crucial technique ensuring long-term structural safety, plays an essential role in early warning, evaluation, and maintenance decisions during the structural life cycle. Accelerometers, one kind of data sensors widely used in structural health monitoring practices, can directly measure the acceleration response of structures while less effective in the low-frequency range [1]. In contrast, displacement sensors can obtain sensitive low-frequency dynamic measurements but see relatively few applications due to the inconvenience of installation and monitoring. For instance, a linear variable differential transformer (LVDT) requires a fixed reference point to measure the displacement, and the laser Doppler vibrometer needs installing close to the object.
In recent years, with the development of image data acquisition and computer vision techniques, vision-based displacement measurement technology has attracted extensive research attention due to its advantages of no contact, long distance, and low cost [2,3]. According to target markers required or not, the computer vision technologies for displacement measurement can be classified into two categories. The first category using customized target markers [4,5,6,7,8,9,10,11] can increase the differentiation between the tracking target and the background to improve measurement accuracy. In the actual measurement, however, the installation of markers is labor-intensive, especially when multi-point displacements are desired.
In contrast, the non-target measurement category uses the texture features of the structure surface as pseudo markers. Various methods have been developed to track the features points and identify the corresponding displacements. For instance, with feature-point matching algorithms, Khuc [12] and Dong [13] adopted professional high-speed cameras to achieve displacement measurement without attaching any artificial target markers. This kind of algorithms mostly needs a clear texture of the object to reduce the false recognition of features. Moreover, the use of high-speed cameras also requires high-brightness lighting conditions to achieve a sharp contrast of imaging objects. Yoon [14] applied an optical flow algorithm and adopted a consumer-grade camera to extract the displacement response and dynamic characteristics. Feng [15] proposed a sub-pixel template matching technique based on image correlation, improving the accuracy of displacement extraction. It is worth noting that all the above studies are applied under ideal measurement conditions and require a stable lighting environment. The influence of unfavorable environmental factors such as illumination change, fog interference, camera motion, and complex background on the measurement results is not involved. Regarding this, Dong [16] applied a spatio-temporal context (STC) algorithm to implement the displacement measurement with a high-speed camera in the illumination change and fog environment. The research realized single-point dynamic displacement measurement but not multi-point synchronous monitoring that is more favorable in structural health monitoring practice. Furthermore, the applied STC algorithm treats the whole region of the context equally, which weakens the effectiveness of the context information. As for the camera motion problem, which is getting increasing attention in vision measurement practices, a general solution is to compensate for camera motion-induced errors by subtracting the movement of reference points on the background [17,18]. The methods applied therein to identify the reference-point movement are critical to the measurement accuracy, especially when camera jitter, one particular kind of camera motion undergoing dramatic change and hence increasing challenges, is taken into account.
The present study proposed an improved displacement measurement method with computer vision technology, which allows the use of smartphone-recording video to achieve simultaneous monitoring of multi-point dynamic displacements under changing environmental conditions. It is noted that current smartphones with high-resolution cameras preserve many advantages, such as low cost, portability, and ease of use, making it a competitive device for optical sensing under various conditions [19]. In the proposed method, after calibration of the sequential distorted images shot from the smartphone, multiple areas centered at the targeting points are predefined with permissible overlapping. A motion-enhanced spatio-temporal context (MSTC) algorithm is then developed for robust tracking of the moving targets. Therein, a motion-enhanced function is constructed with an optical flow (OF) algorithm to strengthen the beneficial information and weaken the interference information in the whole context area. Finally, based on the tracked pixel-level results from the MSTC algorithm, the OF algorithm is re-applied to estimate the subpixel-level moving, benefiting eventually to the achievement of high-precision displacement results. To validate the feasibility of the proposed method for monitoring vibrating displacements in a wide frequency range, a sine-sweep vibration experiment on a cantilever sphere model is carried out. In the test, an iPhone 7 was used to shoot the vibration process of the sine-sweep-excited sphere, and illumination change, fog interference, and camera jitter were artificially simulated to represent the interference environment. By referring to the displacement sensor data, comparative studies between results from the proposed method and current vision methods were presented.

2. Integrated Process for Multi-Point Displacement Measurement

Implementation of the proposed measurement method focuses on three aspects: smartphone application, multi-point monitoring, and high-precision displacement extraction.
The first aspect needs to solve the calibration of the wide-angle distortion from the consumer-grade camera of the smartphone. Therein, Zhang’s method [20] is employed to implement the camera calibration, and we further propose an iterative calibration procedure based on a continually expanding image library to achieve calibrated parameters converged.
The second aspect concerns the issue of multi-target selection and definition. Therein, through introducing multi-region compatibility and parallel design, we extend the range of the STC algorithm from conventional single-target tracking to multi-target tracking.
The last aspect deals with the algorithms about multi-target tracking and high-precision displacement extraction under environmental interference, which is the most crucial part of the integrated process. Therein, we develop a new tracking algorithm, i.e., the MSTC algorithm, to implement robust tracking of multiple objects. A sub-pixel displacement estimation method [21] is then employed to refine the measurement.
Figure 1 shows the flowchart of the monitoring process, in which three main parts are integrated with sequence and will be illustrated in detail in the following.

3. Smartphone-Shoot Image Calibration

In most smartphones, the consumer-grade cameras are equipped with a wide-angle lens to increase the shooting range, while it will introduce large radial and tangential distortion to the image. Therefore, to reduce the position error caused by the image distortion, it is necessary to calibrate the camera and identify the internal/external parameters and distortion coefficients.
According to the pinhole model used for camera imaging [20], the relationship between a real point A (X, Y, Z) and the corresponding point A’(u, v) in the camera image can be described as
s x y 1 = X Y Z 1 R t
u v 1 = x y 1 K
where (X, Y, Z), (x, y), and (u, v) represent the coordinates of point A in the real-world coordinate system, camera coordinate system, and pixel coordinate system, respectively; s is a scaling factor; the external parameter matrices R and t, respectively, represent the rotation matrix and the translation matrix, related to the physical position of the camera; the internal parameter matrix K is determined by the camera structure and material, represented as
K =   f x 0 0 θ f y 0 u 0 v 0 1
where f x and f y are scale factors of the x and y axes in the pixel coordinate system; θ is the inclination parameter of the axis; u 0 and v 0 are coordinates of the origin.
Equations (1) and (2) represent an ideal imaging relationship. However, the camera lens produces radial and tangential distortions, causing the imaging point to shift. A nonlinear distortion model [22] can be used to describe the distortions as
x e y e = 1 + k 1 r 2 + k 2 r 4 x d y d   + 2 p 1 x d y d + p 2 r 2 + 2 x d 2 2 p 2 x d y d + p 1 r 2 + 2 y d 2
in which ( x e , y e ) and ( x d , y d ) are the ideal exact coordinates and actual distorted coordinates of the target point in the camera coordinate system, respectively; k 1 and k 2 are radial distortion coefficients; p 1 and p 2 are tangential distortion coefficients; r is a radius defined by
r 2 = x d 2 + y d 2
Identification of internal/external parameters and distortion coefficients, namely, camera calibration, is based on the accurate coordinates (X, Y, Z) of several predefined points and the distorted pixel coordinates ( x d , y d ) of the corresponding imaging points. Specific values of all the parameters and coefficients can be solved by minimizing the sum of square errors between the distorted pixel coordinates ( x d , y d ) and the pixel coordinates projected from (X, Y, Z) with Equations (1)–(5).
In the present study, we use the method proposed in Reference [20] to implement the camera calibration. Firstly, an image database is established by taking photos at a predefined chessboard with the smartphone, an iPhone 7, of different locations and postures. Then a series of calibration analyses adopting various numbers of images are carried out until a converged result is reached, as listed in Table 1 and Table 2 for final results of all the parameters and coefficients of the iPhone 7 rear camera. Figure 2 shows the variation of identified coefficients k 1 and k 2 versus the number of images used. It can be observed that nearly 20 images are sufficient to meet a convergence requirement for the camera calibration.
After the camera calibration, we apply Equation (4) for every interested image to adjust the distorted pixel coordinates ( x d , y d ) to a more rational one x c , y c , which is finally used in Equations (1) and (2) to get the real-world coordinates. Figure 3 shows a comparison of a partial image of the chessboard before and after calibration. It can be seen in the original image that, due to radial and tangential distortions, vertical edges of black and white squares are projected as curves and fail to form a straight line, especially for the most marginal black square in the bottom left corner. At the same time, apparent zigzag happens on horizontal edges. In the calibrated image, the distorted vertical edges have been recovered nearly straight, and the zigzagged horizontal edges have been partly smoothed. The calibration, although not 100% recovery, still favors in an error reduction of the final displacement.

4. Multi-Target Definition

In the structural vibration practice, multi-point responses are substantially desired for a comprehensive understanding of vibration and consequent modal analysis. By use of video-based methods to achieve multi-point displacement measurement, one needs to make a clear definition of all the points and track their positions in each frame image of the video.
The STC method used in the present study was initially developed to monitor a single-target movement [16,23]. Herein, we extend its application range to the multi-target case by introducing the regional compatibility of multiple target areas into the parallel design of the target-tracking algorithm. The definition of three target areas in the first frame image is shown in Figure 4. Each area is of a rectangle shape and centers at a predefined target point of cross mark. It is worth noting that the target areas allow partially overlapped, and the number of targets is indeed unlimited. Movements of the corresponding target areas then represent those of the target points.

5. Target Tracking Based on MSTC Algorithm

5.1. Conventional STC Tracking Algorithm

The STC algorithm formulates the spatio-temporal statistical correlation between the image intensity from the object of interest and its local context and obtains the best target position by maximizing the following confidence function [23],
c x = P x | o = a e x x α β
where x R 2 is the target position; o denotes the target present in the scene; x is the center point of the target area; a is a normalization constant; α and β are a scale parameter and a shape parameter, respectively.
In a current frame image, the context feature set is defined as,
X c = c z = I z , z | z Ω C x
where I z denotes image intensity at location z ; Ω C x is the context area around the target. As shown in Figure 5, the yellow box defines the tracking target, and the red box represents its local context.
According to the total probability formula, the confidence function in Equation (6) can be computed by
c x = P x | o = c z X c P x , c z | o = c z X c P x | c z , o P c z | o
where, P x , c z | o is the joint probability; P x | c z , o is the conditional probability, modeling the spatial relationship between the target and its context;   P c z | o is the context prior probability, modeling the appearance characteristics of the context. The main task in STC is to learn P x | c z , o as bridging the target position and its spatial context.
With the intensity of the current frame image, the context prior probability P c z | o is simply modeled as
P c z | o = I z ω σ z x
where
ω σ z x =   b e z x 2 σ 2
σ = s w + s h / 2
in which, b is a normalization constant; σ is a scale parameter; s w and s h are the length and width of the target area, respectively.
The conditional probability P x | c z , o is modeled to be a function of the relative distance and direction between the target and its context,
P x | c z , o = h x z
Substituting Equations (9) and (12) into Equation (8), we can get
c x = c z X c h x z I z ω σ z x = h x I x ω σ x x
where denotes the convolution operator. For the sake of efficient operation, Equation (13) is transformed into the frequency domain so that the Fast Fourier Transform (FFT) algorithm can be used for fast convolution, leading to
    F c x = F h x F I x ω σ x x  
where F denotes the FFT operation and represents the pixel-wise product. Therefore, we have
h x = F 1 F c x F I x ω σ x x = F 1 F a e x x α β F I x ω σ x x  
where F 1   denotes the inverse FFT operation.
After h x is obtained, the tracking problem with STC is simply formulated as a detection task. In the (t + 1)-th frame image, the object position x t + 1 is determined by maximizing the new confidence function
x t + 1 = a r g m a x x Ω C x t   c t + 1 x
where c t + 1 x is represented as
c t + 1 x = F 1 F H t + 1 x F I t + 1 x ω σ t x x
in which H t + 1 x , a replacement of h x in Equation (14) with additional temporal information, is updated by
H t + 1 x   = 1 ρ H t x + ρ h t x
where ρ is a learning rate factor.

5.2. Motion-Enhanced STC (MSTC) Tracking Algorithm

5.2.1. Basic Idea

With the spatio-temporal context model in Equation (18), the conventional STC algorithm can filter out the image noise caused by appearance variations, thereby leading to stable tracking results. However, the STC algorithm treats the entire context equivalently by taking it for granted that all regions of the context have the same contribution in tracking the target movement, leading to ineffective outcomes in a severe noise environment. Regarding this shortcoming, Xu [24] proposed a weighted STC algorithm by applying Harris corner displacement identification and linear interpolation to construct a weight function. However, the Harris corner tracking may fail in an interference environment, as shown in Section 7, and the linear interpolation with a few feature points can only yield an approximate contribution of the context.
We notice in practical videos that pixels on the same moving object tend to have a more similar motion than the other pixels, implying different pixels in the context may have different correlations with the target in terms of motion tendency. An example of the pixel-wise motion tendency obtained from the optical flow (OF) algorithm (details in Section 5.2.2) is shown in Figure 6. Regarding the similarity with the target center, all pixels in the context can be roughly grouped into three sets: (1) the auxiliary set, e.g., regions A, E, and G; (2) the irrelevant set, e.g., regions B and F; (3) the interference set, e.g., regions C, D, and H. Furthermore, in line with varying movements in the video process, these three sets will also vary frame to frame during the tracking process. Such new spatio-temporal information, if integrated into the tracking algorithm, will benefit a lot to improve the measurement accuracy in a noisy environment.
Therefore, the basic idea of the MSTC algorithm we develop here is to construct a dense pixel-wise influence matrix by using the motion-tendency similarity between pixels in the context and the target. Pixels in the regions of the auxiliary set will have the most substantial influence on target tracking, as opposed to the interference set. The influence matrix is then integrated into the tracking framework of the STC algorithm to achieve more robust tracking performance.

5.2.2. Derivation of the Context Influence Matrix

The context influence matrix builds a mathematical model based on the motion-tendency similarity between pixels in different regions of the context and the tracking target. Here, we apply the OF approach to identify the motion tendency of all pixels in the context to generate the context influence matrix. The procedure is as follows:
(1)
Get the intensity I t 1 x , y of the (t − 1)-th frame image and I t x , y of the t-th frame image.
(2)
Identify the pixel-wise velocity by use of the least-square OF method [25],
x , y I t x 2   x , y I t x I t y x , y I t x I t y x , y I t y 2   v x t v y t =   x , y I t t I t x x , y I t t I t y
where all the partial derivatives can be estimated by averaging the first-order finite differential values, i.e.,
  I t x 1 4 I t 1 x + 1 , y + I t 1 x + 1 , y + 1 + I t x + 1 , y + I t x + 1 , y + 1 1 4 I t 1 x , y + I t 1 x , y + 1 + I t x , y + I t x , y + 1
I t y 1 4 I t 1 x , y + 1 + I t 1 x + 1 , y + 1 + I t x , y + 1 + I t x + 1 , y + 1 1 4 I t 1 x , y + I t 1 x + 1 , y + I t x , y + I t x + 1 , y .  
  I t t 1 4 I t x , y + I t x + 1 , y + I t x , y + 1 + I t x + 1 , y + 1 1 4 I t 1 x , y + I t 1 x + 1 , y + I t 1 x , y + 1 + I t 1 x + 1 , y + 1
(3)
Calculate the velocity difference between the target center and each pixel in the context,
  d t i = v x t v x i t 2 + v y t v y i t 2
(4)
For each pixel in the context, convert the corresponding velocity difference with a negative exponential function to generate the influence coefficient,
m t i = e d t i 2 2 γ 2
where γ is a scale parameter. As can be seen that more similar the two velocities are, i.e., less velocity difference, the more considerable value the influence coefficient will take.
As shown in Figure 7 is an example of the influence matrix, which we obtained in the laboratory verification for tracking Target Point 2 under fog interference (Scenario III, see details in Section 7). The bright region in the context is almost consistent with the profile of the moving object, implying that a correct relationship is successfully built between the target center and the pixels in the context. Such new information helps to include more contributions from the context into the tracking of the target, especially when severe interference is involved.

5.2.3. Updated Formulation

By combining the influence of the context, we renew the conditional probability as
P x | c z , o = h x z m z
where m z is the dense pixel-wise influence matrix.
Correspondingly, all the related formulas are updated as
c x = c z X c h x z m z I z ω σ z x = h x m x I x ω σ x x
    F c x = F h x F m x I x ω σ x x  
h x = F 1 F c x F m x I x ω σ x x = F 1 F a e x x α β F m x I x ω σ x x   .  

5.2.4. MSTC Tracking Process

After m x is obtained, the confidence function to be maximized during the tracking process now becomes
c t + 1 x = F 1 F H t + 1 x F M t + 1 x I t + 1 x ω σ t x x .  
in which M t + 1 x , a temporal version of m x and similar to H t + 1 x in Equation (18), is updated by
M t + 1 x   = 1 η M t x   +   η m t x
where η is a learning rate factor and the initial matrix M 1 is defined as a regularized matrix with all entries equal to 1.
The integrated MSTC tracking process is illustrated in Figure 8. For the sake of a robust tracking, we recommend several crucial parameters taking values as, α = 2.25, β = 1, ρ = 0.075, γ = 4, η = 0.3.

6. Sub-Pixel Displacement Estimation with OF Algorithm

The target position obtained above is of integer pixel, resulting in a pixel-level displacement. To achieve more accurate movement, besides choosing a higher-resolution camera, one can also use the OF algorithm to estimate the sub-pixel displacement [21].
In general, the following relationship holds between the target intensities in two adjacent frame images, e.g., I t 1 x , y and I t x , y ,
I t 1 x , y = I t x + Δ x , y + Δ y
where the target displacement Δ x ,   Δ y can be expressed as the sum of the pixel-level part Δ x ¯ ,   Δ y ¯ and the sub-pixel part δ x ,   δ y
Δ x = Δ x ¯ + δ x Δ y = Δ y ¯ + δ y
in which δ x ≪1 and δ y ≪1.
After the pixel-level displacement Δ x ¯ ,   Δ y ¯ has been identified by using the MSTC algorithm, we can construct a new template image relating to the pixel-level displacement, then apply the OF algorithm to expand the I t 1 x , y in terms of δ x and δ y ,
I t 1 x , y   =   I t x + Δ x ¯ + δ x , y + Δ y ¯ + δ y I ¯ t x , y + I ¯ t x , y x δ x + I ¯ t x , y y δ y
where I ¯ t x , y = I t x + Δ x ¯ , y + Δ y ¯ represents the intensity of the new template image.
Then, by applying the same least-square OF method as in Section 5.2.2, we can get
x , y I ¯ t x 2   x , y I ¯ t x I ¯ t y x , y I ¯ t x I ¯ t y x , y I ¯ t y 2   δ x δ y =   x , y I t 1 I ¯ t I ¯ t x x , y I t 1 I ¯ t I ¯ t y
where all the partial derivative can be estimated similarly as in Section 5.2.2.
Related research [21] has shown that the accuracy of sub-pixel displacement estimated by the OF algorithm can reach as small as 0.0125 pixel. Therefore, the combination of the MSTC tracking algorithm to identify the pixel-level movement and the OF algorithm to estimate the sub-pixel displacement can effectively improve the measurement precision of the target displacement. It is also worth noting that the same OF algorithm is adopted in the present study to estimate both the pixel-wise motion tendency and the sub-pixel displacement, facilitating the implementation of compact and unified programming.

7. Laboratory Verification

In the following, we will carry out forced vibration experiments of a cantilever sphere model to verify whether the proposed method can accurately obtain the multi-point displacement time history of the model subject to light change, fog interference, and camera jitter. Based on smartphone-shoot videos recording the model vibration process, measurement results with the proposed MSTC method are compared to the displacement sensor data, as well as those obtained with two current vision methods, e.g., the conventional STC method [16] and the characteristic OF method [14].

7.1. Experiment Setup

Figure 9 shows the experiment layout. A steel sphere is installed on top of a vertical steel rod that is bottom-fixed on a rigid support and connected on the middle to a vibration exciter, which will input external excitation to the structural model and force the sphere to vibrate horizontally. A linear variable differential transformer (LVDT), in which one end connected to the sphere and the other end fixed to the left steel column, is used to measure the displacement of the sphere, with data automatically collected by a computer. An iPhone7 smartphone mounted on a tripod in front of the model is used to shoot the sphere vibration video. The resolution of the iPhone7 rear camera is set to 1920 × 1080 pixels and the theoretical frame rate set to 60 frames per second (FPS). The actual frame rate of the video, however, is 59.96 FPS. This subtle FPS difference may cause significant errors to the displacement result, as will be illustrated in detail later. It should also be noted that we deliberately keep the complex background, without cover like usual doing, to approach the real application environment.
In the experiment, for verification in a wide-band frequency range, the exciter is set to produce repetitive sine-sweep excitation with frequency linearly varying from 1 Hz to 10 Hz. We design four scenarios to examine the application effect of the vision methods in different environments,
(1)
Scenario I: no interference in the environment during the vibration.
(2)
Scenario II: illumination variation in the environment during the vibration. We place a lamp near the sphere and switch it on/off several times during the vibration. As shown in Figure 10, the left image represents the case when the lamp is switched off, and the right image when switched on. A significant difference in image brightness can be observed between the two cases.
(3)
Scenario III: fog interference in the environment during the vibration. We place a humidifier between the sphere and the smartphone and turn it on during the vibration to simulate the foggy environment. As shown in Figure 11, the left image represents the case when the humidifier is turned off and the right image when turned on. A significant difference in image contrast can be observed between the two cases.
(4)
Scenario IV: camera jitter during the vibration. We use a pencil to tap the tripod three times during the vibration to make the camera jitter happen after each tapping. As shown in Figure 12 are typical frame images taken from the video at the time right after each tapping. A comparison between Figure 12 and Figure 13 (see later for non-vibrating state) shows that the images are severely blurred due to the camera jitter, in which one can hardly find clear features points.

7.2. Measurement Results

We define three points as the target points that locates at a similar height as the measured point of the LVDT. As shown in Figure 13 are the definition, in the first frame image, of the three target points denoted with blue cross marks and the corresponding target areas indicated with green, red, and yellow boxes. It can be seen that the first target point is sitting on the tip end of the LVDT, the second one is nearly the center of the sphere, and the third one is on the fixed end of the LVDT. During the vibration, the third target point is indeed non-moving. Therefore, in Scenarios I-III, we only track the first and second target points. Moreover, in Scenario IV, we further track the third target point for the sake of excluding the camera-jitter influence from the measurements of the first and second target points.
We firstly adjust the original vibration video shot by the smartphone to a new one by using all the calibrated parameters and coefficients obtained in Section 3, then apply three vision-based measurement methods, i.e., MSTC, STC, and characteristic OF, to the new video to get the corresponding time–history displacements. The characteristic OF method [14] applied herein uses a tracking algorithm that combines Lucas-Kanade optical flow and Harris corners, in which the points for tracking cannot be artificially defined but automatically determined with critical features in the first frame image. So, the tracked points with this OF method cannot be exactly the ones defined above for the MSTC and STC, and we choose one of those feature points that is nearest to the second target point for comparative study.
In the comparative study, we use two types of evaluation indicators [26], e.g., the normalized mean square error indicator (NMSE) and the correlation coefficient indicator ( ρ ), by referring to the LVDT data to evaluate the overall performances of the vision-based measurement methods,
N M S E = 1 i n d v i d s i 2 i n d s i d s ¯ 2
ρ = i = 1 n d v i d v ¯ d s i d s ¯ i = 1 n d v i d v ¯ 2 · i = 1 n d s i d s ¯ 2
where, d v i and d s i are the i-th time-instant displacements obtained with the vision method and the LVDT, respectively; d v ¯ and d s ¯ represent the corresponding mean values of the time–history displacements. For both indicators, a value closer to 1 indicates a better agreement between the vision-measured result and the LVDT data. It is noted that, besides these two overall indicators, some local indices such as deviations from maximum or minimum values are also investigated in the present study for comprehensive understanding.
With the first 15-s time history, the final measurement results in all four scenarios are shown and compared in Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 and Table 3, Table 4, Table 5 and Table 6. Therein, MSTC-1 and MSTC-2 denote the results of the first target point and the second target point identified by the MSTC method, respectively; STC-1 and STC-2 denote the ones by the STC method; LK denotes the ones by the characteristic OF method; sensor denotes the LVDT data. It is noted that the sampling rate for the LVDT is 100 Hz and it is down-sampled to be comparable to the vision measurements with the theoretical or actual frame rate [27].
In Scenario IV, MSTC-1_0, MSTC-2_0, and MSTC-3_0 denote the overall results of the three target points identified by the MSTC method, which include the camera-jitter influence. Corresponding displacements of the first and second target points, denoted by MSTC-1 and MSTC-2, respectively, are then obtained by excluding the camera-jitter influence, e.g., subtracting MSTC-3_0 from MSTC-1_0 and MSTC-2_0. With the STC method, similar definitions hold for STC-1_0, STC-2_0, STC-3_0, STC-1, and STC-2.
Observations on the results will be illustrated in the following.

7.2.1. Scenario I

From Figure 14 and Figure 15 and Table 3, it can be seen that,
(1)
In the absence of interference factors, all three vision methods can accurately capture the sphere’s movement under the complex background, among which the two STC-type methods perform better than the OF method, and the MSTC method provides the most consistent results with the LVDT data.
(2)
Both overall indicators, i.e., NMSE and ρ , can reach the same conclusion in terms of distinguishing the advantages and disadvantages of different methods, and NMSE shows higher sensitivity. See the 59.95 FPS case, for example, ρ keeps larger than 0.99 with all methods while NMSE takes 0.992, 0.988, and 0.987 for MSTC-1, STC-1, and LK, respectively.
(3)
The measurement results obtained with the actual frame rate (59.96 FPS) are more accurate than the ones with the theoretical frame rate (60 FPS). The inconsistency of the frame rate of smartphones between theory and practice has been noticed by Yoon [14], and our study further shows that corresponding errors in displacement measurement even deteriorate in a frequency-varying vibration practice. In the theoretical FPS case, significant shifting happens during the whole time history and exaggerates in later-stage vibration due to error accumulation. Variations of overall indicators also demonstrate the influence. See the MSTC-1 measurement, for example, due to inadequate choice of 60 FPS instead of 59.96 FPS, NMSE drops from 0.992 to 0.953 and ρ from 0.996 to 0.977.
(4)
The maximum and minimum values from the sensor vary with different frame rates due to the down-sampled operation. The corresponding deviations, although not wholly accurate, can still provide an estimation of the local accuracy. For the MSTC method, deviation results are relatively small, e.g., less than 10%, implying a good accuracy of the proposed method.

7.2.2. Scenario II

From Figure 16 and Figure 17 and Table 4, it can be seen that,
(1)
When the illumination undergoes a sudden change during the vibration, the displacement time history obtained with the OF method varies abruptly, indicating failure of the target tracking. However, results with the STC-type methods still keep stable and consistent with the LVDT data in shape and trend. Besides, again, the MSTC method gives the best performance.
(2)
Both overall indicators can lead to the same conclusion in comparing different methods’ performance under illumination variation, and, again, NMSE is more sensitive. See the 59.95 FPS case, for example, ρ keeps larger than 0.96 with all methods, while NMSE takes 0.977, 0.957, and 0.778 for MSTC-1, STC-1, and LK, respectively. Besides, when compared with Scenario I, NMSE drops from 0.987 to 0.778 for LK, from 0.988 to 0.957 for STC-1, and from 0.992 to 0.977 for MSTC-1. The minimum loss of NMSE indicates the best performance of the MSTC method in resisting the interference of illumination variation.
(3)
A significant influence on the measurement results similar to Scenario I still holds for the frame rate of the smartphone. See the MSTC-1 measurement, for example, due to an inadequate choice of 60 FPS instead of 59.96 FPS, NMSE drops from 0.977 to 0.885 and ρ from 0.989 to 0.945.
(4)
The MSTC method and the STC method perform almost the same in measuring target 2, while the former one shows better performance in measuring target 1. See also in Table 4, NMSE takes 0.988 and 0.987 for MSTC-2 and STC-2, respectively, while 0.977 and 0.957 for MSTC-1 and STC-1, respectively. As shown in Figure 13, the proportion of pixels in the target-1 context with similar motion tendency to the target center is much smaller than that in the target-2 one, implying fewer auxiliary information and more interference one for target 1. It results in a more difficult tracking of target 1 under an interference environment. At the same time, the MSTC method can achieve better indices than the STC method by extracting the favorable and unfavorable information from the context and correspondingly strengthening or weakening its influence during the target tracking.
(5)
The most significant deviations happen to the OF method. Moreover, deviations of the MSTC method are smaller than those of the STC method, in which the target 2 shows better performance than the target 1.

7.2.3. Scenario III

From Figure 18 and Figure 19 and Table 5, it can be seen that,
(1)
When the fog continuously interferes with the shooting environment during the vibration, the displacement time history obtained with the OF method significantly shifts from the accurate one, indicating apparent failure of the target tracking. However, the same excellent performance of the MSTC method and the STC method, as in Scenario II, still holds.
(2)
Both overall indicators can lead to the same conclusion in comparing different methods’ performance under fog interference, and, again, NMSE is more sensitive. See the 59.95 FPS case, for example, ρ keeps larger than 0.91 with all methods while NMSE takes 0.969, 0.948, and 0.501 for MSTC-1, STC-1, and LK, respectively. Besides, when compared with Scenario I, NMSE drops from 0.987 to 0.501 for LK, from 0.988 to 0.948 for STC-1, and from 0.992 to 0.969 for MSTC-1. The minimum loss of NMSE indicates the best performance of the MSTC method in resisting the fog interference.
(3)
A significant influence on the measurement results similar to Scenarios I and II still holds for the frame rate of the smartphone. See the MSTC-1 measurement, for example, due to an inadequate choice of 60 FPS instead of 59.96 FPS, NMSE drops from 0.969 to 0.872 and ρ from 0.986 to 0.939.
(4)
The MSTC method, again, shows similar performance as the STC method in measuring target 2 and better performance in measuring target 1, indicating a better anti-interference ability of the proposed method. See also in Table 5, NMSE takes 0.972 and 0.971 for MSTC-2 and STC-2, respectively, while 0.969 and 0.948 for MSTC-1 and STC-1, respectively.
(5)
The most significant deviations happen to the OF method. Moreover, deviations of the MSTC method are smaller than those of the STC method.

7.2.4. Scenario IV

From Figure 20, Figure 21 and Figure 22 and Table 6, it can be seen that,
(1)
When camera jitter happens during the vibration, the OF method fails to find and match the critical feature points in the blurred images. The STC method may significantly mistrack the target point when subjected to substantial camera jitter, as shown in STC-1_0 after the second tapping. The MSTC method, however, can deliver excellent performance in tracking all three target points under all camera-jitter circumstances.
(2)
Both overall indicators can lead to the same conclusion in comparing different methods’ performance under camera jitter, and, again, NMSE is more sensitive. Due to the failure of tracking feature points after camera jitter, the OF method produces no values for both indicators. Moreover, for the STC method, the target mistracking caused by substantial camera jitter even leads to a negative value of NMSE, much smaller than the positive ρ . Besides, when compared with Scenario I, NMSE varies from 0.987 to no value for LK, from 0.988 to negative value for STC-1, and from 0.992 to 0.925 for MSTC-1. The minimum loss of NMSE indicates the best performance of the MSTC method in resisting the interference of camera jitter.
(3)
A significant influence on the measurement results similar to Scenarios I-III still holds for the frame rate of the smartphone. See the MSTC-1 measurement, for example, due to inadequate choice of 60 FPS instead of 59.96 FPS, NMSE drops from 0.925 to 0.894 and ρ from 0.964 to 0.948.
(4)
The MSTC method, as compared to the STC method, shows slightly better performance in measuring target 2 and much better performance in measuring target 1, proving once again a better anti-interference ability of the proposed method. See also in Table 6, NMSE takes 0.933 and 0.926 for MSTC-2 and STC-2, respectively, while 0.925 and negative value for MSTC-1 and STC-1, respectively.
(5)
Deviations of the MSTC method are smaller than those of the STC method, in which the target 2 shows better performance than the target 1.

7.3. Computation Efficiency

After the iPhone 7 recorded the videos of sphere vibration in four scenarios, we implemented all the above displacement identification with Matlab codes on a laptop computer, which is of moderate hardware configuration, e.g., Intel Core i5-4200H 2.8GHz and 4G memory.
During the identification, each 15-s vibration video with 60 FPS contains 900 frame images to be analyzed, among which the first one is shown in Figure 13. The time consumed for the identification with different vision methods in four scenarios is listed in Table 7.
It can be seen from Table 7 that the MSTC method takes 0.14s~0.18s to identify the displacement of one target between two adjacent frame images. Comparatively, the STC method and the OF method take 0.12~0.16 s and 0.08~0.10 s, respectively. Despite a slight increment, i.e., 0.02 s introduced by additional calculation of the context influence matrix, the MSTC method is still applicable for the design of online measurement via its high efficiency, similarly to the STC method that has seen online tracking practice.

7.4. Summary

To sum up, several conclusions can be drawn from all the above observations,
(1)
In the vision-based measurement of engineering vibration, the actual frame rate, instead of the theoretical one of the camera, is crucial to achieving a high-accuracy result of displacement time history.
(2)
To evaluate the performance of vision methods, the indicator, NMSE, is more critical than ρ when interference is involved.
(3)
In an interference environment, the OF method is prone to mismatch the feature points and lead to data deviated or lost. The conventional STC method is sensitive to target selection and can effectively track those targets having a large proportion of pixels in the context with motion tendency similar to the target center. The proposed MSTC method, however, can ease the sensitivity to target selection through in-depth processing of the information in the context and finally enhance the robustness of the target tracking.
(4)
The OF method has the highest computation efficiency in measuring target displacement, followed by the STC method and the MSTC method. Although the MSTC method requires nearly 10% more time than the STC method to measure displacement, the average time cost consumed for tracking one target between adjacent frame images is far less than 1 s, favoring its potential application in online measurement.

8. Further Discussion

In the present study, we develop an improved vision method for structural displacement measurement. This method uses a smartphone to perform multi-point displacement measurements in environments with illumination variation, fog interference, and camera jitter. Despite the high-accuracy results, there still exist two issues to be addressed:
(1)
Temporal aliasing: Compared with professional high-speed cameras, smartphone cameras have a lower sampling frequency, i.e., frame rate, for video shooting, mostly 30/60 FPS, which is vulnerable to time aliasing. According to the Nyquist sampling theorem, the sampling frequency of measurement should be at least two times the highest frequency of response to avoid high-frequency aliasing information. In wired or wireless sensor systems, anti-aliasing filters are usually used to eliminate aliasing effects. In vision systems, however, temporal aliasing cannot be removed because the image has already been aliased when it is taken. Therefore, when the Nyquist sampling theorem cannot be satisfied with the smartphone, a professional camera with a higher frame rate has to be applied for accurate measurement.
(2)
Generalized camera motion: In the experiment, the present method is verified effective under camera jitter, and target displacements can be obtained through directly subtracting the measurement of a non-moving reference point from the overall measurements. In engineering practice, however, the camera motion can be more generalized and complex, such as that of a UAV-carried camera, and reference points cannot fully provide 6-DOF (degrees-of-freedom) camera motions. In this case, to measure the dynamic displacements, it is necessary to introduce new technologies such as dual-camera based motion compensation [28] to exclude the influence of camera motion.

9. Conclusions

In this paper, we studied the application of computer vision technology in the field of structural vibration. A multi-target monitoring method based on a motion-enhanced spatio-temporal context (MSTC) algorithm developed herein was proposed, which realizes the synchronous measurement of multi-point dynamic displacement in an interference environment by using smartphone-recording video. A series of sine-sweep vibration experiments considering different interference factors were carried out to investigate the performance of the proposed method. Test results show good consistency in different scenarios, indicating that the proposed method preserves good adaptability to complex backgrounds and strong resistance to illumination variation, fog interference, and camera jitter. Besides, the proposed method acts insensitive in selecting the target region and efficient in tracking targets, which are beneficial to potential engineering applications.
Although the designed experiments have involved several interference factors, it is no doubt that, for comprehensive performance validation, more investigations on other objects of various profiles and textures are still required, as well as field tests on real vibrating structures. These ongoing validations and further studies on more complex environment changes, such as occlusion, strong wind, and rain, are crucial to extending the vision-related applications.

Author Contributions

Conceptualization, T.C.; methodology, T.C. and Z.Z.; validation, T.C. and Z.Z.; formal analysis, T.C.; resources, Z.Z.; writing—original draft preparation, Z.Z.; writing—review and editing, T.C.; funding acquisition, T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China State Key Laboratory of Subtropical Building Science (grant number 2017KB13).

Acknowledgments

The authors would like to thank Xuan Wang for his technical support in the implementation of the experiment, as well as Zhou Li for his inspiring advice on experiment validation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cho, S.; Spencer, B.F. Sensor Attitude Correction of Wireless Sensor Network for Acceleration-Based Monitoring of Civil Structures. Comput. Civ. Infrastruct. Eng. 2015, 30, 859–871. [Google Scholar] [CrossRef]
  2. Spencer, B.F.; Hoskere, V.; Narazaki, Y. Advances in computer vision-based civil infrastructure inspection and monitoring. Engineering 2019, 5, 199–248. [Google Scholar] [CrossRef]
  3. Feng, D.; Feng, M.Q. Computer vision for SHM of civil infrastructure: From dynamic response measurement to damage detection−A review. Eng. Struct. 2018, 156, 105–117. [Google Scholar] [CrossRef]
  4. Dworakowski, Z.; Kohut, P.; Gallina, A.; Holak, K.; Uhl, T. Vision-based algorithms for damage detection and localization in structural health monitoring. Struct. Control. Health Monit. 2016, 23, 35–50. [Google Scholar] [CrossRef]
  5. Feng, M.Q.; Fukuda, Y.; Feng, D.; Mizuta, M. Non-target vision sensor for remote measurement of bridge dynamic response. J. Bridge Eng. 2015, 20, 04015023. [Google Scholar] [CrossRef]
  6. Luo, L.; Feng, M.Q.; Wu, Z.Y. Robust vision sensor for multi-point displacement monitoring of bridges in the field. Eng. Struct. 2018, 163, 255–266. [Google Scholar] [CrossRef]
  7. Kim, H.; Shin, S. Reliability verification of a vision-based dynamic displacement measurement for system identification. J. Wind Eng. Ind. Aerodyn. 2019, 191, 22–31. [Google Scholar] [CrossRef]
  8. Lee, J.; Lee, K.C.; Cho, S.; Sim, S.H. Computer vision-based structural displacement measurement robust to light-induced image degradation for in-service bridges. Sensors 2017, 17, 2317. [Google Scholar] [CrossRef] [Green Version]
  9. Shariati, A.; Schumacher, T. Eulerian-based virtual visual sensors to measure dynamic displacements of structures. Struct. Control. Health Monit. 2017, 24, 1–9. [Google Scholar] [CrossRef]
  10. Ye, X.W.; Yi, T.H.; Dong, C.Z.; Liu, T. Vision-based structural displacement measurement: System performance evaluation and influence factor analysis. Measurement 2016, 88, 372–384. [Google Scholar] [CrossRef] [Green Version]
  11. He, L.; Tan, J.; Hu, Q.; He, S.; Cai, Q.; Fu, Y.; Tang, S. Non-contact measurement of the surface displacement of a slope based on a smart binocular vision system. Sensors 2018, 18, 2890. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Khuc, T.; Catbas, F.N. Completely contactless structural health monitoring of real-life structures using cameras and computer vision. Struct. Control. Health Monit. 2017, 24, e1852. [Google Scholar] [CrossRef]
  13. Dong, C.Z.; Catbas, F.N. A non-target structural displacement measurement method using advanced feature matching strategy. Adv. Struct. Eng. 2019, 22, 3461–3472. [Google Scholar] [CrossRef]
  14. Yoon, H.; Elanwar, H.; Choi, H.; Golparvar-Fard, M.; Spencer, B.F. Target-free approach for vision-based structural system identification using consumer-grade cameras. Struct. Control. Health Monit. 2016, 23, 1405–1416. [Google Scholar] [CrossRef]
  15. Feng, D.; Feng, M.Q.; Ozer, E.; Fukuda, Y. A vision-based sensor for noncontact structural displacement measurement. Sensors 2015, 15, 16557–16575. [Google Scholar] [CrossRef] [PubMed]
  16. Dong, C.Z.; Celik, O.; Catbas, F.N.; Obrien, E.; Taylor, S. A robust vision-based method for displacement measurement under adverse environmental factors using spatio-temporal context learning and taylor approximation. Sensors 2019, 19, 3197. [Google Scholar] [CrossRef] [Green Version]
  17. Chen, J.G.; Davis, A.; Wadhwa, N.; Durand, F.; Freeman, W.T.; Büyüköztürk, O. Video camera–based vibration measurement for civil infrastructure applications. J. Infrastruct. Syst. 2016, 23, B4016013. [Google Scholar] [CrossRef]
  18. Dong, C.Z.; Celik, O.; Catbas, F.N.; Obrien, E.; Taylor, S. Structural displacement monitoring using deep learning-based full field optical flow methods. Struct. Infrastruct. Eng. 2020, 16, 51–71. [Google Scholar] [CrossRef]
  19. Mai, H.H.; Le, T.T.h. Testing edible oil authenticity by using smartphone based spectrometer. Comput. Opt. 2020, 44, 189–194. [Google Scholar] [CrossRef]
  20. Zhang, Z.Y. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  21. Chan, S.H.; Võ, D.T.; Nguyen, T.Q. Subpixel motion estimation without interpolation. IEEE Int. Conf. Acoust. Speech Sig. Process. 2010, 722–725. [Google Scholar]
  22. Gao, X.; Zhang, T.; Yan, Q.; Liu, Y. 14 Lectures on Visual SLAM: From Theory to Practice; Publishing House of Electronics Industry: Beijing, China, 2017. [Google Scholar]
  23. Zhang, K.; Zhang, L.; Liu, Q.; Zhang, D.; Yang, M.H. Fast visual tracking via dense spatio-temporal context learning. Proc. Eur. Conf. Comput. Vis. 2014, 127–141. [Google Scholar] [CrossRef] [Green Version]
  24. Xu, J.; Lu, Y.; Liu, J. Robust tracking via weighted spatio-temporal context learning. IEEE Int. Conf. Image Process. 2014, 413–416. [Google Scholar] [CrossRef]
  25. Fleet, D.J.; Weiss, Y. Optical flow estimation. In Mathematical Models for Computer Vision: The Handbook; Springer: Berlin/Heidelberg, Germany, 2005; pp. 239–257. [Google Scholar]
  26. Khuc, T. Computer Vision Based Structural Identification Framework for Bridge Health Monitoring. Ph.D. Thesis, University of Central Florida, Orlando, FL, USA, 2016. [Google Scholar]
  27. Dong, C.Z.; Celik, O.; Catbas, F.N. Marker-free monitoring of the grandstand structures and modal identification using computer vision methods. Struct. Health Monit. 2019, 18, 1491–1509. [Google Scholar] [CrossRef]
  28. Lee, J.; Lee, K.C.; Jeong, S.; Lee, Y.J.; Sim, S.H. Long-term displacement measurement of full-scale bridges using camera ego-motion compensation. Mech. Syst. Signal Proc. 2020, 140, 106651. [Google Scholar] [CrossRef]
Figure 1. Flowchart of multi-displacement measurement with smartphone.
Figure 1. Flowchart of multi-displacement measurement with smartphone.
Sensors 20 05929 g001
Figure 2. Identification trend of radial distortion coefficients.
Figure 2. Identification trend of radial distortion coefficients.
Sensors 20 05929 g002
Figure 3. Pre- and post-calibration images of B&W chessboard.
Figure 3. Pre- and post-calibration images of B&W chessboard.
Sensors 20 05929 g003
Figure 4. Definition of three target areas in the first frame image.
Figure 4. Definition of three target areas in the first frame image.
Sensors 20 05929 g004
Figure 5. Graphical model of the spatial relationship between the target and its local context.
Figure 5. Graphical model of the spatial relationship between the target and its local context.
Sensors 20 05929 g005
Figure 6. Context and division result.
Figure 6. Context and division result.
Sensors 20 05929 g006
Figure 7. Local context and influence matrix.
Figure 7. Local context and influence matrix.
Sensors 20 05929 g007
Figure 8. Flow chart of motion-enhanced spatio-temporal context (MSTC) algorithm.
Figure 8. Flow chart of motion-enhanced spatio-temporal context (MSTC) algorithm.
Sensors 20 05929 g008
Figure 9. Experiment layout.
Figure 9. Experiment layout.
Sensors 20 05929 g009
Figure 10. Comparison of variable illumination.
Figure 10. Comparison of variable illumination.
Sensors 20 05929 g010
Figure 11. Comparison of fog simulation.
Figure 11. Comparison of fog simulation.
Sensors 20 05929 g011
Figure 12. Blurred images after each tapping.
Figure 12. Blurred images after each tapping.
Sensors 20 05929 g012
Figure 13. Initial definition of three tracking targets.
Figure 13. Initial definition of three tracking targets.
Sensors 20 05929 g013
Figure 14. Displacement time history and local amplification in Scenario I (59.96 FPS).
Figure 14. Displacement time history and local amplification in Scenario I (59.96 FPS).
Sensors 20 05929 g014
Figure 15. Displacement time history and local amplification in Scenario I (60 FPS).
Figure 15. Displacement time history and local amplification in Scenario I (60 FPS).
Sensors 20 05929 g015
Figure 16. Displacement time history and local amplification in Scenario II (59.96 FPS).
Figure 16. Displacement time history and local amplification in Scenario II (59.96 FPS).
Sensors 20 05929 g016
Figure 17. Displacement time history and local amplification in Scenario II (60 FPS).
Figure 17. Displacement time history and local amplification in Scenario II (60 FPS).
Sensors 20 05929 g017
Figure 18. Displacement time history and local amplification in Scenario III (59.96 FPS).
Figure 18. Displacement time history and local amplification in Scenario III (59.96 FPS).
Sensors 20 05929 g018
Figure 19. Displacement time history and local amplification in Scenario III (60 FPS).
Figure 19. Displacement time history and local amplification in Scenario III (60 FPS).
Sensors 20 05929 g019
Figure 20. Overall measurement time history and local amplification in Scenario IV.
Figure 20. Overall measurement time history and local amplification in Scenario IV.
Sensors 20 05929 g020aSensors 20 05929 g020b
Figure 21. Displacement time history and local amplification in Scenario IV (59.96 FPS).
Figure 21. Displacement time history and local amplification in Scenario IV (59.96 FPS).
Sensors 20 05929 g021
Figure 22. Displacement time history and local amplification in Scenario IV (60 FPS).
Figure 22. Displacement time history and local amplification in Scenario IV (60 FPS).
Sensors 20 05929 g022
Table 1. Intrinsic parameters of iPhone 7 back camera.
Table 1. Intrinsic parameters of iPhone 7 back camera.
Parameter f x f y θ x 0 y 0
value3322.53322.31.68692010.31480.6
Table 2. Distortion coefficients of iPhone 7 back camera.
Table 2. Distortion coefficients of iPhone 7 back camera.
Coefficient k 1 k 2 p 1 p 2
value0.2999−1.66053.03 × 10−43.80 × 10−3
Table 3. Evaluation of displacement results without interference factors.
Table 3. Evaluation of displacement results without interference factors.
59.96 FPS
NMSE ρ Maximum DisplacementMinimum Displacement
Vision (mm)Sensor (mm)DeviationVision (mm)Sensor (mm)Deviation
MSTC-10.9920.9965.7685.4645.55%−5.405−5.495−1.65%
MSTC-20.9920.9965.381−1.52%−5.9618.48%
STC-10.9880.9955.379−1.56%−6.07110.48%
STC-20.9910.9965.6493.37%−5.9949.09%
LK0.9870.9945.460−0.09%−6.26614.03%
60FPS
MSTC-10.9530.9775.7685.900−2.24%−5.405−5.938−8.98%
MSTC-20.9660.9845.381−8.79%−5.9610.39%
STC-10.9480.9755.379−8.82%−6.0712.24%
STC-20.9600.9835.649−4.26%−5.9940.95%
LK0.9350.9695.460−7.46%−6.2665.53%
Table 4. Evaluation of displacement results with illumination variation.
Table 4. Evaluation of displacement results with illumination variation.
59.96 FPS
NMSE ρ Maximum DisplacementMinimum Displacement
Vision (mm)Sensor (mm)DeviationVision (mm)Sensor (mm)Deviation
MSTC-10.9770.9895.2775.758−8.35%−6.000−5.7803.81%
MSTC-20.9880.9945.555−3.53%−5.620−2.76%
STC-10.9570.9884.821−16.27%−6.60014.19%
STC-20.9870.9945.524−4.06%−6.0003.81%
LK0.7780.9686.94820.67%−4.569−20.95%
60 FPS
MSTC-10.8850.9455.2775.679−7.09%−6.000−5.5508.12%
MSTC-20.9030.9535.555−2.20%−5.6201.27%
STC-10.8650.9414.821−15.12%−6.60018.93%
STC-20.9010.9525.524−2.73%−6.0008.12%
LK0.6520.9086.94822.33%−4.569−17.67%
Table 5. Evaluation of displacement monitoring results with fog interference.
Table 5. Evaluation of displacement monitoring results with fog interference.
59.96 FPS
NMSE ρ Maximum DisplacementMinimum Displacement
Vision (mm)Sensor (mm)DeviationVision (mm)Sensor (mm)Deviation
MSTC-10.9690.9865.6025.4532.75%−6.050−5.5289.45%
MSTC-20.9720.9905.7455.36%−6.0008.54%
STC-10.9480.9845.6673.93%−6.80023.01%
STC-20.9710.9885.7876.13%−6.15011.26%
LK0.5010.91710.12985.76%−4.562−17.48%
60 FPS
MSTC-10.8720.9395.6025.5151.58%−6.150−5.55710.66%
MSTC-20.8880.9455.7454.17%−6.0007.96%
STC-10.8500.9375.6672.75%−6.80022.36%
STC-20.8820.9445.7874.93%−6.15010.66%
LK0.4130.87510.12983.65%−4.562−17.92%
Table 6. Evaluation of displacement monitoring results with camera jitter.
Table 6. Evaluation of displacement monitoring results with camera jitter.
59.96 FPS
NMSE ρ Maximum DisplacementMinimum Displacement
Vision (mm)Sensor (mm)DeviationVision (mm)Sensor (mm)Deviation
MSTC-10.9250.9645.2865.868−9.92%−5.643−5.860−3.71%
MSTC-20.9330.9675.514−6.03%−5.808−0.89%
STC-1−29.2270.15629.796407.79%−19.659235.47%
STC-20.9260.9655.400−7.97%−6.0813.77%
LK//////
60 FPS
MSTC-10.8940.9485.1865.840−11.20%−5.643−5.742−1.73%
MSTC-20.9000.9505.514−5.58%−5.8081.15%
STC-1−36.5390.14229.796410.22%−19.659242.36%
STC-20.8870.9455.400−7.53%−6.0815.90%
LK//////
Table 7. Time consumption for displacement identification.
Table 7. Time consumption for displacement identification.
Total Number of TargetsTotal Number of Frame ImagesTotal Time Cost (s)Average Time Cost (s/Target/Frame)
MSTC29002580.143
Scenario ISTC29002300.128
LK1900790.088
MSTC29002620.146
Scenario IISTC29002320.129
LK1900840.093
MSTC29002680.149
Scenario IIISTC29002350.131
LK1900850.094
MSTC39004720.175
Scenario IVSTC39004160.154
LK1///
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, T.; Zhou, Z. An Improved Vision Method for Robust Monitoring of Multi-Point Dynamic Displacements with Smartphones in an Interference Environment. Sensors 2020, 20, 5929. https://doi.org/10.3390/s20205929

AMA Style

Chen T, Zhou Z. An Improved Vision Method for Robust Monitoring of Multi-Point Dynamic Displacements with Smartphones in an Interference Environment. Sensors. 2020; 20(20):5929. https://doi.org/10.3390/s20205929

Chicago/Turabian Style

Chen, Taicong, and Zhou Zhou. 2020. "An Improved Vision Method for Robust Monitoring of Multi-Point Dynamic Displacements with Smartphones in an Interference Environment" Sensors 20, no. 20: 5929. https://doi.org/10.3390/s20205929

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop