Next Article in Journal
Identification of Individual Mobility Anchor Places and Patterns Based on Mobile Phone GPS Data
Previous Article in Journal
Re-Supplying Autonomous Mobile Parcel Lockers in Last-Mile Distribution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Methodology for Automatically Detecting Pan–Tilt–Zoom CCTV Camera Drift in Advanced Traffic Management System Networks

Department of Civil and Construction Engineering, Purdue University, West Lafayette, IN 47907, USA
*
Author to whom correspondence should be addressed.
Future Transp. 2024, 4(4), 1297-1317; https://doi.org/10.3390/futuretransp4040062
Submission received: 20 August 2024 / Revised: 4 October 2024 / Accepted: 17 October 2024 / Published: 1 November 2024

Abstract

Many transportation agencies have deployed pan–tilt–zoom (PTZ) closed-circuit television (CCTV) cameras to monitor roadway conditions and coordinate traffic incident management (TIM), particularly in urbanized areas. Pre-programmed “presets” provide the ability to rapidly position a camera on regions of highways. However, camera views occasionally develop systematic deviations from their original presets due to a variety of factors, such as camera change-outs, routine maintenance, drive belt slippage, bracket movements, and even minor vehicle crashes into the camera support structures. Scheduled manual calibration is one way to systematically eliminate these positioning problems, but it is more desirable to develop automated techniques to detect and alert agencies of potential drift. This is particularly useful for agencies with large camera networks, often numbering in the 1000’s. This paper proposes a methodology using the mean Structured Similarity Index Measure (SSIM) to compare images for a current observation to a stored original image with identical PTZ coordinates. Analyzing images using the mean SSIM generates a single value, which is then aggregated every week to generate potential drift alerts. This methodology was applied to 2200 images from 49 cameras over a 12-month period, which generated less than 30 alerts that required manual validation to determine the confirmed drift detection rate. Approximately 57% of those alerts were confirmed to be camera drift. This paper concludes with the limitations of the methodology and future research opportunities to possibly increase alert accuracy in an active deployment.

1. Introduction

State and local agencies have deployed large pan–tilt–zoom (PTZ) closed-circuit television (CCTV) networks to monitor and manage roadway conditions. Modern advanced traffic management systems (ATMSs) are increasingly becoming equipped with enhanced repositioning and monitoring features [1]. All enhanced repositioning and monitoring features require calibrating and georeferencing a camera’s internal images to the physical world. Through the standard use and maintenance of these cameras, it is likely that cameras will drift from a calibrated position over time.
Limited research on the topic and the wide variety of calibration methodologies with various recalibration processes have motivated this research. Drift analysis using LiDAR georeferenced and calibrated cameras has previously been performed but requires driving the corridor with the proper equipment and significant post-processing before implementing any alerting methodology [2]. Additionally, cameras need to be thoroughly calibrated to derive several parameters describing internal and external orientations for these self-recalibration processes. Thus, a drift detection system that can be applied without much initial setup is desired.
This paper proposes a methodology using the mean Structured Similarity Index Measure (mSSIM) to compare a current “comparison image” to an “original image” previously collected. This novel approach to automating PTZ CCTV camera drift detection saves several routine person-hours of manually inspecting cameras. Often, it is the case that when a camera has recently been inspected and deemed fit for operation, shortly after it “drifts”, leaving enhanced features inoperable. Therefore, the proposed automated methodology is crucial to reduce the time needed to find camera drift, alert maintenance personnel, and increase the operable time of enhanced repositioning features of these tools.
The structure of this paper is as follows: a detailed literature review is presented highlighting the past work and research gaps, followed by the objectives of this research and methodologies for assessing images from a single camera, as well as multiple cameras in a network. This is followed by the results and a brief discussion of the findings. This paper concludes with a concise summary of the research and potential future research opportunities.

2. Literature Review

2.1. Early Cameras in Traffic Management

The use of pan–tilt–zoom (PTZ) closed-circuit television (CCTV) cameras in advanced traffic management systems (ATMSs) has its roots in the early 1970s [3,4]. Initially, these systems were fixed-mounted systems routed to Traffic Management Center (TMC) operators located in a central office [3,5]. From these fixed-mounted systems, TMCs shifted to the installation of PTZ cameras in the late 1990s to increase the robustness of their operations. These new PTZ cameras allowed operators to monitor an entire area overhead and facilitate communications in real time [6,7,8,9].

2.2. Developments in Image Analysis

With the newly installed PTZ cameras, all parties (private and public) quickly began experimenting with and developing systems to offer enhanced features. These efforts were a direct result of multi-disciplinary efforts. The fields of computer vision and machine learning concurrently developed several methodologies often incorporated under the topic of “visual information analysis” [10]. Several notable developments with interest in transportation and traffic management are image-difference detection and object detection [11,12,13,14,15]. One notable contribution came from Kodratoff and Moscatelli, who pushed for machine learning in object detection and scene analysis [16]. In the field of image comparison, the early 2000s hosted several foundational publications. Wang developed his “Structural Similarity Index” (SSIM) algorithm [17]. This was later developed into several specialized applications [18,19].

2.3. Preliminary Applications in Transportation

With several tools at their disposal, transportation professionals set out to build principal Intelligent Transportation Systems (ITSs) and Automated Traffic Management Systems (ATMSs). Mak and Fan were among the first to develop an incident detection algorithm using two adjacent video-based prevalence stations [20]. Schoepflin (and others) worked to create algorithms for estimating mean vehicle speed [21] and other traffic characteristics. This trend was a significant focus in early developments, that is, creating video analytic methodologies to characterize traffic flows [22,23].

2.4. The Need for Camera Calibration

Apart from these video analytics, the parallel development branch was focused on assisting those TMC operators by automating specific job tasks. This branch focuses on geographic calibration, or “geo-calibration”, which aligns a camera’s internal coordinates with the environment around it, enabling automated movements derived solely from real-world coordinates. Galego [24] was one of the first publications on this topic, proposing a geometric approach for calibrating traffic cameras. Shortly after, Dadashi proposed a semi-automated approach to geo-calibration [25]. Recent works implementing these foundations have developed fully autonomous movement methodologies, which allow for cameras to reposition onto positional data related to traffic incidents [1,2,26]. Though these methodologies provide high value for traffic management, the source of success is always in the geo-calibration of these cameras; as such, a method of detecting when deviation or drift from a calibrated state takes place is highly desirable.

2.5. Camera Drift Detection

This apparent need has been a topic of research since the implementation of geo-calibration in PTZ traffic cameras. Lisanti proposed a “continuous localization” technique [27], a landmark tracking algorithm utilizing SURF key-point detection [28]. Other approaches utilize multiple cameras to offer the “recalibration” of drifted cameras [29,30]. While these detect “drift”, initial calibration may be cumbersome on larger-scale systems continually under preventative maintenance regimes and upgrades. Others, motivated to simplify the calibration processes that are more suited for more extensive networks, developed procedures that are less capable of automatically adapting to drift [31]. By accepting a simpler model, practitioners take the responsibility of manually locating and correcting drift events.
One of the first published approaches for these scenarios utilizes wide-angle feature point (FP) matching followed by zoom-in FP matching [32]. This approach requires several well-defined feature points, which may not be applicable to many scenarios due to changing traffic conditions. Other approaches suggest the use of laser scanners to generate 3D point clouds to accurately calculate the geometric relations of cameras and their environments in combination with intrinsic and extrinsic parameters [2,33]. While these offer more accurate calibration and drift detection options, for physically large networks (e.g., statewide systems), collecting the necessary data is costly and time-consuming, often only being performed every few years. This lack of updated data can completely exclude its use for newly installed cameras not previously installed in the network at the time of the data collection [2].
In the current literature, these two approaches define the main approaches to camera calibration deviation. Feature matching can be performed by utilizing one or more cameras for continuous localization. These require intensive calibration processes that are not easily scaled to large camera networks. Laser scanning offers a fast and relatively easy calibration process with accurate ground truth to compare to for detecting drift. These methods require significant capital and time investments to collect data for physically large systems. These approaches leave a gap in research. That is where this research proposes a simplified approach based on the comparison of two images, generating a similarity score and generating alerts for where to focus manual maintenance efforts. Although this proposition appears straightforward, several vital decisions must be made. What image comparison algorithm should be used? What similarity index best captures the results of the algorithm? What is the criterion for alerting? What is the criterion for collecting images to compare? These are just a few examples of these decisions.

2.6. Approaches for Image Comparison

2.6.1. SSIM

The first identified approach is the Structural Similarity Index Measure (SSIM) [17]. This approach offers a statistical measure of the luminance, variance, and correlation of image structure [18]. Often, the mean value SSIM (called mean SSIM or mSSIM) is taken to quantify the similarity of the image. A value of 1 is an identical image (based on the structural measure previously mentioned), and 0 is two images that share no commonality [18].
Several adaptations of the SSIM have also been developed to handle several different use cases: Wang’s Multiscale SSIM (MS-SSIM) [34] varies the viewing conditions (e.g., contrast); Sampat et al.’s Complex Wavelet SSIM (CW-SSIM) [19] has been shown to handle minute variations in the image better than the original but is not necessarily as relevant to the stated problem; 3D-SSIM [35] is used for videos; and spherical SSIM (S-SSIM) [36] is used for spherical projections. This family of algorithms offers an easy-to-use algorithm that generates a single, quantified index for comparison.

2.6.2. SIFT

SIFT (Scale Invariant Feature Transformation), first published by Lowe [12], utilizes distinctive image features. SIFT is able to handle feature descriptor matching between several different views. This algorithm offers increased flexibility when combined with geomatic derivations to determine orientation from a separate view (quantifying the approximate drifted amount) [37]. This algorithm offers calculable components of orientation change, which, when compared to a known orientation, could allow for quantifiable drift. However, to implement this, key features need to be identified, requiring an added layer of “calibration.”

2.6.3. SURF

Partially developed from SIFT, SURF (Speeded-Up Robust Features) was proposed by Bay et al. [28] and performs similar actions as SIFT (object recognition and matching), but in many applications is several times faster. Considering its likeness to SIFT, several of its drawbacks are related to the identification of critical features in a rapidly changing environment.
Several studies have proposed various techniques to identify camera drift. However, most of them require additional calibration, which is time-consuming and cumbersome. Additionally, some of the methods involving lasers require extensive data collection and are computationally expensive. This research proposes a systematic and simple methodology that leverages the mSSIM approach to identify potential drift among a network of agency-maintained PTZ cameras without the need for any external recalibration procedures.

3. Study Objectives

The objective of this paper is to develop a framework for assessing if a PTZ CCTV camera has drifted so that it should be recalibrated. This objective is further partitioned into seven sub-objectives:
  • Define PTZ CCTV drift using a set of images and recorded PTZ values from one camera;
  • Apply the mean SSIM algorithm and determine if it is feasible to detect drift with collected samples;
  • Expand and scale the methodology to 49 cameras along a 53-mile route (on Interstate 465) using one sample for each day of the week (seven per week), and show the systematic application of the methodology;
  • Reduce the number of samples from seven per week to two per week (Sunday and Saturday) to show that the methodology is still valid at lower sample sizes (important when considering a real-time deployment);
  • Establish criteria to classify a potential drift alert;
  • Establish analysis criteria to detect and eliminate false positives;
  • Analyze generated alerts to establish the drift detection rate recommendations.

4. Motivation

Figure is called “camera drift” or “drift.” Drift is defined by the change of a frame with respect to constant features within the image without a change in PTZ values. The drift event in Figure 1 is easily identified by visual inspection, but to analyze a network of 100’s of cameras on a weekly basis would require many person-hours. Therefore, it is desired to have an automated algorithm to detect and alert to drift events.

5. Methods

Figure 2a is a transparent overlay of Figure 1b onto Figure 1a, where Figure 1b is red-shifted at 50% opacity, and Figure 1a is blue-shifted at 100% opacity. Figure 2a shows many areas of common registration, including the following callouts: I, a mile reference post; iii, the year and month aspects of the time overlay; and v, the lane markings. Objects different between images appear as a “ghost” feature, being see-through. Callout ii shows a tractor-trailer in Figure 1b, but not in Figure 1a. Callout iv shows the day aspect of the timestamp overlay, which is less defined due to conflicting numbers being overlayed. Callout vi shows the shadow from a tree on the left of the image.
Figure 2b is the SSIM difference image produced from the algorithm described in [17]. Common pixels approach a value of 1, appearing lighter. Different pixels approach a value of 0, appearing darker. Callouts I, iii, and v appear as lighter values due to their fixed nature between images. In callout ii, the areas of difference and agreeance are caused by the structure of luminance and contrast, distinguishing objects themselves rather than their general relationships. The tractor-trailer area contains areas resembling the roadway in Figure 1a, causing the mixed similarity. Callout iv shows the day aspect of the timestamp overlay as being different. Callout iv demonstrates that the SSIM algorithm works best with large luminance and contrast differences and defined edges [18]. Finally, callout vi shows mixed results due to the random nature of the shadow. Taking the mean SSIM difference frame, a value of 0.49752 is observed, which is important for determining if drift detection is possible using the SSIM algorithm. This approach ignores the decline the mean SSIM due to passing vehicles and assumes the mean SSIM will vary enough to detect when drift occurs.
Figure 3 performs the same analysis as Figure 2, except comparing Figure 1b-c. In both subfigures, callout I is the mile reference post in the original image, callout ii is the drifted position of the mile reference post in the comparison image, and callout iii is the path of the drift. Drift is apparent through Figure 3 and has a corresponding mean SSIM value of 0.25186. The drop in the SSIM is assumed to be directly caused by a drift in the camera position.
The process applied in Figure 2b and Figure 3b establishes the methodology for drift detection. Next, this methodology is applied to 12 months of historical observations to determine the viability of the methodology for camera 52.
The methodology is summarized by the following numbered list:
  • Capture the “original image”.
  • Capture the “comparison image”.
  • Compare the images using the mSSIM algorithm.
  • Store the mSSIM value for the two compared images.
  • Compute the 2-day average mSSIM.
  • Check if the 2-day average mSSIM is greater than the alerting threshold:
    a.
    If greater than or equal to the alerting threshold, do not send an alert;
    b.
    If less than the alerting threshold, send an alert.
  • Repeat 2–6 during the operation of the camera at set intervals.

5.1. Sample Selection Criteria

To apply the developed methodology temporarily, the collected data need to be filtered to common times to limit influences from differential light intensity throughout the day. Data were collected from 1 April 2023 to 1 April 2024 at 5 min intervals and limited by the following:
  • Images whose times occurred at 16:00 UTC ± 10 min, best ensuring constant overhead lighting;
  • Images that have a common PTZ value with the “original image,” defined in the next section.

5.2. Original Image Selection Criteria

A constant comparison image must be selected, called the “original image.” The original image is selected by meeting the following:
  • Images whose times occur at 16:00 UTC ± 1 h;
  • Images that have the most common observed combination of PTZ values;
  • Images that occur within the first week of August 2023.
The first week of August 2023 was selected, as images then were likely to be on clear, sunny days for the study areas. This helps mitigate the effects of weather on SSIM scores, covered in later sections. The original image does not have to be within the data being compared to consider more possible data if the first criterion did not yield an original image.

5.3. Single Camera

Figure 4a plots the mean SSIM versus time for camera 52, shown in Figure 1. Callout I in Figure 4a identifies the approximate date drift was observed in camera 52. Callout ii is the data point closest to the original, having the greatest mean SSIM. Callout iii corresponds to Figure 1a, callout iv to Figure 1b, and callout v to Figure 1c. Callouts iii and iv occur before the drift and have a higher SSIM score than after drift, both greater than 0.3 mean SSIM. In fact, no observations are less than 0.3 mean SSIM until callout I, from which no value exceeds 0.3 mean SSIM. This shows the drift event being detected for camera 52 at the confirmed time and suggests a value of 0.3 mean SSIM as an early suitable threshold for detecting drift.
Data from Figure 4a are aggregated by the mean observed value for each week, Sunday to Saturday, and plotted in Figure 4b. Callouts between both figures are identical, with the addition of a border for weeks containing points from callouts in Figure 4a. Figure 4b reports the timing of the drift event from Figure 4a as slightly delayed due to the aggregation process.

5.4. Multiple Cameras

This methodology was applied to 49 cameras along Interstate 465, a beltway encircling Indianapolis, Indiana (IN), in the United States of America. Each camera adheres to the selection criterion listed in the Sample Selection Criteria and Original Image Selection Criteria sections of this manuscript. The results are depicted as a heat map for the average mean SSIM per week plotted in Figure 5. Callout I is the drift event of camera 52, the same as Figure 4b. Callout ii is the week in which originals were selected, having a higher aggregated mean SSIM value. Callout iii is a sample space where data were not collected due to inaccessibility, or no data met the selection criteria. Figure 5 uses 1 daily sample aggregated per week, resulting in nearly full coverage over consecutive weeks. However, further analysis is performed to check if the reduction in samples can produce similar results. This will also reduce the data collection efforts and associated computational complexity required for processing the images.
To minimize samples, an additional criterion that samples are taken on either Sunday or Saturday is added, as shown in Figure 6. For comparison, Figure 5, which uses 7 samples per week (1 for each day), has a total sample size of 7934 SSIM values aggregated into 1893 weeks. Out of these, 399 aggregated weeks had an average mean SSIM value of 0.3 or less; Figure 6, which only uses two samples (Sunday and Saturday), has a total of 2203 samples aggregated into 1528 weeks. Out of these, 347 aggregated weeks had an average mean SSIM value of 0.3 or less. Limiting the selection criteria resulted in approximately 28% of original samples, 81% of original aggregated weeks, and 87% of aggregated weeks with an average mean SSIM value of 0.3 or less. It is assumed that 2 weekly samples using the updated criteria are sufficient to detect drift when analyzing the historical data.
Callout I in Figure 7 (same data as Figure 6) represents camera 52. Callouts ii, iii, and iv are additional alerts discussed in later figures for cameras 48, 142, and 7, respectively. Callouts v, vi, vii, and viii are samples from camera 43, which had no alert over the observation period and serves as an example of no drift.
Those points with cyan diamonds are generated with the following algorithm, representing the automated alerting algorithm:
  • Extract all aggregated bins with an average mean SSIM value of 0.3 or less;
  • Group extracted bins by camera number;
  • Select the first occurrence in bins with at least two consecutive alerted bins as a new drift alert.
Figure 8 shows ideal comparisons corresponding to callouts v, vi, vii, and viii in Figure 7. Each subfigure uses dashed registration aligned to the top-most point of a light post within the image frame to visually compare alignment. Figure 8a is the first image frame occurring on 11 June 2023, followed sequentially by Figure 8b–d corresponding to callouts v, vi, vii, and viii of Figure 7. All subfigures of Figure 8 are identically registered to the same light post. Therefore, no drift occurred during the observation period.
Figure 9 compares the original and drifted image of camera 48 (Figure 7, callout ii). Figure 9a is the original with callout I pointing to the top of a light post in the frame, the registration point. Figure 9b shows the drifted image where callout ii is the drifted position of the registration point, visually different between images. This case is a true positive; an alert was reported with an observable drift event.
Figure 10 compares the original and drifted images for Figure 7, callout iii. Figure 10a is the original with callout I pointing to the top of a bridge support, the registration point. Figure 10b is the drifted image. The registration between both images is identical. Therefore, this case is a false positive; an alert was reported with no observable drift event.
Callout ii of Figure 10 is the difference in weather conditions in the comparison image, being overcast or raining. Callout iii is the change in foliage due to the change in seasons. The callouts allude to limitations of the SSIM algorithm from varied luminance and contrast values between images. Therefore, alerts need to be verified manually to determine the likely cause of an alert.
Figure 11 compares the original drifted image for Figure 7, callout iv. Figure 11a is the original with callout I as the top of a sign, the registration point. Figure 11b is the drifted image. This case is a false positive, likely due to additional roadway and riprap along the side slopes, callout ii. Consideration should be made to update the original when considerable construction occurs.

6. Results

Figure 12 displays all 31 alerts generated from the alerting algorithm. Three alerts were caused by errors in image capture and removed from the analysis. The remaining alerts were analyzed with the visual methods from Figure 8, Figure 9, Figure 10 and Figure 11, and the results are summarized in Figure 13.
Figure 13 summarizes the application of the proposed drift detection and alerting methodology on 49 cameras on I-465 from 1 April 2023 to 1 April 2024. For the 49 cameras, 2203 unique, original samples were collected, meeting the selection criteria in the Sample Selection Criteria section. The 2203 original samples also represent samples that occurred on a Sunday or Saturday (being, at most, 2 samples per week per camera). Those samples were then aggregated by week, taking the average mean SSIM value, resulting in 1528 total aggregated weeks for 49 cameras, plotted in Figure 6. The 1528 aggregated weeks represent 69% of the original samples.
From the 1528 aggregated bins, the number of samples being assessed for drift was further reduced using the alerting criteria established earlier in this study. Applying the alerting criteria to the 1528 aggregated weeks resulted in 28 alerts. Importantly, these alerts represent estimated drift occurrence, not drift itself, needing to be manually verified. These 28 alerts represent only 1% of the original 2203 samples, or 2% of the 1528 aggregated bins, and were generated by the presented methodology automatically.
Finally, each of those 28 alerts was analyzed manually to determine if a drift had occurred. Analysis was performed by adding registration lines to a point in the original image, adding identical lines to the alerted image, and comparing the two. This resulted in the discovery of 16 alerts, resulting in a drift event having occurred, that is, a 57% positive drift alert rate. These 16 positive alerts represent 1% of both the original and aggregated samples, which resulted in drift.

7. Discussion

7.1. Limitations

This methodology uses the original variant of the SSIM algorithm without any additional processing. Thus, the comparisons are susceptible to luminance and contrast differences caused by weather and vegetation changes (Figure 10), roadway changes (Figure 11), or other occlusions (e.g., support cables) resulting in false-positive alerts.
The historical observations limited data collection as a passive operation, collecting data wherever the camera was repositioned during normal use. The selection criteria further limited the usable samples for this analysis. An active data collection approach can anticipate larger sample sizes. Higher true-positive alert rates are expected because of larger sample sizes in cases with weather effects or other discussed factors.
PTZ CCTV cameras are limited in the precision of internal PTZ measurements. The PTZ values for this study were limited to 1° of pan precision, 0.1° of tilt precision, and 1× of zoom (integer precision). Thus, images with reported identical PTZ values are not necessarily in the exact same position, which can result in a false positive for “drift”, resulting in calibration disruption.

7.2. Future Research

Future research on this topic has vast opportunity. This work presented the results from a single image comparison algorithm proofed on historic observations. Future work should analyze the application of other image comparison algorithms and approaches such as CW-SSIM, SURF, or SIFT.
Additionally, a more comprehensive examination of analysis techniques for determining the effectiveness of any given technique is required. The proposed qualitative assessment based on quantitative alerts mirrors the approach an operator may use in determining if a drift event has occurred.
Regarding the effectiveness of a technique, future research should study the environmental effects of each technique, such as weather, traffic conditions, lighting conditions, or natural environmental conditions. Environmental effects were shown to play a significant role in the technique, but further work is necessary to quantify what these effects are and the best approach to mitigate their impacts.
An adjacent topic for future research is the application of these techniques for determining when movements occur to better the working relationship between automated systems and operators. Interference between operators and automated systems is of major concern in this field, and thus a solution is highly desirable.
Overall, further analysis of this technique with future developments is a necessity in finding the best possible solution that is easily implemented, computationally light, and performs with few false positives.

8. Conclusions

Automated pan–tilt–zoom CCTV camera drift detection is vital to the implementation and maintenance of enhanced repositioning and monitoring features. Using the mean SSIM algorithm to assign each sampled image is efficient for 49 cameras in a network and easily scaled to much larger networks. Aggregating scores mitigate artificially low scores from external factors (e.g., weather, construction, or traffic changes). For this methodology, 12 months of data for 49 cameras on Interstate 465 with historical data only yielded 2203 samples (automatically collected) and resulted in 28 alerts requiring manual validation. Alerts are derived using a threshold approach, where this work used a 0.3 mSSIM in at least two consecutive days to trigger an alert. The elimination of false positives was performed by a manual analysis of alerts. Future research is required to identify techniques to better the approach in eliminating false positives using this methodology. The resulting 1% of initial samples requiring manual validation is a more manageable size, requiring far fewer person-hours and lessening the potential downtime of enhanced features.
Modifications to the standard SSIM algorithm have the potential to yield higher passing rates. Additionally, applications of preprocessing methodologies or other similarity indexing algorithms could see improved results.
The proposed methodology is an efficient, lightweight approach to monitoring potential drift in ATMS CCTV camera networks and generating alerts for less than 1% of samples over 12 months. This reduction in sample size to alert size will prove powerful in maintaining large ASTM CCTV networks as more enhanced features are deployed in the coming years by limiting the required person-hours for detecting drift.

Author Contributions

Conceptualization, C.G., J.K.M., and D.B.; data curation, J.K.M.; formal analysis, C.G. and J.K.M.; investigation, C.G., J.K.M., and D.B.; methodology, C.G., J.K.M., and D.B.; software, C.G.; supervision, D.B.; validation, C.G.; visualization, C.G., J.K.M., and D.B.; writing—original draft, C.G.; writing—review and editing, J.K.M. and D.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data for this research are not available due to the large storage requirements associated with the collected images.

Acknowledgments

The contents of this paper reflect the views of the authors, who are responsible for the facts and the accuracy of the data presented herein, and do not necessarily reflect the official views or policies of the sponsoring organization.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Malackowski, H.A. Techniques For Reducing Traffic Management Center Camera Positioning Latency For Accelerated Incident Response. Master’s Thesis, Purdue University Graduate School, West Lafayette, IN, USA, 2024. [Google Scholar]
  2. Mathew, J.K.; Malackowski, H.A.; Koshan, Y.; Gartner, C.; Desai, J.; Li, H.; Cox, E.D.; Habib, A.; Bullock, D.M. Joint Transportation Research Program. Development of Latitude/Longitude (and Route/Milepost) Positioning Traffic Management Cameras; Purdue University: West Lafayette, IN, USA, 2023. [Google Scholar]
  3. Castleman, M. A History of Traffic Management Technology. 2019. Available online: https://streets.mn/2019/10/16/a-history-of-traffic-management-technology/ (accessed on 3 October 2024).
  4. Stehr, R.A. Minnesota Department of Transportation Experience in the Application of Advanced Traffic Management Systems; ASCE: Minneapolis, MN, USA, 1991; pp. 458–462. [Google Scholar]
  5. Anonymous. City Keeps an Eye on Traffic with Video Technology. Am. City Cty. 2003, 118, 22. [Google Scholar]
  6. Anonymous. Optelecom-NKF Awarded Projects by Imtech: [1]. Telecomworldwire 2008. Available online: https://www.proquest.com/trade-journals/optelecom-nkf-awarded-projects-imtech/docview/190999863/se-2?accountid=13360 (accessed on 3 October 2024).
  7. McAuliffe, T.P. Pan, Tilt, Zoom, Better than Ever. Access Control Secur. Syst. 2002, 45, 50. [Google Scholar]
  8. Careless, J. Government Video; Future Publishing Ltd.: New York, NY, USA; London, UK, 2011; pp. 20–23. [Google Scholar]
  9. Anonymous. Caltrans “Converts” for the Future. Commun. News 2004, 41, 26–30. [Google Scholar]
  10. Petkovic, D.; Wilder, J.; Ejiri, M.; Haralick, R.M.; Jain, R.; Ruetz, P.A.; Sklansky, J.; Swonger, C.W. Machine Vision in the 1990s: Applications and How to Get There. Mach. Vis. Apps. 1991, 4, 113–126. [Google Scholar] [CrossRef]
  11. Rosin, P.; Ellis, T. Image Difference Threshold Strategies and Shadow Detection. In Proceedings of the British Machine Conference, Birmingham, UK, 11–14 September 1995; Volume 1. [Google Scholar] [CrossRef]
  12. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  13. LeCun, Y.; Bengio, Y.; Laboratories, T.B. Convolutional Networks for Images, Speech, and Time-Series. In The Handbook of Brain Theory and Neural Networks; ACM: New York, NY, USA, 1998. [Google Scholar]
  14. Bruzzone, L.; Prieto, D.F. Automatic Analysis of the Difference Image for Unsupervised Change Detection. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1171–1182. [Google Scholar] [CrossRef]
  15. Lowe, D.G. Object Recognition from Local Scale-Invariant Features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  16. Kodratoff, Y.; Moscatelli, S. Machine Learning for Object Recognition and Scene Analysis. Int. J. Patt. Recogn. Artif. Intell. 1994, 08, 259–304. [Google Scholar] [CrossRef]
  17. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  18. Nilsson, J.; Akenine-Möller, T. Understanding SSIM. arXiv 2020, arXiv:2006.13846. [Google Scholar]
  19. Sampat, M.P.; Wang, Z.; Gupta, S.; Bovik, A.C.; Markey, M.K. Complex Wavelet Structural Similarity: A New Image Similarity Index. IEEE Trans. Image Process. 2009, 18, 2385–2401. [Google Scholar] [CrossRef]
  20. Mak, C.L.; Fan, H.S.L. Heavy Flow-Based Incident Detection Algorithm Using Information From Two Adjacent Detector Stations. J. Intell. Transp. Syst. 2006, 10, 23–31. [Google Scholar] [CrossRef]
  21. Schoepflin, T.N. Algorithms for Estimating Mean Vehicle Speed Using Uncalibrated Traffic Management Cameras. Ph.D. Thesis, University of Washington, Seattle, WA, USA, 2003. [Google Scholar]
  22. Thornton, J.; Baran-Gale, J.; Yahr, A. An Assessment of the Video Analytics Technology Gap for Transportation Facilities. In Proceedings of the 2009 IEEE Conference on Technologies for Homeland Security, Waltham, MA, USA, 11–12 May 2009; pp. 135–142. [Google Scholar]
  23. Zheng, X.-Q.; Wu, Z.; Xu, S.-X.; Guo, M.-M.; Lin, Z.-X.; Zhang, Y.-Y. Video-Based Measurement and Data Analysis of Traffic Flow on Urban Expressways. Acta Mech. Sin. 2011, 27, 346–353. [Google Scholar] [CrossRef]
  24. Galego, R.; Bernardino, A.; Gaspar, J. Auto-Calibration of Pan-Tilt Cameras Including Radial Distortion and Zoom. In Proceedings of the Advances in Visual Computing; Bebis, G., Boyle, R., Parvin, B., Koracin, D., Fowlkes, C., Wang, S., Choi, M.-H., Mantler, S., Schulze, J., Acevedo, D., et al., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 169–178. [Google Scholar]
  25. Dadashi, N.; Stedmon, A.W.; Pridmore, T.P. Semi-Automated CCTV Surveillance: The Effects of System Confidence, System Accuracy and Task Complexity on Operator Vigilance, Reliance and Workload. Appl. Ergon. 2013, 44, 730–738. [Google Scholar] [CrossRef]
  26. Mathew, J.K.; Malackowski, H.A.; Gartner, C.M.; Desai, J.; Cox, E.D.; Habib, A.F.; Bullock, D.M. Methodology for Automatically Setting Camera View to Mile Marker for Traffic Incident Management. J. Transp. Technol. 2023, 13, 708–730. [Google Scholar] [CrossRef]
  27. Lisanti, G.; Masi, I.; Pernici, F.; Del Bimbo, A. Continuous Localization and Mapping of a Pan–Tilt–Zoom Camera for Wide Area Tracking. Mach. Vis. Appl. 2016, 27, 1071–1085. [Google Scholar] [CrossRef]
  28. Bay, H. Surf: Speeded Up Robust Features; Computer Vision—ECCV 2006; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  29. Eslami, M.A.; Rzasa, J.R.; Milner, S.D.; Davis, C.C. Dual Dynamic PTZ Tracking Using Cooperating Cameras. Intell. Control Autom. 2015, 6, 45–54. [Google Scholar] [CrossRef]
  30. Mao, K.; Xu, Y.; Wang, R. A General Calibration Method for Dual-PTZ Cameras Based on Feedback Parameters. Appl. Sci. 2022, 12, 9148. [Google Scholar] [CrossRef]
  31. Bastanlar, Y. A Simplified Two-View Geometry Based External Calibration Method for Omnidirectional and PTZ Camera Pairs. Pattern Recognit. Lett. 2016, 71, 1–7. [Google Scholar] [CrossRef]
  32. Liao, H.-C.; Pan, M.-H.; Hwang, H.-W.; Chang, M.-C.; Chen, P.-C. An Automatic Calibration Method Based On Feature Point Matching For The Cooperation Of Wide-Angle And Pan-Tilt-Zoom Cameras. Inf. Technol. Control 2011, 40, 41–47. [Google Scholar] [CrossRef]
  33. Hillemann, M.; Jutzi, B. UCalMiCeL—Unified Intrinsic And Extrinsic Calibration Of A Multi-Camera-System And A Laserscanner. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, IV-2-W3, 17–24. [Google Scholar] [CrossRef]
  34. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale Structural Similarity for Image Quality Assessment. In Proceedings of the Thirty-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; pp. 1398–1402. [Google Scholar]
  35. Zeng, K.; Wang, Z. 3D-SSIM for Video Quality Assessment. In Proceedings of the 2012 19th IEEE International Conference on Image Processing, Orlando, FL, USA, 30 September–3 October 2012; pp. 621–624. [Google Scholar]
  36. Chen, S.; Zhang, Y.; Li, Y.; Chen, Z.; Wang, Z. Spherical Structural Similarity Index for Objective Omnidirectional Video Quality Assessment. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23–27 July 2018; pp. 1–6. [Google Scholar]
  37. Castillo-Carrión, S.; Guerrero-Ginel, J.-E. SIFT Optimization and Automation for Matching Images from Multiple Temporal Sources. Int. J. Appl. Earth Obs. Geoinf. 2017, 57, 113–122. [Google Scholar] [CrossRef]
Figure 1. Camera 52 image drift: (a) view on 8 September 2023 at 12:20:09 with PTZ 284°, −12.4°, 1×; (b) view on 13 September 2023 at 12:20:09 with PTZ 284°, −12.4°, 1×; (c) view on 20 September 2023 at 12:35:12 with PTZ 284°, −12.4°, 1×.
Figure 1. Camera 52 image drift: (a) view on 8 September 2023 at 12:20:09 with PTZ 284°, −12.4°, 1×; (b) view on 13 September 2023 at 12:20:09 with PTZ 284°, −12.4°, 1×; (c) view on 20 September 2023 at 12:35:12 with PTZ 284°, −12.4°, 1×.
Futuretransp 04 00062 g001
Figure 2. Image comparison of Figure 1a,b: (a) transparent overlay of Figure 1b onto Figure 1a; (b) SSIM difference of Figure 1a,b, mean SSIM value of 0.49752.
Figure 2. Image comparison of Figure 1a,b: (a) transparent overlay of Figure 1b onto Figure 1a; (b) SSIM difference of Figure 1a,b, mean SSIM value of 0.49752.
Futuretransp 04 00062 g002aFuturetransp 04 00062 g002b
Figure 3. Image comparison of Figure 1b,c: (a) transparent overlay of Figure 1c onto Figure 1b; (b) SSIM difference of Figure 1b,c, mean SSIM value of 0.25186.
Figure 3. Image comparison of Figure 1b,c: (a) transparent overlay of Figure 1c onto Figure 1b; (b) SSIM difference of Figure 1b,c, mean SSIM value of 0.25186.
Futuretransp 04 00062 g003aFuturetransp 04 00062 g003b
Figure 4. Mean SSIM vs. time for camera 52: (a) daily samples meeting selection criteria; (b) daily samples aggregated by average mean SSIM per week.
Figure 4. Mean SSIM vs. time for camera 52: (a) daily samples meeting selection criteria; (b) daily samples aggregated by average mean SSIM per week.
Futuretransp 04 00062 g004
Figure 5. Average mean SSIM by week for cameras on I-465 using 7 samples per week (1 for each day).
Figure 5. Average mean SSIM by week for cameras on I-465 using 7 samples per week (1 for each day).
Futuretransp 04 00062 g005
Figure 6. Average mean SSIM by week for cameras on I-465 using 2 samples per week (Sunday and Saturday).
Figure 6. Average mean SSIM by week for cameras on I-465 using 2 samples per week (Sunday and Saturday).
Futuretransp 04 00062 g006
Figure 7. Selected alerts and non-alerts for cameras on I-465 using 2 samples per week (Sunday and Saturday).
Figure 7. Selected alerts and non-alerts for cameras on I-465 using 2 samples per week (Sunday and Saturday).
Futuretransp 04 00062 g007
Figure 8. Camera 43 image comparison, “no drift”: (a) view on 11 June 2023 at 12:00:10 (Figure 7, callout v); (b) view on 20 August 2023 at 12:00:07 (Figure 7, callout vi); (c) view on 3 December 2023 at 11:05:09 (Figure 7, callout vii); (d) view on 25 February 2023 at 11:05:08 (Figure 7, callout viii).
Figure 8. Camera 43 image comparison, “no drift”: (a) view on 11 June 2023 at 12:00:10 (Figure 7, callout v); (b) view on 20 August 2023 at 12:00:07 (Figure 7, callout vi); (c) view on 3 December 2023 at 11:05:09 (Figure 7, callout vii); (d) view on 25 February 2023 at 11:05:08 (Figure 7, callout viii).
Futuretransp 04 00062 g008
Figure 9. Camera 48 alert comparison, “true positive”: (a) original; (b) alerted image (Figure 7, callout ii).
Figure 9. Camera 48 alert comparison, “true positive”: (a) original; (b) alerted image (Figure 7, callout ii).
Futuretransp 04 00062 g009aFuturetransp 04 00062 g009b
Figure 10. Camera 142 alert comparison, “false positive” with environmental factors: (a) comparison image; (b) alerted image, callout iii of Figure 7.
Figure 10. Camera 142 alert comparison, “false positive” with environmental factors: (a) comparison image; (b) alerted image, callout iii of Figure 7.
Futuretransp 04 00062 g010aFuturetransp 04 00062 g010b
Figure 11. Camera 7 alert comparison, “false positive” with environmental influences: (a) comparison image; (b) alerted image, callout iv of Figure 7.
Figure 11. Camera 7 alert comparison, “false positive” with environmental influences: (a) comparison image; (b) alerted image, callout iv of Figure 7.
Futuretransp 04 00062 g011aFuturetransp 04 00062 g011b
Figure 12. All alerts for cameras on I-465 using 2 samples per week (Sundays and Saturdays).
Figure 12. All alerts for cameras on I-465 using 2 samples per week (Sundays and Saturdays).
Futuretransp 04 00062 g012
Figure 13. Funnel chart for 12 months of data for 49 cameras on I-465.
Figure 13. Funnel chart for 12 months of data for 49 cameras on I-465.
Futuretransp 04 00062 g013
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gartner, C.; Mathew, J.K.; Bullock, D. Methodology for Automatically Detecting Pan–Tilt–Zoom CCTV Camera Drift in Advanced Traffic Management System Networks. Future Transp. 2024, 4, 1297-1317. https://doi.org/10.3390/futuretransp4040062

AMA Style

Gartner C, Mathew JK, Bullock D. Methodology for Automatically Detecting Pan–Tilt–Zoom CCTV Camera Drift in Advanced Traffic Management System Networks. Future Transportation. 2024; 4(4):1297-1317. https://doi.org/10.3390/futuretransp4040062

Chicago/Turabian Style

Gartner, Christopher, Jijo K. Mathew, and Darcy Bullock. 2024. "Methodology for Automatically Detecting Pan–Tilt–Zoom CCTV Camera Drift in Advanced Traffic Management System Networks" Future Transportation 4, no. 4: 1297-1317. https://doi.org/10.3390/futuretransp4040062

APA Style

Gartner, C., Mathew, J. K., & Bullock, D. (2024). Methodology for Automatically Detecting Pan–Tilt–Zoom CCTV Camera Drift in Advanced Traffic Management System Networks. Future Transportation, 4(4), 1297-1317. https://doi.org/10.3390/futuretransp4040062

Article Metrics

Back to TopTop