Next Article in Journal
Intelligent Replica Selection in Edge and IoT Environments Using Artificial Neural Networks
Previous Article in Journal
A Survey on MAC-Based Physical Layer Security over Wireless Sensor Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Geo-Location Method for Images of Damaged Roads

Air and Missile Defense College, Air Force Engineering University, Xi’an 710051, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(16), 2530; https://doi.org/10.3390/electronics11162530
Submission received: 21 July 2022 / Revised: 10 August 2022 / Accepted: 10 August 2022 / Published: 12 August 2022
(This article belongs to the Topic Computer Vision and Image Processing)

Abstract

:
Due to the large difference between normal conditions and damaged road images, geo-location in damaged areas often fails due to occlusion or damage to buildings and iconic signage in the image. In order to study the influence of post-war building and landmark damage conditions on the geolocation results of localization algorithms, and to improve the geolocation effect of such algorithms under damaged conditions, this paper used informative reference images and key point selection. Aiming at the negative effects of occlusion and landmark building damage in the retrieval process, a retrieval method called reliability- and repeatability-based deep learning feature points is proposed. In order to verify the effectiveness of the above algorithm, this paper constructed a data set consisting of urban, rural and technological parks and other road segments as a training set to generate a database. It consists of 11,896 reference images. Considering the cost of damaged landmarks, an artificially generated method is used to construct images of damaged landmarks with different damage ratios as a test set. Experiments show that the database optimization method can effectively compress the storage capacity of the feature index and can also speed up the positioning speed without affecting the accuracy rate. The proposed image retrieval method optimizes feature points and feature indices to make them reliable against damaged terrain and images. The improved algorithm improved the accuracy of geo-location for damaged roads, and the method based on deep learning has a higher effect on the geo-location of damaged roads than the traditional algorithm. Furthermore, we fully demonstrated the effectiveness of our proposed method by constructing a multi-segment road image dataset.

1. Introduction

The Global Navigation Satellite System (GNSS) has been a relatively mature posi-tioning technology, and it has been widely used in many fields. In the actual use of the GNSS positioning method, the ground signal strength is weak, and the civil coding structure is open, which makes the GNSS signal vulnerable to interference under complex electromagnetic interference and is prone to positioning failure when it is interfered by malicious deception. Although nowadays, smart devices have the function of obtaining geographic tags, when the GNSS signal is interfered and still needs to be positioned, users cannot rely on GNSS to obtain positioning information, and they are located in cities, post-disaster damaged areas, and geodetic control points. Being damaged and unable to be used normally, the inertial navigation equipment cannot be effectively calibrated, so it is difficult to use the inertial navigation equipment for precise positioning. In recent years, visual place recognition has received great attention in the field of machine vision, which can be used to solve the problem of location information localization. If relatively accurate geo-location information is added to these images, they can be of great benefit in areas such as outdoor localization [1], pedestrian detection [2], autonomous driving [3], etc. In addition, pictures with geolocation information can also help environment perception technology for robots [4] and urban construction. Therefore, it is a problem that needs to be researched to identify the visual position of damaged road images under the condition of interference with GNSS signals, and to perform geolocation at the same time.
In order to correctly localize a road image, the image-based localization (IBL) task matches image features with unknown location information to image feature labels with GNSS information in the database [5]. Precise image geo-location has long been a challenge, and geolocation using images involves image retrieval, including identifying, extracting, and indexing geographic information features from massive databases. Simultaneous localization and mapping (SLAM) is also widely used in image geolocation to map and locate objects. However, under the condition of interference of GNSS signals, such as damaged landmarks, traditional image retrieval methods cannot capture enough features, resulting in retrieval failure, and the matching results of SLAM will also be affected. It is impossible to draw maps and calculate latitude and longitude under conditions where landmark features in the image are corrupted, or where the image is obscured or contaminated. Additionally, different types of features will be highlighted under different conditions [6]. Under extreme conditions such as occlusion and landmark damage, the altered landform hides most of the features, making visual location recognition difficult. In the future network system, it is also a very important requirement for the unification of location information in the same coordinate system with high accuracy. The traditional method of map-based emergency positioning has gradually been difficult to meet the requirements of high precision.
Content-based image retrieval (CBIR) is one of the important applications of deep learning in computer vision. The purpose of CBIR is to search images with similar content from the database. Under the condition that the landmarks in the image are damaged, the change of the scene in the image will increase the difficulty of image retrieval, and the feature points used for positioning in the image will change. Deep learning has been successful in many fields in recent years, including computer vision (CV) [7,8,9,10].
The descriptors of the whole image are calculated directly using the global descriptor method. The Gist descriptor proposed by Olive [11] et al. is a widely used global feature descriptor, which uses Gabor filters to extract image information in different directions and frequencies and compress them into vectors as image descriptions. Lowry et al. [12] used online learning to train PCA transformation and pointed out that the PCA features corresponding to the first half of the dimension represent similar information in continuous image sequences and are susceptible to environmental changes, while the second half has environmental conditions invariance and can better correspond to the environment variety. Ulrich et al. [13] used panoramic color image histograms combined with nearest neighbors to learn to match images. Local feature point descriptors can also generate global image descriptions. For example, Sunderhauf et al. [14] first down-sample the image and then calculate the brief descriptor around the center of the down-sampled image, which is suitable for some large-scale applications.
Local features are generally more robust to occlusion, scale, rotation, and illumination changes. These methods start with a detection phase, where points of interest are found in the image, followed by a description phase, where a bit of metric is extracted from around these key points. Local features have better recognition capability, thus improving the recognition rate and reducing the detection error. The most commonly used one is the SIFT algorithm proposed by Lowe [15], which uses a Gaussian convolution kernel to construct a scale space and extracts feature points in an image pyramid, where the descriptor of each feature point consists of a 128-dimensional vector. The algorithm is invariant to scale, rotation, and illumination, making it widely used in early visual localization [16,17,18], but because the SIFT algorithm is very time-consuming to extract feature points and descriptors, the subsequent development of algorithms such as the SURF algorithm proposed by Bay et al. [19] and the ORB algorithm proposed by Rubbee et al. [20] mostly improve efficiency at the expense of performance.
The complementary advantages of local feature descriptors and global feature descriptors have led many researchers to use global description methods to generate descriptors for local regions of images. Gradually, many object feature extraction methods have emerged, such as the RPN [21] network and the edge-boxes algorithm [22]. The RPN obtains the potential area of a specific target through the learning method, and the edge-boxes algorithm judges whether it contains an object by the size of the outline information inside the box, which is universal. The method of generating local regions from the original image and then generating global descriptors for the local regions takes into account appearance invariance and perspective invariance and makes the scene definition more flexible [23]. A large amount of literature [24,25,26,27] has shown that deep learning-based feature extraction methods outperform traditional methods, especially deep convolutional neural network (DCNN)-based feature extraction methods for image retrieval and feature point extraction by deep learning-based features for location identification, which can achieve results unattainable by traditional algorithms.
Accurate geo-localization has long been a challenge [28], especially under obscured content in images, and there are still many problems to be solved. First, errors in position recognition can cause the wrong positioning of the positioning algorithm, thus reducing the positioning accuracy and even leading to ultimate positioning failure. Secondly, most of the existing visual location recognition algorithms are content-based. However, typical landmarks and building features on the same road are easily altered by damage from various factors, resulting in a large change in the content of the image even in the same place.
According to the image representation method, a reasonable decision model is used to infer the final recognition point, and the decision model can effectively improve the accuracy and recall rate of location recognition. The matching of input images to stored database images treats the recognition problem as a large-scale retrieval problem. There are many solutions around the problem of how to retrieve images quickly and accurately, such as bag-of-words (BoW) [29] and its derivatives. The core of this series of methods is to encode images into dictionaries with reference to the idea of text search. Search and match in the form of a dictionary. This derivative method also appears to be more effective. For example, Cummins [30] et al. combined BoW with concepts to search and match images and achieved very good results. In addition, FAB-MAP2.0 [31] uses an inverse index structure to store map description information. The images that own the term are stored under each term, not under each image, making the search space size only related to the number of terms and not limited by the map size. It is evident from the related works that most researchers have tried to eliminate the effects of extreme conditions using various techniques.
Aiming at the problem of damaged image geolocation, this paper proposes a road image geolocation method for damaged buildings and landmark areas. The method employs a three-step strategy.
  • For data with low occlusion or useful information, we used improved semantic segmentation algorithms to filter the dataset and reduce the number of images in the dataset to speed up localization time;
  • Aiming at the phenomenon that the image damage area interferes with the image retrieval results, to perform coarse geo-location, we propose a deep learning feature point-based image retrieval method;
  • Fine-grained geo-location using heading angle information was finally completed with experimental validation on our dataset, proving the effectiveness of our method.
The remainder of this paper is organized as follows. Section 2 introduces the implementation of the method and expounds its rationale. Section 3 constructs the database using our self-harvested data as the training set and the generated damaged road images as the test set and describes the experimental procedure and results in detail. Section 4 lists the conclusions and future work.

2. Methods

Although previous methods have also provided relatively accurate geographic location information, the proposed method provided the ability to maintain accurate geographic location information under the condition of environmental damage to road images. This is achieved through the design of reference dataset optimization and retrieval algorithms. The database optimization is performed by an improved semantic segmentation algorithm that automatically filters the information in it in batch. The proposed system consists of two key parts: an online part and an offline part, as shown in Figure 1. In the user’s daily work, the system is used to collect data of typical features that may be applied in urban, desert and Gobi environmental areas. After data training processing, a corresponding relationship can be established through the time information and the location information obtained by applying the multi-source navigation and positioning data fusion method to form a data pair and store it in the offline feature database. When the user cannot quickly obtain accurate positioning during emergency positioning, the online system can respond quickly and obtain high-precision positioning information. When the location is around the objects stored in the offline database. It can use the stereo data binocular vision fast matching method to calculate and match the similarity between the real-time captured image and the database image. At the same time, the spatial position information of the typical feature is read, and the relative position relationship between the database image shooting point and the online point to be measured is established according to the rotation matching matrix obtained by the visual method to transmit the position information, and finally, the positioning information is obtained.

2.1. Dataset Filtering and Damaged Territory Dataset Construction Based on Improved Semantic Segmentation Methods

2.1.1. Improved Semantic Segmentation Algorithm

For the current Deeplab v3+ [32] semantic segmentation algorithm that does not use high-resolution shallow features, the phenomenon of wrong segmentation and omission of segmentation occurs. This paper uses the improved semantic segmentation algorithm proposed by the author [33]. The improved Deeplab v3+ network structure is shown in Figure 2. Compared with the original Deeplab v3+ algorithm, the main structure is still the encoder–decoder structure, and the backbone network used in the original algorithm is ResNet [34]. The purpose of the improved network is to improve the “loss” of information in the semantic segmentation task for damaged buildings and damaged road signs in damaged road images. The purpose of the improved algorithm is to segment the edge information of the damaged object more accurately and accurately and to improve or prevent this phenomenon of wrong segmentation and missing segmentation.
The main improvements are as follows:
  • In the coding layer, the backbone network is improved, and PyConv [35] is introduced. The main idea of introducing PyConv is to divide the input features into different groups for pyramid convolution and perform convolution calculations independently. Compared with standard convolution, PyConv can enlarge the receptive field of the kernel without increasing the computational cost. It can also apply different types of kernels in parallel to process the input, with different spatial resolutions and depths. Along the Resnet network, we can identify four main stages based on the spatial size of the feature maps. The pyramid convolution kernel divides the output channels into four groups in the first stage of the network, each group uses the convolution kernel size of 3 × 3, 5 × 5, 7 × 7, 9 × 9, and the number of output channels is 64, 64, 64, 64. The corresponding grouped convolution group numbers are 1, 4, 8, and 16, respectively. The second stage divides the output channels into three groups, and each group uses convolution kernel sizes of 3 × 3, 5 × 5, and 7 × 7, respectively. The number of output channels is 128, 128, 256, respectively. The corresponding grouped convolution group numbers are 1, 4, and 8, respectively. The third stage divides the output channels into two groups. The size of the convolution kernels used in each group is 3 × 3 and 5 × 5, respectively, and the number of output channels is 512 and 512, respectively. The corresponding grouped convolution group numbers are 1 and 4, respectively. The fourth stage convolves all channels with a 3 × 3 convolution kernel, and the number of groups is 1. The specific parameter details of pyconv are shown in Table 1.
  • Replace the normal convolution in atrous spatial pyramid pooling (ASPP) with depth-wise separable convolution. That is, all 3 × 3 convolutional layers are changed to 3 × 3 depth-wise separable convolutions, reducing the parameters of the network layer and speeding up the training efficiency with little impact on the running results.
  • In the decoding layer, the original algorithm is improved mainly by combining the outputs of different stages of the backbone residual network. Since the feature maps generated by each layer of the backbone network are critical to the final segmentation map, while the original segmentation network Deeplab v3+ uses only the high-resolution features of the first layer, i.e., it uses a quarter-sized feature map. In this paper, the feature maps output by the second and third layers are used. The size of the feature maps is 1/8, 1/16 and 1/16, respectively, and the number of channels is 512, 1024 and 2048, respectively. The output of each feature layer is upsampled by a bilinear interpolation operation, and the number of channels is reduced to 64 by 1 × 1 convolution. The output three feature maps are channel superimposed into a feature map of 1/4 size and 128 channels, which is passed through the attention mechanism module CBAM [36]. This makes it easier for the network to focus on the key locations where features are extracted and increases the network’s ability to extract edge features. The feature map processed by attention mechanism is superimposed with shallow features and deep features after quadruple up-sampling, the number of channels is adjusted to 256 by 3 × 3 convolution, and finally, the feature map is restored to the original image size by upsampling, and the feature map is divided into pre-defined categories by the classifier and output.
This approach helps to encode both global and local environments due to the use of features learned at multiple scales to enrich the representation of shallow features. In this paper, we borrow the idea of multi-scale self-guided attention network design for medical image segmentation proposed by Sinha et al. [37]. A new combinatorial network of multi-scale attention mechanisms is proposed, which is embedded in the decoding layer of the original Deeplab v3+ through the attention mechanism CBAM after the combination. Specifically, in our setting, features at multiple scales are denoted as F S , where S represents the number of layers where the feature map is located. Since the features of each layer have different resolutions, they are upsampled to the same resolution by bilinear interpolation, so that the output amplified feature map is denoted as F S and the corresponding S denotes the number of layers corresponding to the feature map. Then, without changing the original network structure, F 1 is output by 1 × 1 convolution, the subsequent F 2 , F 3 , and F 4 are connected to form a tensor (tensor), a multi-scale feature map is obtained by convolution, and then the output multi-scale feature map is passed through the CBAM module to obtain F M S , where F M S is shown in Equation (1):
F M S = C B A M ( C o n v ( F 2 + F 3 + F 4 ) )

2.1.2. Database Optimization

Using an improved semantic segmentation algorithm based on deep learning can segment semantic information from images and identify what different regions of the image represent. Buildings and road signs play an indispensable role in road image matching and localization. Relatively speaking, static and dynamic objects in complex scenes of road images, such as pedestrians, vehicles, trees, etc., contribute unstable or even almost meaningless to the matching and recognition effects of road images. The semantic segmentation approach is proposed to address this problem, and in this paper, its main contribution is to identify the effective feature regions in the image and to identify the portion of buildings and road signs in the image.
The filtering of the data is achieved by removing privacy-infringing, duplicate and unfocused images from the data through semantic segmentation and eliminating images where buildings and road signs make up too small a proportion of the images before the dataset is built. The specific method is as follows. The images are fed into the network for segmentation process to obtain the classification matrix of the images. The number of pixel points belonging to buildings and landmarks in the image is counted and compared with a fixed threshold. The comparison result is used to determine whether the image should be sieved or not. For a 2048 × 1080 image, the pixel threshold is 10,000 for buildings and 5000 for landmarks, and when the number of corresponding pixel points is less than the threshold, it is determined that the image has less content, and the image is excluded. Otherwise, it will be kept.

2.1.3. Destruction Area Dataset Construction

The damaged road condition studied in this paper refers to the road environment under various regional conditions. Landmarks are difficult to identify in images due to thick smoke, fog, or rust, later man-made damage to road signs and buildings on both sides of the road. Many of these regional conditions include highway sections, urban roads, and rural sections. In this case, ordinary image search and localization algorithms will face great challenges. In order to simulate the above-mentioned damaged area conditions under realistic conditions, this paper uses partial masks from the irregular mask dataset proposed in the paper by Nvidia [38] to constitute the damage condition dataset. The irregular mask dataset is generated by collecting random streaks and masks of arbitrarily shaped holes, the source of which is described in [39], using occlusion or de-occlusion mask estimation methods between two consecutive frames of the video result. This dataset has a total of six masks with different hole-to-area ratios of (0.01,0.1], (0.1,0.2], (0.2,0.3], (0.3,0.4], (0.4,0.5], and (0.5,0.6]. The mask images used in this paper are 48 images from these six classes of masks with different hole-to-area ratios selected randomly, with eight images selected from each class. The selected mask masks are distributed in the center and all around the periphery of the images, some of which are shown in Figure 3.
The damage condition dataset consists of the above irregular mask dataset, four kinds of mask maps simulating various masking factors in real environment and original images, where the mask maps include corrosion, light smoke, rust and smoke, etc. These mask maps are created by Photoshop software and saved in picture form by means of layers. The mask maps used in this paper are shown in Figure 4.
On the basis of the collected data set of a certain road section, it is damaged and processed. Road signs and buildings in the image are first identified using an improved semantic segmentation algorithm. These two types of information are not easily disturbed by seasonal and illumination factors, so they can be considered as valid information. The randomly generated “damage pollution” is used to cover the image, and the damage ratio of each group is calculated according to the area ratio of the occluded road signs and buildings to the whole image. Generate the damage image data set. The process of damage masking on the original image is shown in Figure 5. The part of the generated damaged geographical condition dataset is shown in Figure 6.

2.2. Coarse Geo-Location Based on Image Retrieval

In order to perform accurate geolocation calculations, coarse geolocation based on image retrieval is performed first in the online phase after optimizing the dataset using an improved semantic segmentation algorithm. The purpose of coarse localization by image retrieval is to find the best match for the query image.
In this paragraph, we will describe the process of image retrieval using our method. This includes using deep learning feature points for image retrieval. First, we input the images into the neural network. Reliable and repeatable feature points and descriptors are extracted by this algorithm. The feature point and descriptor extraction method used in this paper refers to the feature point detection method proposed by Jerome Revaud [40] et al. in 2019. The introduced model is shown in Figure 7. Among them, repeatable feature points can still be detected in various shooting angles, light, and seasonal changes, and reliable ones can make the descriptors match correctly. This feature point extraction method believes that descriptors should be learned only in regions with high confidence because good keypoints should not only be repeatable, but also discriminative. Therefore, in the learning process, the detection and description processes are seamlessly combined to learn, thereby improving the reliability of the descriptor because the damaged road image content targeted in this paper is subject to external interference and occlusion. Using content-based retrieval methods will reduce the retrieval accuracy in the cases addressed in this paper. Using this feature point detection method can adapt to the situation that the image content is changed due to damage and man-made damage under the damaged condition, and the damaged road section can also be considered as the situation where the content in the image is occluded. Using the introduced method can extract more reliable and effective feature points and descriptors in this case. Damaged images can be expressed more effectively. After that, the clustering method can be used to train the codebook, the features of each image can be accumulated with the nearest cluster center to obtain a feature matrix, and then a feature vector can be created from the matrix. Finally, the PCA dimension reduction and normalization operations are performed on the accumulated descriptors.
The feature extraction process in image retrieval is shown in Figure 7. The specific process is as follows.
  • Reliable and repeatable feature points and descriptors are extracted using a feature point extraction network. These descriptors are later clustered using the KMeans algorithm and the training codebook;
  • The features of each image are accumulated with the nearest cluster center;
  • PCA dimension reduction is performed on the accumulated feature vectors, and they are normalized;
  • A shortlist of the top-matched images is generated based on the feature vector;
  • The images in the shortlist are re-ranked by local descriptor matching using the Hamming distance and by checking the geometrical consistency using the distance ratio coherence.
The backbone network is L2-Net [41], and two modifications have been made. The first is to use extended convolution to replace the original down-sampling at the down-sampling place, so that the original resolution of the feature map is maintained at each stage. The second is to replace the final 8 × 8 convolution with three 2 × 2 convolutions.
The output of the backbone network is a 128-dimensional feature map, followed by three outputs. (1) The descriptor X of each pixel is obtained by L2 normalization; (2) S is obtained by a square operation, 1 × 1 convolution, and softmax; (3) R is obtained by the same operation as (2).
The three outputs of this network are as follows.
  • X H × W × D corresponds to the descriptor;
  • S [ 0 , 1 ] H × W corresponds to feature descriptor location (repetitiveness);
  • R [ 0 , 1 ] H × W corresponds to descriptor reliability.
The network outputs dense local descriptors and two associated repeatability and reliability confidence maps. One of them estimates keypoints to be repeatable, and the other estimates that their descriptors are separable. Finally, keypoints are taken from where the response is maximized for these two graphs.

2.3. Fine-Grained Geographic Alignment Implementation Based on Feature Matching

Our method usually uses the same camera with two images taken at two locations. Let the coordinates of the space point P in the world coordinate system be P. Because the world coordinate system and the camera coordinate system of the left view overlap, the coordinates are also P in the left camera coordinate system and RP + t in the right view camera coordinate system.
R and T are calculated from the essential matrix and fundamental matrix. The calculation method is shown in Equations (2) and (3).
E = t R
F = K T t R K 1
where K represents the camera’s internal parameter matrix, R represents the external parameter rotation matrix, and t represents the external parameter translation matrix.
After solving the above equation for F or E, and decomposing F and E to obtain R and t, the rotation matrix R and the translation vector t can be solved to find the distance of translation and the angle of rotation between the two images. Assume that the GNSS information of the first image is already known, such as latitude and longitude information and heading angle. The latitude and longitude information of the second image can be calculated by the distance of translation and the angle of rotation. The formula for calculating the latitude and longitude is as Equations (4) and (5).
φ 2 = arcsin ( sin φ 1 × cos δ + cos φ 1 × sin δ × cos θ )
λ 2 = λ 1 + arctan 2 ( sin θ × sin δ × cos φ 1 , cos δ sin φ 1 × sin δ 2 )
where φ is the latitude, λ is the longitude, θ is the azimuth (clockwise from north), δ is the angular distance d/R; d is the distance traveled, and R is the radius of the earth.

3. Experimental Results and Discussion

3.1. Dataset

For our geo-location experiments of damaged road images, we constructed and used a dataset consisting of urban, rural, and technopark segments. For each reference image in the database, a combined inertial guidance device was used to compute the velocity, heading angle and position information in the navigation coordinate system at the moment the image was taken, which can be used to compute GNSS information during the shooting process.
In this paper, the calculation method of the occlusion rate of the image is proposed as shown in Equation (6).
O c c = 1 C p i x e l C m a s k
C m a s k is the total number of pixels occupied by the object to be calculated (such as buildings, street signs) when it is not blocked and disturbed.
C p i x e l is the number of pixels of the object to be calculated in the image after partial occlusion.
Our test images generate 19 sets of damaged road images using our damaged image generation method, with the damage ratio of each set increasing by 5% of the total damage ratio as the number of sets increases. A total of 26,468 images of damaged roads ranging from 5% damage to 95% damage. We randomly select 400 images from each group as a test set of damaged images for each damage ratio. Additionally, our test images consist of 400 uncorrupted images.

3.2. Implementation Details

For the semantic segmentation algorithm, the stochastic gradient descent (SGD) optimization algorithm is used for training on the CityScapes dataset due to equipment performance limitations, with momentum set to 0.9, weight decay set to 0.0001, base learning rate set to 0.1, and learning rate decay using the “Poly” decay strategy is used for end-to-end training by back propagation. When training on the public dataset CityScapes, the maximum number of iteration steps is set to 50,000. We use a pre-trained model trained on ImageNet with PyConvResNet-50 to initialize our network weights and fine-tune them on this basis. In this way, the problems of gradient disappearance and gradient explosion can be effectively prevented, and the training speed and convergence speed of the network can be made faster, which saves the learning time and improves the learning efficiency.
Cross-entropy is used for the training loss function, which is mainly used to measure the variability between two probability distributions. The formula of the multiclassification cross-entropy loss function is described in Equation (7)
L = 1 N l c = 1 M y i c log ( p i c )
where M is the number of categories. y i c is the indicator variable. If the category and the category of sample i are the same, it is 1; otherwise, it is 0. p i c is the predicted probability that the observed sample i belongs to category c.
For the image retrieval algorithm, when we process the query image and the reference image, we select 1000 feature points and descriptors with repeatability and reliability for clustering and subsequent use to extract descriptors in order to improve computational efficiency.

3.3. Evaluation Measure

We declare that the latitude and longitude, actual location and estimated location in this article are in the BD-09 coordinate system used by Baidu, which is a coordinate system formed by encrypting and offsetting again on the basis of GCJ-02 coordinate system and is only applicable to Baidu maps. The true locations of these query images are the coordinates obtained by field mapping with the help of distinctive features (such as intersections and street lights) at the time the photos were taken. GNSS information acquired by smart devices is not used. This is because this information tends to produce significant errors in the positioning of urban buildings.
In this study, the validation data used are the latitude and longitude coordinates of the query image, and we use the total error to evaluate the accuracy of our proposed geo-location method by calculating the horizontal distance between the actual location and the estimated location. If the horizontal distance is smaller, the accuracy is higher.
In the related literature [42,43], an image is assumed to be correctly localized if one of the first K retrieved images is within d = 25 m from its true position error. At the same time, the percentage of correctly localized queries according to the top K candidates is a general evaluation measure from the literature.

3.4. Performance Analysis

In this section, we will list the experimental results for discussion and analysis. The final correct localization rate of all experiments is shown in Table 2. Metadata_SIFT represents the test set to perform experiments on datasets that have not been filtered for semantic segmentation. Coarse geolocation is performed using SIFT feature point retrieval. Deeplab v3+_SIFT represents the test set and conducts experiments on the dataset constructed by the segmentation and screening of the Deeplab v3+ algorithm. Coarse geolocation is performed using SIFT feature point retrieval. Ours_SIFT represents the test set for experiments on the dataset constructed by screening our improved semantic segmentation algorithm. Coarse geolocation is performed using SIFT feature point retrieval. Metadata_resnet represents the test set for experiments on datasets that have not been filtered for semantic segmentation. A feature vector database is built using resnet. Finally, the retrieval experiment is carried out using Euclidean distance comparison to complete the process of rough geolocation. Deeplab v3+_resnet and Ours_resnet represent pre-screened datasets using Deeplab v3+ and our method, respectively. After that, the experimental results of resnet feature vector retrieval of images are determined. Metadata_R2D2 represents the test set for experiments on datasets that have not been filtered for semantic segmentation. The R2D2 deep learning feature points are used to replace the traditional SIFT feature points for retrieval experiments to complete the experimental results of coarse geolocation. Deeplab v3+_R2D2 and Ours_R2D2 represent pre-screened datasets using Deeplab v3+ and our method, respectively. After that, the experimental results of R2D2 deep learning feature point retrieval image are carried out. Our proposed Ours_R2D2 method uses an improved semantic segmentation algorithm to segment the dataset. An image with better edge segmentation effect can be obtained, which is convenient for subsequent retrieval. Additionally, we use deep learning feature points instead of traditional hand-crafted feature points and descriptors. The time spent for single image retrieval and localization is shown in Table 3.

3.4.1. Comparison of Time Cost

The cost of time is one of the important indicators to measure the positioning method. The time for geo-location includes image retrieval time and precise geo-location calculation time. We list in the table below the time spent for localization using different retrieval algorithms. Each queried image includes image retrieval, matching and exact geo-location calculation. Additionally, the positioning time also includes judgment time, that is, calculating the distance between the estimated query image position and the actual position. We consider a correct localization result only when the distance is less than a threshold value of 25 m. Due to the computational complexity of our experiments and the large amount of data used in the experiments. This results in a longer operation time for the overall experiment, which causes the experimental equipment to overheat. The operation speed of the device will be reduced due to factors such as overheating. After many experiments, we found that the overall operation time of the first part of each experiment is relatively low. Therefore, our algorithm selects the average time of the first 100 images of each experiment as the single image processing time of the algorithm. The specific positioning time is shown in Table 4 below.
In our experiments, the neural network extracts feature vectors showing its advantage in terms of time cost. and its time overhead in applying image feature retrieval is low. The matching time without the method is shown in Table 4, and the time cost using SIFT feature points is shorter than our method. However, the overall situation is similar, the average time to extract deep learning feature points using our method is 0.025 s longer than that using traditional feature points, but our method is 14.8% higher in retrieval accuracy than the retrieval method implemented by traditional feature points and 13% higher than the accuracy of applying deep learning feature vector retrieval.

3.4.2. Effectiveness of Semantic Segmentation

Reduce Storage and Speed Up Retrieval

As shown in Table 5, we will list the screening of the dataset after the semantic segmentation method used and separately list the screening of the storage space of the original dataset.
As shown in Table 5, the dataset images are significantly reduced by 10% compared to the original dataset by Deeplab v3+. Furthermore, our improved method reduces dataset images by 18% compared to Deeplab v3+.
As shown in Figure 8, it can be concluded that as the number of images in the dataset increases, the final positioning time of several retrieval algorithms also increases. When using SIFT features and our deep learning feature points for localization, the time for retrieval and localization in the original database is significantly higher than the time for localization after semantic segmentation and screening of the dataset, while when using deep learning feature vectors for retrieval and localization, The time difference is not significant. This shows that reducing the number of datasets has a great effect on the retrieval and positioning time when using image feature points for retrieval and positioning tests. It can also illustrate that our method is more competitive in engineering applications at the metropolitan scale.

Effect on Positioning

The role of semantic segmentation in the whole localization process is to screen out images with too many interfering elements and less useful information, as shown in Figure 8. In Figure 8a–c, the retrieval effects of different retrieval methods on the pre-screened and post-screened datasets are basically similar, i.e., the method of reducing images with less useful information in the database by semantic segmentation has no or little effect on the retrieval and localization effects. It can be demonstrated that our method not only speeds up retrieval localization but also achieves better localization results without affecting the correct localization rate. These also prove the significance of our work.

3.4.3. Effectiveness of Improved Image Retrieval Algorithm

Reduce Storage

As shown in Table 6, the neural network method of extracting feature vectors can effectively reduce the size of stored feature files. It can also be seen that the SIFT feature index requires more storage space, and from this point of view, our proposed deep learning-based feature point retrieval method outperforms the hand-crafted feature point extraction method.

The Effect on the Search Effect

In recent years, in the field of image recognition and matching, deep learning features often outperform artificial features. In Table 2, the results using deep learning features achieve higher accuracy than the experiments retrieved without deep learning methods. In all experiments, the highest value of almost every correct rate is obtained by our improved algorithm ours_r2d2.
Figure 9, Figure 10 and Figure 11 show the retrieval results of the query image in detail.
In Figure 9 and Figure 11, the localization task of the corrupted images is very difficult, and the top 5 retrieval results of SIFT and Resnet do not or only a few complete the localization task. However, after using our method, the retrieval of images was completed with correct results.
In Figure 9, Figure 10 and Figure 11, the result image of SIFT is not very similar to the original image. Additionally, in all search results, the correct result will be displayed in the top image list. According to the failure of SIFT feature retrieval caused by some damages in the figure, it shows that the damage area has great interference to SIFT features. This demonstrates the success of our adoption of deep learning feature points. The correct rate of retrieval results also proves the effectiveness of using deep learning feature points in adversarial damaged region retrieval.
This section may be divided by subheadings. It should provide a concise and precise description of the experimental results, their interpretation, as well as the experimental conclusions that can be drawn.

4. Conclusions

The goal of this paper is to perform geo-location in damaged road conditions. When dealing with the image-based localization task of damaged road sections, the existing methods do not specifically optimize the damaged road images in the feature extraction stage. Such complex scenes after buildings or road signs are damaged are not conducive to feature extraction. If the features extracted from the damaged image are used for matching, these features falling on meaningless objects will also affect the accuracy of the localization results. For this purpose, this paper describes a method for geo-location of damaged road images, which can achieve geo-location from non-geo-tagged to geo-tagged images in urban damaged areas. The method uses a variety of roadway images acquired by itself as the reference dataset constructs disturbed images using an improved semantic segmentation algorithm. An improved semantic segmentation algorithm is used to construct the disturbed images and use them as the damage dataset. The image localization task is accomplished by first rough geo-localization through image retrieval, and then fine geographic alignment through feature matching, thus adding accurate geotags to the damaged images. The main contribution of this paper is that the method has higher correctness and geo-localization speed than previous methods by using current road images and combining retrieval and matching algorithms, especially for road areas under damage and destruction conditions. Through experiments on the dataset, we fully demonstrate the effectiveness of the proposed algorithm and process. Compared with previous methods, our method improves the retrieval accuracy. Under the error requirement of 25 m, the correct rate of geolocation query images for damaged road images is higher than that of traditional algorithms. Our approach is very practical due to the metropolitan scale and data diversity of the datasets we use.

Author Contributions

Conceptualization, W.W. and W.Z.; methodology, W.Z.; software, W.Z.; validation, W.W., W.Z. and J.Q.; formal analysis, W.Z.; investigation, W.Z.; resources, W.W.; data curation, W.Z.; writing—original draft preparation, W.Z.; writing—review and editing, W.Z.; visualization, J.H.; supervision, J.L.; project administration, W.Z.; funding acquisition, J.Q. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the National Natural Science Foundation of China (grant number 52,175,282) Engineering University, Xi’an 710051, Shaanxi, China.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Y.; Ding, A.Y.; Ott, J.; Yuan, M.; Zeng, J.; Zhang, K.; Rao, W. Transfer learning-based outdoor position recovery with Cellular Data. IEEE Trans. Mob. Comput. 2021, 20, 2094–2110. [Google Scholar] [CrossRef]
  2. Zhang, Z.; Tao, W.; Sun, K.; Hu, W.; Yao, L. Transfer learning-based outdoor position recovery with Cellular Data. Pattern Recognit. 2016, 60, 227–238. [Google Scholar] [CrossRef]
  3. Large, N.L.; Bieder, F.; Lauer, M. Comparison of different slam approaches for a driverless race car. Tm Tech. Mess. 2021, 88, 227–236. [Google Scholar] [CrossRef]
  4. Ren, J.; Wu, T.; Zhou, X.; Yang, C.; Sun, J.; Li, M.; Jiang, H.; Zhang, A. SLAM, Path Planning Algorithm and Application Researchof an Indoor Substation Wheeled Robot Navigation System. Electronics 2022, 11, 1838. [Google Scholar] [CrossRef]
  5. Zhi, L.; Xiao, Z.; Qiang, Y.; Qian, L. Street-level image localization based on building-aware features via patch-region retrieval under Metropolitan-scale. Remote Sens. 2021, 13, 4876. [Google Scholar] [CrossRef]
  6. Yadav, R.; Kala, R. Fusion of visual odometry and place recognition for slam in extreme conditions. Appl. Intell. 2022, 52, 1–20. [Google Scholar] [CrossRef]
  7. Rong, D.; Xie, L.; Ying, Y. Computer vision detection of foreign objects in walnuts using deep learning. Comput. Electron. Agric. 2019, 162, 1001–1010. [Google Scholar] [CrossRef]
  8. White, J.; Kameneva, T.; McCarthy, C. Vision processing for assistive vision: A deep reinforcement learning approach. IEEE Trans. Hum. Mach. Syst. 2022, 52, 123–133. [Google Scholar] [CrossRef]
  9. Xue, B.; He, Y.; Jing, F.; Ren, Y.; Jiao, L.; Huang, Y. Robot target recognition using Deep Federated Learning. Int. J. Intell. Syst. 2021, 36, 7754–7769. [Google Scholar] [CrossRef]
  10. Amit, D.; Shah, N.; Adhikari, P.; Kumbhar, S.; Dhanjal, I.S.; Mehendale, N. Firefighting robot with Deep Learning and Machine Vision. Neural Comput. Appl. 2021, 34, 2831–2839. [Google Scholar]
  11. Oliva, A.; Torralba, A. Chapter 2 building the Gist of a scene: The role of Global Image Features in recognition. Prog. Brain Res. 2006, 155, 23–36. [Google Scholar] [PubMed]
  12. Lowry, S.; Wyeth, G.; Milford, M. Unsupervised online learning of condition-invariant images for place recognition. Procedia Soc. Behav. Sci. 2014, 106, 1418–1427. [Google Scholar]
  13. Ulrich, I.; Nourbakhsh, I. Appearance-based place recognition for topological localization. In Proceedings of the 2000 ICRA. Millennium Conference, IEEE International Conference on Robotics and Automation, Symposia Proceedings (Cat. No.00CH37065), San Francisco, CA, USA, 24–28 April 2000. [Google Scholar]
  14. Sunderhauf, N.; Protzel, P. Brief-gist—Closing the loop by simple means. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011. [Google Scholar]
  15. Lowe, D.G. Distinctive Image Feature from Scale-Invariant Key points. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  16. Biswas, B.; Ghosh, S.K.; Hore, M.; Ghosh, A. SIFT-based visual tracking using optical flow and belief propagation algorithm. Comput. J. 2020, 65, 1–17. [Google Scholar] [CrossRef]
  17. Se, S.; Lowe, D.; Little, J. Mobile robot localization and mapping with uncertainty using scale-invariant visual landmarks. Int. J. Robot. Res. 2002, 21, 735–758. [Google Scholar] [CrossRef]
  18. Stumm, E.; Mei, C.; Lacroix, S. Probabilistic Place Recognition with Covisibility Maps. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013. [Google Scholar]
  19. Bay, H.; Tuytelaars, T.; Gool, L.V. SURF: Speeded up robust features. In Proceedings of the 9th European Conference on Computer Vision—Volume Part I; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  20. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. Orb: An efficient alternative to SIFT or surf. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  22. Zitnick, C.L.; Dollar, P. Edge Boxes: Locating Object Proposals from Edges. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; pp. 391–405. [Google Scholar]
  23. Mei, C.; Sibley, G.; Newman, P. Closing loops without places. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010. [Google Scholar]
  24. Ma, C.; Guo, W.; Zhang, H.; Samuel, O.W.; Ji, X.; Xu, L.; Li, G.A. A novel and efficient feature extraction method for deep learning based continuous estimation. IEEE Robot. Autom. Lett. 2021, 6, 7341–7348. [Google Scholar] [CrossRef]
  25. Jayalaxmi, P.; Saha, R.; Kumar, G.; Kim, T.H. Machine and deep learning amalgamation for feature extraction in industrial internet-of-things. Comput. Electr. Eng. 2022, 97, 107610. [Google Scholar] [CrossRef]
  26. Xu, C.; Zhu, G.; Shu, J. A combination of lie group machine learning and Deep Learning for remote sensing scene classification using multi-layer heterogeneous feature extraction and fusion. Remote Sens. 2022, 14, 1445. [Google Scholar] [CrossRef]
  27. Apostolopoulos, I.D.; Tzani, M.A. Industrial Object and defect recognition utilizing multilevel feature extraction from industrial scenes with Deep Learning Approach. J. Ambient. Intell. Humaniz. Comput. 2022. [Google Scholar] [CrossRef]
  28. Zamir, A.R.; Hakeem, A.; Gool, L.V.; Shah, M.; Szeliski, R. Introduction to large-scale visual geo-localization. In Large-Scale Visual Geo-Localization; Springer: Berlin/Heidelberg, Germany, 2016; pp. 1–18. [Google Scholar]
  29. Zhang, X.; Wang, L.; Su, Y. Visual Place Recognition: A Survey from Deep Learning Perspective. Pattern Recognit. 2020, 113, 107760. [Google Scholar] [CrossRef]
  30. Cummins, M.; Newman, P. FAB-MAP: Probabilistic Localization and Mapping in the Space of Appearance. Int. J. Robot. Res. 2008, 27, 647–665. [Google Scholar] [CrossRef]
  31. Angeli, A.; Doncieux, S.; Meyer, J.A.; Filliat, D. Incremental vision-based topological SLAM. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008. [Google Scholar]
  32. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for Semantic Image segmentation. In Proceedings of the Computer Vision–ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 833–851. [Google Scholar] [CrossRef]
  33. Zhang, W.; Wei, W.; Jue, Q.; Hu, J.; Wang, Q.-L. Improved Deeplab v3+ Image Semantic Segmentation Algorithm Fusion Multi-scale Features. Electrooptics Control. 2022. Available online: https://cf.cnki.net/kcms/detail/detail.aspx?filename=DGKQ2022071100G&dbcode=XWCJ&dbname=XWCTLKCAPJLAST&v= (accessed on 20 July 2022).
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  35. Duta, I.C.; Liu, L.; Zhu, F.; Shao, L. Pyramidal Convolution: Rethinking Convolutional Neural Networks for Visual Recognition. 2020. Available online: https://www.paepper.com/blog/posts/pyramidal-convolution-rethinking-convolutional-neural-networks-for-visual-recognition/ (accessed on 29 November 2020).
  36. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision–ECCV, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  37. Sinha, A.; Dolz, J. Multi-scale self-guided attention for medical image segmentation. IEEE J. Biomed. Health Inform. 2021, 25, 121–130. [Google Scholar] [CrossRef] [PubMed]
  38. Liu, G.; Reda, F.A.; Shih, K.J.; Wang, T.C.; Tao, A.; Catanzaro, B. Image inpainting for irregular holes using partial convolutions. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 89–105. [Google Scholar]
  39. Sundaram, N.; Brox, T.; Keutzer, K. Dense point trajectories by GPU-accelerated large displacement optical flow. In Proceedings of the Computer Vision–ECCV 2010, Crete, Greece, 5–11 September 2010; pp. 438–451. [Google Scholar]
  40. Revaud, J.; Weinzaep, P.F.L.; Souza, C.D.; Pion, N.; Humenberger, M. R2D2: Reliable and Repeatable Detectors and Descriptors for Joint Sparse Keypoint Detection and Local Feature Extraction. arXiv 2019, arXiv:1906.06195. [Google Scholar]
  41. Tian, Y.; Fan, B.; Wu, F. L2-net: Deep learning of discriminative patch descriptor in euclidean space. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  42. Torii, A.; Arandjelovic, R.; Sivic, J.; Okutomi, M.; Pajdla, T. 24/7 place recognition by View Synthesis. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  43. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN architecture for weakly supervised place recogni-tion. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
Figure 1. Proposed visual place recognition pipeline.
Figure 1. Proposed visual place recognition pipeline.
Electronics 11 02530 g001
Figure 2. Improved Deeplab v3+ network structure diagram.
Figure 2. Improved Deeplab v3+ network structure diagram.
Electronics 11 02530 g002
Figure 3. Partially irregular mask dataset used.
Figure 3. Partially irregular mask dataset used.
Electronics 11 02530 g003
Figure 4. Simulate the real environment mask map.
Figure 4. Simulate the real environment mask map.
Electronics 11 02530 g004
Figure 5. Damaged image generation process diagram.
Figure 5. Damaged image generation process diagram.
Electronics 11 02530 g005
Figure 6. Partially damaged area data.
Figure 6. Partially damaged area data.
Electronics 11 02530 g006
Figure 7. Schematic diagram of feature extraction process in image retrieval.
Figure 7. Schematic diagram of feature extraction process in image retrieval.
Electronics 11 02530 g007
Figure 8. The positioning accuracy and retrieval time of images with different damage rates using different retrieval algorithms. Figure (a) represents the time and accuracy of image retrieval and localization using SIFT features, Metadata represent the original data set composed of data without semantic segmentation and screening, and Deeplab represents the data set composed of filtered data using the original Deeplab v3+ algorithm, Ours represents a dataset composed of data filtered using an improved semantic segmentation algorithm, Figure (b) represents the time and accuracy of image retrieval and localization using Resnet features, and Figure (c) represents the use of R2D2 deep learning feature points for images and the time and accuracy of retrieval and location.
Figure 8. The positioning accuracy and retrieval time of images with different damage rates using different retrieval algorithms. Figure (a) represents the time and accuracy of image retrieval and localization using SIFT features, Metadata represent the original data set composed of data without semantic segmentation and screening, and Deeplab represents the data set composed of filtered data using the original Deeplab v3+ algorithm, Ours represents a dataset composed of data filtered using an improved semantic segmentation algorithm, Figure (b) represents the time and accuracy of image retrieval and localization using Resnet features, and Figure (c) represents the use of R2D2 deep learning feature points for images and the time and accuracy of retrieval and location.
Electronics 11 02530 g008aElectronics 11 02530 g008b
Figure 9. Top5 retrieval results of the database using the same query image after filtering the dataset by our algorithm.
Figure 9. Top5 retrieval results of the database using the same query image after filtering the dataset by our algorithm.
Electronics 11 02530 g009
Figure 10. Top 5 retrieval results of the database using the same query image after filtering the dataset with the original semantic segmentation algorithm.
Figure 10. Top 5 retrieval results of the database using the same query image after filtering the dataset with the original semantic segmentation algorithm.
Electronics 11 02530 g010
Figure 11. Top 5 retrieval results in unfiltered dataset using the same query image.
Figure 11. Top 5 retrieval results in unfiltered dataset using the same query image.
Electronics 11 02530 g011
Table 1. Parameter table of each layer of pyconv.
Table 1. Parameter table of each layer of pyconv.
LayersPyconv KernelsPyconv Kernel Output ChanelsPyconv_Groups
Layer1(3,5,7,9)(64,64,64,64)(1,4,8,16)
Layer2(3,5,7)(128,128,256)(1,4,8)
Layer3(3,5)(512,512)(1,4)
Layer4(3)(2048)(1)
Table 2. Correct rates of images location with different damage rates.
Table 2. Correct rates of images location with different damage rates.
Damage RatesMetadata_SIFTDeeplab v3+_SIFTOurs_SIFTMetadata_ResnetDeeplab v3+_ResnetOurs_ResnetMetadata_R2D2Deeplab v3+_R2D2Ours_R2D2
00.820.8040.7720.7440.7620.7960.8920.8980.898
50.7840.7220.7320.6460.6960.7340.8920.9220.936
100.7520.6630.7560.6080.6840.7160.8820.9420.956
150.6980.6960.720.6720.6720.7260.8740.9660.968
200.7260.7020.7380.6760.6520.6880.8580.9260.97
250.7460.6980.7360.6320.6620.6760.8640.9160.926
300.7520.7320.7280.6260.6860.6960.8960.8990.896
350.7520.720.7520.6180.6420.6620.8760.9060.926
400.7400.7280.740.6060.6260.6580.8860.9120.914
450.7320.7260.7380.6320.6360.650.8840.8780.894
500.7620.7160.7360.6260.6960.6860.8860.8360.862
550.7020.7120.7240.5960.6340.6460.8780.8660.848
600.7100.7460.7360.6120.6180.6580.8920.8620.896
650.7020.7340.7260.6160.6060.6220.8760.8620.876
700.70.7160.7220.6320.6040.6340.8920.840.854
750.7220.6420.6880.5880.5980.650.8620.8060.846
800.6120.6320.6460.5220.640.6360.8120.8560.874
850.60.5940.6120.5260.6120.6240.8060.7980.862
900.6620.6520.6960.4620.6360.6620.8760.9320.95
950.6420.6360.6880.4960.6560.6220.8720.9620.968
Table 3. Positioning takes time.
Table 3. Positioning takes time.
Damage RatesMetadata_SIFTDeeplab v3+_SIFTOurs_SIFTMetadata_ResnetDeeplab v3+_ResnetOurs_ResnetMetadata_R2D2Deeplab v3+_R2D2Ours_R2D2
01.3951.3251.1951.7821.6671.6131.8011.5661.373
51.4371.3381.1661.7331.6831.6141.7661.6081.402
101.4551.3351.1721.7081.6781.6211.7791.6031.374
151.4351.3271.1821.7311.6811.6191.7751.6521.379
201.4331.3281.1821.7571.6611.6281.7431.6291.401
251.4291.3321.1841.7471.6721.6271.7531.6241.375
301.4631.3541.1751.731.6691.6031.7641.5951.351
351.4711.3271.1771.6841.6511.6221.7521.5831.399
401.4841.3381.1741.731.6691.6211.7781.5821.375
451.4611.3441.1891.7211.6761.6281.7761.6161.362
501.4371.3031.1811.7351.6641.6431.7621.5941.353
551.4431.3121.1731.7461.6731.6531.7971.6081.377
601.4631.3211.1661.7321.7051.6481.7741.6151.351
651.4691.3391.1951.7181.7011.6571.7311.5891.384
701.4511.3311.1621.7421.6671.6331.7391.5871.368
751.4621.2961.1651.7331.7041.6161.7441.531.357
801.4561.3421.1561.7171.6741.6411.7331.5841.371
851.4641.3181.1611.7181.6941.6281.7781.5921.373
901.4351.3121.1421.7331.6841.5941.7671.5621.388
951.4141.3061.1651.7511.6911.6361.7851.5811.367
Table 4. Time consumption of different features.
Table 4. Time consumption of different features.
FeatureTime of Feature ExtractionTime of Image RetrievalPrecise Geo-Location Calculation
SIFT0.3230.5700.775
Resnet0.0320.0760.793
Ours0.3480.5760.923
Table 5. Reduction in storage space of original dataset after semantic segmentation.
Table 5. Reduction in storage space of original dataset after semantic segmentation.
FeatureNumber of ImagesFilter Rate
Metadata11,896--
Deeplab_seg10,7360.10
Ours_seg86200.28
Table 6. Memory size occupied by image retrieval and storage feature files.
Table 6. Memory size occupied by image retrieval and storage feature files.
DatasetMetadataDeeplab_segOurs_seg
Feature
SIFT Feature918 MB807 MB675 MB
Resnet94.6 MB82.9 MB68.5 MB
R2D2 Feature790 MB693 MB572 MB
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, W.; Qu, J.; Wang, W.; Hu, J.; Li, J. Geo-Location Method for Images of Damaged Roads. Electronics 2022, 11, 2530. https://doi.org/10.3390/electronics11162530

AMA Style

Zhang W, Qu J, Wang W, Hu J, Li J. Geo-Location Method for Images of Damaged Roads. Electronics. 2022; 11(16):2530. https://doi.org/10.3390/electronics11162530

Chicago/Turabian Style

Zhang, Wenbo, Jue Qu, Wei Wang, Jun Hu, and Jie Li. 2022. "Geo-Location Method for Images of Damaged Roads" Electronics 11, no. 16: 2530. https://doi.org/10.3390/electronics11162530

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop