Next Article in Journal
A Multi-Platform Optical Sensor for In Vivo and In Vitro Algae Classification
Next Article in Special Issue
Dynamics of Hierarchical Urban Green Space Patches and Implications for Management Policy
Previous Article in Journal
Investigating Surface and Near-Surface Bushfire Fuel Attributes: A Comparison between Visual Assessments and Image-Based Point Clouds
Previous Article in Special Issue
Downscaling Land Surface Temperature in Complex Regions by Using Multiple Scale Factors with Adaptive Thresholds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rapid Texture Optimization of Three-Dimensional Urban Model Based on Oblique Images

1
State Key Laboratory of Information Engineering in Surveying Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
2
Collaborative Innovation Center for Geospatial Technology, Wuhan University, Wuhan 430079, China
*
Authors to whom correspondence should be addressed.
Sensors 2017, 17(4), 911; https://doi.org/10.3390/s17040911
Submission received: 10 March 2017 / Revised: 12 April 2017 / Accepted: 14 April 2017 / Published: 20 April 2017

Abstract

:
Seamless texture mapping is one of the key technologies for photorealistic 3D texture reconstruction. In this paper, a method of rapid texture optimization of 3D urban reconstruction based on oblique images is proposed aiming at the existence of texture fragments, seams, and inconsistency of color in urban 3D texture mapping based on low-altitude oblique images. First, we explore implementing radiation correction on the experimental images with a radiation procession algorithm. Then, an efficient occlusion detection algorithm based on OpenGL is proposed according to the mapping relation between the terrain triangular mesh surface and the images to implement the occlusion detection of the visible texture on the triangular facets as well as create a list of visible images. Finally, a texture clustering algorithm is put forward based on Markov Random Field utilizing the inherent attributes of the images and solve the energy function minimization by Graph-Cuts. The experimental results display that the method is capable of decreasing the existence of texture fragments, seams, and inconsistency of color in the 3D texture model reconstruction.

1. Introduction

Low-altitude oblique images can be photographed from the side and the top of the city, which can not only provide a precise 3D surface model, but also obtain reliable texture information, becoming one of the principal data resources in urban 3D texture reconstruction and updating. Now, some reliable methods in the reconstruction of urban 3D surface models from oblique images have been proposed [1,2,3,4,5], especially for the commercialization of some software like Smart3D Capture. However, problems existed in low-altitude oblique images—such as great changes in field of view, large volume of data, inconsistency of light, perspective distortion, and occlusion—will the vast texture fragments and the discontinuity of color common, which needs to be solved. Research about these problems has been carried out in numerous of papers. Take the vertex rendering method epidemic in the early days, [6,7] assign the vertexes of the triangular by rear projection after color components of RGB of the aerial images. The texture generated lost great details with low resolution. References [8,9] utilized the method of weighting different images to generate texture, which did not work well with satisfactory consistency of the color of texture in many cases. In order to get a superior solution of these problems, many studies started to address how to choose the priority images as the texture resource for seamless texture mapping in a triangular mesh model. References [10,11] chose the priority image for each triangular facet as the texture resource by taking the visibility condition between the images and 3D surface model in to consideration. References [12,13] attempted to choose the highest quality image for every triangular facet as its texture after matching the corresponding line segments in the images and 3D surface model and making overall consideration of image occlusion, resolution, as well as the normal vector of the model. Reference [14] selected the priority texture for indoor small objects with close shots utilizing the Markov random field (MRF) method after overall optimization which obtained good results. Nevertheless, neither of these could address the problem of texture fragments and inconsistency of color in big urban-scene texture mapping. Moreover, the methods above require excessive computer resources resulting in low efficiency. On account of these problems, a method of rapid texture optimization will be presented in this paper to accelerate and improve the quality of seamless 3D urban texture reconstruction. In this paper, the experimental oblique images have been corrected for distortion, we will adopt a radiation procession algorithm to preprocess the oblique images first; and then, an efficient occlusion detection algorithm will be proposed for creating a priority list of visible images; at last, a Graph-Cuts based texture clustering and selecting algorithm will be put forward.

2. 3D Urban Surface Model

In the paper, the procedure of 3D urban terrain surface model reconstruction is as follows. First, extracting and matching feature points in the multi-view oblique images utilizing SIFT (Scale Invariant Feature Transform) [3] is combined with calculating the geometrical relationship between the multiple views. Then, the matching result by SFM (Structure from Motion) [1] is calculated to get scene information such as the camera position and then reconstruct sparse by the obtained 3D point cloud. After that, we combine the CMVS (Cluster Multi-View Stereo) and PMVS (Patch-based Multi-View Stereo) technique to acquire a reconstructed 3D dense point cloud [4,5]. The result is a constructed 3D urban surface model (expressed by Triangulated Irregular Network) for texture optimization and mapping. Figure 1 shows a local 3D model of the experimental area.

3. Image Radiation Preprocessing

On account of the time difference, environmental implications, and camera system error existing when shooting in multi-camera oblique photograph systems, there is a difference in brightness and color in different images shot on the same object. If the difference is left alone, it could result in color jump when reconstructing texture. Therefore, in this paper, a new algorithm is proposed which can correct the radiation inconsistency in multi-view images. In our algorithm, the local and overall environmental impacting factors are adjusted and the systemic error between and inside the camera is corrected, which could recover the mean value and the variance of the image to the initial radiation brightness value of the scene as far as possible and improve the brightness and color of the image used for texture extraction.

3.1. Correction of Environmental Radiation Error

The uniform color and light mechanism provided by Wallis filter is a relative radiation correction algorithm [15,16]. The algorithm can remove environmental radiation errors caused by global multi view by adjusting the mean value and variance of the image, which obtains a considerable result in uniform color and bright in aerial images. Formula (1) [15,16] is Wallis’ mapping function. However, during the processing of oblique photographs, on account of the vast gap of shooting distance on the oblique surface, a resultant effect of white mist appeared in some images caused by the inconsistency of atmospheric influence. As for the haze phenomena which has shrouded the cities of China for a long time; a superior result is not attainable by Wallis’ mapping function alone. Therefore, in this paper, the algorithm based on the dark channel in [17] is brought in to remove the phenomena of white mist caused by the inconsistency of depth. The experiment proved that the environmental radiation problem could be solved from the depth information of the images. Formula (2) [17] is the recovery function of the scene based on dark channel.
f ( x , y ) = [ g ( x , y ) m g ] c s f c s g + ( 1 c ) c s f + b m f + ( 1 b ) m g
J ( x , y ) = I ( x , y ) A max [ t 0 , t ( x , y ) ] + A
In Formula (1), f(x,y) is the image after conversion and g(x,y) is the raw image. sf, sg and mf, mg are the standard deviation and mean value before and after conversion respectively. c ∈ [0, 1], which is the expansion constant of the variance of the image. b ∈ [0, 1], which is the brightness coefficient of the image. In Formula (2), J(x,y) is the image under ideal consideration; A is the intensity of atmospheric light; t(x,y) is the transmittance of transmission medium; t0 is the threshold of transmission.

3.2. Correction of Camera System Error

After the processing above in Section 3.1, the effect of environmental factor in different images is removed fundamentally, recovering their radiance to the level of a single image. Nevertheless, on account of the existing of camera system error, the phenomenon of the difference in brightness and color still appears in different images photographing the same object, which needs correcting. Generally, camera system error could be divided into exposure factor, white balance coefficient, camera response function, and the influence of halo. In the paper, other than the rest algorithms that only the camera response function taken into consideration to correct the camera system error, our proposed generalization camera response function including halo function, white balance coefficient, and exposure factor to correct the mapping from scene radiance to brightness value for color images. Formula (3) [18] is the mapping relation of the generalized camera response estimation function in the paper.
g ( B 1 ) w 1 i e 1 M ( x 1 ) = g ( B 2 ) w 2 i e 2 M ( x 2 )
In Formula (3), g(·) represents the inverse function of camera response function; B represents the brightness value of the image; e represents the exposure factor; M represents the halo function; wi represents the white balance coefficient in different channel; and i represents different channel.

4. Establishing a Visible Images List

Before the texture clustering and selection processed, it is required to know every triangular facet on the 3D surface model is visible in which images. Therefore, a list of visible images for each triangular facet in a 3D surface model needs to be established. In the paper, occlusion detection of triangular facets is implemented at first; then, according to the idea of texture filtering in [19], the occlusion area and normal vector are used as the evaluation criteria and a priority list of visible images is created.

4.1. OpenGL-Based Occlusion Detection

Different from the geometric model obtained by scanning a Lidar point cloud, it is uncertain to reconstruct all the surface features when utilizing oblique-image-based 3D reconstruction, which could lead to the inconsistency of occlusion between images and the 3D surface model. Therefore, algorithms of occlusion detection analogous to Z-buffer do not work well. In order to solve this problem, according to the transformation matrix of OpenGL in relation with the interior and exterior orientation elements in photogrammetry in [20], we completed the transformation matrix expression of OpenGL. Achieving rapid, efficient, and precise occlusion detection.
[ u v 1 ] Τ = V 3 × 4 · P 4 × 4 · M 4 × 4 [ X Y Z 1 ] Τ
Figure 2 shows three basic transformation of the procedure when OpenGL display 3D objects in window, they are model transformation (M), projection transformation (P), and viewport transformation (V). The object space coordinate system can be transformed into the screen coordinate system through these three transformation matrixes. The mathematical expressions of the matrixes are as seen in Formula (4). Formulas (5)–(7) are the specific expression of M, P, and V respectively, in which the model transformation equals the rotation and translation of the exterior orientation elements, R and Ts represent the rotation matrix and translation matrix, respectively. Projection transformation equals to 3D cutting + similarity transformation + affine transformation, in which (f,x0,y0) is the interior orientation elements and (w,h) is the width and height of the camera. Viewport transformation is equal to 2D Euclidean transformation.
M = [ R T R T T s 0 1 ]
P = [ 2 f w x 0 w 2 f h y 0 h F + N F N 2 F N F N 1 ]
V = [ w 2 0 0 w 2 + x 0 0 h 2 0 h 2 + y 0 0 0 0 1 ]

4.2. Adaptive Partial Occluded Triangular Facet Subdivision

After occlusion detection, the images can be divided into three types for each triangular facet: fully occluded images, partially occluded images, and fully visible images. The fully occluded images are removed. However, if we do not process the triangular facets we only have partially occluded images, the texture inconsistencies might appear because of the 3D triangular mesh surface model subdivision is insufficient. In order to solve the problem and improve the occupancy of the non-occluded information in a single triangular facet, a local quartering algorithm is applied in adaptive partial occluded triangular facet subdivision in the paper. The new subdivision algorithm only partially occluded triangular facets and dichotomized the adjacent triangular facets (as Figure 3). In this way, not only the efficient of partial occluded triangular facets subdivision can be guaranteed, it can also decrease the time consumption and data size of segmenting non-occluded triangular facets. At the same time, it can also guarantee the invariability of half-edge data structure and the quality of manifold surfaces of the net form on the 3D surface model.
The shadow area in Figure 3 is the triangular facet which is partially occluded. Figure 3c shows the subdivision result of two different algorithms. Compared with the result of global quartering in Figure 3b, only in the worst situation, the number of triangular facets in Figure 3c is equal to Figure 3b. Generally, the amount of the triangular produced is 4a + 6a by our algorithm, which is smaller than 4b. a is the number of partially occluded triangular facets, b is the number of all of the triangular facets.

4.3. Priority Principle of Visible Images

In the paper, we establish the visible images priority list following three principles below: (1) area maximum principle. That is to say, the larger area of the image on the triangular facet, the better; (2) angle minimum principle. In a non-occluded image, image deformation is proportional to the angle between photography light and terrain normal n. As shown in Figure 4, for surface abcd, image O1 is occlusion image, for abcd surface in the cube, image O3 has higher priority than image O2.

5. Texture Clustering and Selection Based on Graph-Cuts

Selecting the texture with the highest priority as the texture source for every triangular facet in the established priority list of visible images often cannot meet the global optimization condition. Therefore, how to choose the optimal image in the priority list of visible images as the texture mapping resource is still an intractable problem. In computer vision, a discrete multi-label problem like texture mapping can be regarded as a typical MRF model and Graph-Cuts is one of the most efficient global optimization tools to solve the energy function [21,22]. Thus, the Graph-Cuts based texture clustering algorithm under the frame of MRF is proposed to solve the problem of texture selection optimization in the paper.

5.1. Label Graph Construction

Graphing is an intuitive representation of the real word which is composed be a point set V and a set of sides connected between the nodes under a certain rule indicated as G = <V, E>. In this paper, each triangular facet in a 3D triangular mesh surface model is regarded as a node in graph G. Every two adjacent triangles with a common side respond an edge constructed between the nodes in graph G. In this way, we finish constructing the graph that we need. Graph-Cuts is a subset in set E of graph G. Graph-Cuts is a two-labeling problem, which could link the minimum of the energy function with the minimum cut of an image containing source and sink. As seen in the labeled graph constructed in this paper, not all the images could be the candidate texture resource to some certain triangular facets, as every triangular facet may be occluded in some image. That is to say, the mapping relation between triangular facets and images is not a complete graph, which needs to be processed.
As shown in Figure 5, Figure 5a is a general label graph, in which the top and bottom special red terminal nodes are the source s and sink t in the graph, respectively. The rest nodes are different labeled point to each triangular facet. The nodes are connected by side t-link and side n-link which is measured the probabilistic relation of Energy weight. t-link connects the sides of s,t, and other points which contains the energy between different labels. n-link connects the sides of middle points which contains the energy relation between adjacent nodes. The relations mainly reflect the probability of selecting the same texture label between the adjacent nodes. All the nodes constitute label space. Moreover, the surfaces in the same label layer have the same texture source. Figure 5b is the labeled graph constructed in the paper. Compared with Figure 5a, some green blank nodes appear, which means that texture nodes are undesirable in the texture space of this layer. Therefore, all the links between the nodes need to avoid the green blank nodes so that the data size of Graph-Cuts algorithm involved will be decreased and the overall efficiency of the algorithm can be improved.

5.2. Energy Function Construction

The energy function is the bridge between Graph-Cuts theory and a specific problem, it consists of two parts: data item and smooth item, which have one-to-one correspondence to weight of Graph-Cuts sides. The minimum cut in Graph-Cuts equals to the minimum energy function. Formula (8) is the functional expression of the energy function in the paper. In graph G, each node has a label set that contains k factors. The optimization algorithm selects the optimal label in the label set for each node. To guarantee calculation in the valid real range by the Graph-Cuts algorithm and the ability to process big data, the energy function construction needs to meet the following principles: (A) The energy function needs to possess upper and lower bounds; (B) The value range of the energy function needs to be no more than limit times of the number of the data calculated; (C) The data item, smoothed value, and fluctuating value of the energy are supposed to be in the same order of magnitude. In order to accurately reflect the identity of the corresponding texture resources in adjacent triangular facets, we constructed the data item and smooth item for energy functions using the inherent attributes of the candidate images, to achieve selecting optimization and clustering for textures in adjacent triangular facets.
E ( L ) = α D ( L ) + β N ( L )
M = { m i | m i = A V E R A G E ( I k ) } , i [ 0 , n 1 ] , k [ 0 , O 1 ]
D = { d i | d i = S T D E V   ( I k ) } , i [ 0 , n 1 ] , k [ 0 , O 1 ]
F = { f i | i [ 0 , n 1 ] }
L = { l i | l i [ 0 , O 1 ] , i [ 0 , n 1 ] }
In Formula (8), α and β are the control coefficients of the data item D(L) and smooth item N(L), respectively. Formulas (9–12) indicate the means value M, variance D, terrain triangular facet F, and the range of the label L, respectively. O is the number of the images and n is terrain surface. Formulas (13)–(14) are the expressions of the data item D(L) and smooth item N(L) in the energy function, in which αd and βd are internal factors of data items; αn and βn are internal factors of smooth item; Vfili is the included angle between the terrain triangular surface fi and the responding image li; Wfili is the visual ratio, which is the ratio of visible area and real area.
D ( L ) = D P = exp [ 1 P f i ( l i ) ] 1 ,   P f i ( l i ) = ( α d ( π / 2 v f i l i ) + β d w f i l i k [ 0 , n 1 ] [ α d ( π / 2 v f k l k ) + β d w f k l k ] )
{ N ( L ) = { p , q } N N p , q δ ( l p , l q ) , δ ( l p , l q ) = { 0 l p = l q 1 l p l q N p , q = exp ( min ( max ( s q r t ( α n ( m p m q ) 2 2 σ 1 2 + β n ( d p d q ) 2 2 σ 2 2 ) ) , n _ t ) ) 1
The energy function constructed in the paper satisfies not only the basic energy construction principle but also the normal distribution. Part of the parameter values in the formula above are as follows: α = 0.8, β = 0.2; αd = 0.8, βd = 0.2; αn = βn = 0.5. The parameters control the mean value and variance of the image directly. The n_t is the threshold value of the smooth item and here n_t = 1.

5.3. Energy Minimization Solution

In the paper, α-β swap optimization algorithm is utilized to solve energy function, which is an efficient Graph-Cuts based dichotomy optimization. Not only can the initial data set be optimized and segmented, it can also transform the multidimensional digraph to 2D simple digraph which can avoid bounded uncertainty [21,22,23]. The basic idea of the algorithm supposes that there is a known label set L and segmented set P. If α, βL, exchange the label set α and β to form a new label set Lnew. To make sure that the cut corresponding to the Graph-Cuts in the new label set is smaller than the original ones, the new label set is assigned to L. The procession is circulated in turn until the minimum cut of the Graph-Cuts appears. Any label set L to segmenting set P can be equal to P = {Pl|lL} with no difference. L is the definitional domain of the label set. Pl = {pP|lp = l} is the subset label of l which is a label in set P. Given a set of labels α and β, α-β exchange is applied. That is to say, exchange α and β to form a new segmented set Pnew, in which the set lα, β remains unchanged.
Figure 6 shows the construction way of α-β graph. α and β are connected through the Pα set and Pβ set in the α-β graph, where Pαβ = PαPβ to define the weighted value of the corresponding t-link and n-link. Np is the neighborhood of node p, and the dissimilarity degree is labeled by Dp(·), it guides the energy function space of Graph-Cuts; V(·) is the diversity measure of the neighborhood nodes, it has a smoothing effect.

6. Experiments and Analysis

6.1. Experimental Data and Environment

In order to objectively evaluate and verify the validity and superiority of the proposed method in the paper, we chose the real oblique images of an experimental area to test the method. The experimental images acquisition platform is SWDC-5, which consists of an upright camera and four cameras in the four directions in a 45° angle layout. The experimental computer environment is Windows 7 operating system, with 32 G computer memory, intel core 7 (3.6 GHz, four cores and eight threads). Figure 7 shows the five-view images of the experimental area.

6.2. Pyramid Design of PagedLoad

Large-scale urban 3D texture model reconstruction requires a lot of computing resources and storage space. In order to alleviate the limitations and conflicts of computer hardware and massive data, we designed a ready-to-use PageLoad pyramid model during the execution of the algorithm. The main idea of the PageLoad pyramid model is to layer and downsample the reconstruction scene according to the different visual distances and ranges of the observer to the scene. As shown in Figure 8, the grid of the top of the pyramid is the simplest, and the lower layers of the pyramid become more and more complex.

6.3. Results Analysis

6.3.1. Comparison of Image Radiometric Correction

Figure 9 shows the defog effect using dark channels. From left to right in order are the original oblique image, Wallis processing result, dark channel processing result, and depth map estimated by dark channel. It can be easy seen from Figure 9 that the dark channel method has a better defog effect than the Wallis method, which can significantly enhance the contrast and color saturation of the image, and because the dark channel method can estimate the approximate depth range of the original image, we can achieve the desired defog effect without additional depth information.
Figure 10 shows the comparison of the authenticity of our proposed method and Hugin [18] for the camera response curve estimation. The test data set is composed of the experimental oblique images in this paper. There are many ladder-like fluctuations in the camera response curve estimated by Hugin, the main reason is because it did not constrain the local constant function, that leads to a deformed camera response curve.
Since the Wallis has a good global radiometric correction effect in the absence of fog, and the dark channel algorithm is only applied to single image defogging, and the overall image is darker, therefore, in order to avoid the processed image still displaying radiation inconsistency after dark channel algorithm correction and our camera response function. The experimental oblique images are processed with the camera response function and the dark channel algorithm, and then Wallis processing, to make up for the defects of the dark channel algorithm without image-related information.
Figure 11 shows the local effects contrast of the 3D texture reconstruction before and after the radiation pretreatment. The left side of Figure 11 shows the effect of radiation pretreatment without our method; the right side are ours, where we can see that there is no texture color jump problem. After pretreatment of our radiometric correction, mapping textures not only maintain the high contrast of the image, but also the image saturation and white balance have been significantly adjusted, and the visual effect of texture model is more natural and clear.

6.3.2. Efficiency Analysis of Occlusion Detection

Concerning occlusion detection efficiency, our use of the occlusion detection algorithm of Sparse Grid is based on Z-Buffer [24] as the reference object. The time complexity of traditional Z-Buffer is O(nm), the improved algorithm of Sparse Grid can be reduced to O(n), where m is the number of average pixels corresponding to each triangle facet, and n is the number of triangle facets. For the data set of massive triangular facets in a large-scale city scene 3D texture reconstruction, the spatial complexity of the improved algorithm mentioned above is non-uniform linear growth. Figure 12 shows the comparison result of the Sparse Grid algorithm and our occlusion detection algorithm proposed in the paper. It can be seen from Figure 12 that although the algorithm is not dominant when the amount of data is small, due to the optimization of OpenGL off-screen rendering detection mechanism and GPU acceleration, the linear steady growth of processing time can be ensured with by increasing the number of triangular facets. When the number of triangular facets is more than 300,000, the time consumption of the algorithm based on Sparse Grid shows abnormal jump growth with an increasing number of triangular facets, which demonstrates that our proposed occlusion detection algorithm has a significant advantage in dealing with large data volumes.
Figure 13 compares the effects of the occlusion detection. It can be seen from the red boxes that the occlusion phenomenon has been effectively removed in the 3D texture model on the right, guaranteeing the texture disparity coherence and consistency of the reconstructed 3D texture model.

6.3.3. Comparison of Texture Clustering and Selection

Figure 14 shows the effects of texture clustering and selection, the above is the effect of our algorithm, the below effect is not optimal. The zoom view of four corners are the enlarged pictures corresponding to red frame, where the different colors represent the triangular texture taken from the different images. The comparison of the effects in Figure 14 is obvious, after texture clustering optimization, more adjacent triangular facets in the 3D triangular mesh surface model are given the same texture source, which can reduce the problem of texture fragments and seams. Besides, the results can also be used in the texture mapping operation to improve the efficiency of texture mapping by making the same clustering texture block and to obtain the corresponding textures at one time.
Figure 15 shows the clustering statistic results of the triangular facets in the 3D triangular mesh surface model (3117 triangular facets). Figure 15a shows the texture selection distribution before clustering, Figure 15b is the result after clustering. According to the specific statistical data of clustering results, before clustering and selection optimization the isolated clustering (the number of adjacent triangular facets selecting the same texture source is 1 called isolated clustering) is about 35% of the total number of triangular facets, and the maximum clustering capacity is less than 20. After optimization, the isolated clustering is reduced to 15%, and the maximum clustering capacity is more than 40, which shows our optimization algorithm has a significant effect.
Table 1 is the efficiency comparison of energy minimization by Graph-Cuts with four different simplification 3D triangular mesh surface models. It can be seen from Table 1 that the processing time is directly proportional to the number of triangular facets, and it can converge with the global minimization quickly, the number of iterations is around 5, meanwhile, the adjacent iterative energy shows a decreasing trend. It should be noted that the local energy minimum value perhaps appears because the observed value is insufficient when the data set of the triangular facets is small. However, our constructed energy function can skip the local minimum value, and finally obtains the global minimization. This shows that our algorithm is stable and has high convergence efficiency and strong robustness.

6.3.4. Effect Analysis of 3D Texture Reconstruction

Figure 16 shows the experimental results of 3D texture reconstruction of our proposed method. The left side of Figure 16 is the global view of the 3D texture model in the experimental area, the right side of Figure 16 is the zoom view of the local 3D texture model. We can easily see the experimental results are better to maintain the consistency of color, realistic effects, and rich details. Furthermore, in order to more intuitively demonstrate the effectiveness and superiority of the proposed method, our method is compared with the famous 3D reconstruction software Photomodeler. As shown in Figure 17, there is some local amplification of the 3D texture model reconstructed by Photomodeler and our proposed method. From the areas indicated by the arrows in Figure 17, we can see that there are lots of texture inconsistencies, mapping errors, and seams in the results which are reconstructed by Photomodeler. After our texture optimization, the problems such as texture fragments and texture seams are significantly reduced, and the texture color continuity and integrity have been well maintained.

7. Conclusions

In this paper, the experimental images are obtained by oblique photography technology which is used for 3D surface model reconstruction and as a texture extraction source. The main innovations are as follows: with respect to image radiation correction, the camera response function is introduced to deal with image reprocessing; with respect to occlusion detection, an efficient OpenGL detection method and a partial quartering algorithm are proposed; with respect to texture extraction, the discrete multi-label problem is solved by Graph-Cuts under the frame of MRF in texture selection optimization. Through the objective comparison of experimental results, the novel method of 3D urban texture model reconstruction is proved effective; can improve clustering expression and radiation consistency of texture mapping; and demonstrates advantages of high efficiency, fine automation level, realistic effects, and low cost.

Acknowledgments

This study is supported by the National Key Research and Development Program of China (2016YFB0502202), the NSFC (91638203), the State Key Laboratory Research Expenses of LIESMARS.

Author Contributions

W.Z. and M.L. proposed the idea and wrote the manuscript; B.G. and D.L. conceived and designed the experiments; G.G. performed experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goesele, M.; Curless, B.; Seitz, S. Multi-view stereo revisited. In Proceedings of the Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; Volume 2, pp. 2402–2409. [Google Scholar]
  2. Kazhdan, M.; Bolitho, M.; Hoppe, H. Poisson surface reconstruction. In Proceedings of the Symposium on Geometry Processing, Cagliari, Italy, 26–28 June 2006; pp. 61–70. [Google Scholar]
  3. Brown, M.; Lowe, D. Automatic Panoramic Image Stitching using Invariant Features. Int. Comput. Vis. 2007, 74, 59–73. [Google Scholar] [CrossRef]
  4. Furukawa, Y.; Ponce, J. Accurate, dense, and robust multi-view stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 505–512. [Google Scholar] [CrossRef] [PubMed]
  5. Schoning, J.; Heidemann, G. Evaluation of multi-view 3D reconstruction software. In Proceedings of the Computer Analysis of Images and Patterns, Valletta, Malta, 2–4 September 2015; pp. 450–461. [Google Scholar]
  6. Soucy, M.; Godin, G.; Rioux, M. A texture-mapping approach for the compression of colored 3D triangulations. Vis. Comput. 1996, 12, 503–514. [Google Scholar] [CrossRef]
  7. Debevec, P.; Yu, Y.; Boshokov, G. Efficient View-Dependent Image-Based Rendering with Projective Texture-Mapping. In Proceedings of the Eurographics Symposium on Rendering Techniques, Vienna, Austria, 29 June–1 July 1998; pp. 105–116. [Google Scholar]
  8. Bernardini, F.; Martin, I.; Rushmeier, H. High-quality texture reconstruction from multiple scans. IEEE Trans. Vis. Comput. Graph. 2001, 7, 318–332. [Google Scholar] [CrossRef]
  9. Baumberg, A. Blending images for texturing 3D models. In Proceedings of the British Machine Vision Conference, Cardiff, UK, 2–5 September 2002; pp. 404–413. [Google Scholar]
  10. Kim, C.; Rhee, E. Realistic Facade Texturing of Digital Building Models. Softw. Eng. Appl. 2014, 8, 193–202. [Google Scholar] [CrossRef]
  11. Luo, Q.; Zhou, G.; Zhang, G.; Hang, J. The texture extraction and mapping of buildings with occlusion detection. In Proceedings of the International Geoscience and Remote Sensing Symposium, Milan, Italy, 26–31 July 2015; pp. 3002–3005. [Google Scholar]
  12. Frueh, C.; Jain, S.; Zakhor, A. Data Processing Algorithms for Generating Textured 3D Building Facade Meshes from Laser Scans and Camera Images. Int. J. Comput. Vis. 2005, 61, 159–184. [Google Scholar] [CrossRef]
  13. Iwaszczuk, D.; Hoegner, L.; Stilla, U. Quality-based building-texture selection from different sensors. In Proceedings of the Urban Remote Sensing Event (JURSE), Lausanne, Switzerland, 30 March–1 April 2015; pp. 1–4. [Google Scholar]
  14. Gal, R.; Wexler, Y.; Hoppe, H.; Cohenor, D. Seamless Montage for Texturing Models. Comput. Graph. Forum 2010, 29, 479–486. [Google Scholar] [CrossRef]
  15. Li, D.; Wang, M.; Pan, J. Auto-dodging processing and its application for optical RS image. Geomat. Inf. Sci. Wuhan Univ. 2006, 31, 753–756. [Google Scholar]
  16. Luo, S. Improved dodging algorithm based on Wallis principle. Geomat. Sci. Technol. 2015, 3, 51–58. [Google Scholar] [CrossRef]
  17. He, K.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Tran. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar]
  18. Dangelo, P. Radio metric alignment and vignetting calibration. In Proceedings of the International Conference on Computer Vision System, Bielefeld, Germany, 21–24 March 2007. [Google Scholar]
  19. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: London, UK, 2000. [Google Scholar]
  20. Zhang, Z.; Su, G.; Zhen, S.; Zhang, J. Relation OpenGL imaging process with exterior and interior parameters of photogrammetry. Geomat. Inf. Sci. Wuhan Univ. 2004, 29, 570–574. [Google Scholar]
  21. Boykov, Y.; Veksler, O.; Zabih, R. Fast approximate energy minimization via graph cuts. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 1222–1239. [Google Scholar] [CrossRef]
  22. Schmidt, M.; Alahari, K. Generalized Fast Approximate Energy Minimization via Graph Cuts: Alpha-Expansion Beta-Shrink Moves. In Proceedings of the Uncertainty in Artificial Intelligence, Barcelona, Spain, 14–17 July 2011. [Google Scholar]
  23. Hussmann, S.; Ringbeck, T.; Hagebeuker, B. A Performance Review of 3D TOF Vision Systems in Comparison to Stereo Vision Systems; InTech: Vienna, Austria, 2008. [Google Scholar]
  24. Zhang, C.; Zhang, W.; Guo, B.; Liu, J.; Li, M. Rapidly 3D texture reconstruction based on oblique photograph. Acta Geod. Cartogr. Sin. 2015, 44, 782–790. [Google Scholar]
Figure 1. Local 3D model of the experimental area. (a) 3D fitting model; (b) 3D mesh model.
Figure 1. Local 3D model of the experimental area. (a) 3D fitting model; (b) 3D mesh model.
Sensors 17 00911 g001
Figure 2. OpenGL transform procession.
Figure 2. OpenGL transform procession.
Sensors 17 00911 g002
Figure 3. Triangle facet subdivision: (a) original mesh; (b) global quartering; (c) our method.
Figure 3. Triangle facet subdivision: (a) original mesh; (b) global quartering; (c) our method.
Sensors 17 00911 g003
Figure 4. Occlusion conditions.
Figure 4. Occlusion conditions.
Sensors 17 00911 g004
Figure 5. Labeled graph: (a) shows the general labeled graph; (b) shows our labeled graph.
Figure 5. Labeled graph: (a) shows the general labeled graph; (b) shows our labeled graph.
Sensors 17 00911 g005
Figure 6. The α-β graph.
Figure 6. The α-β graph.
Sensors 17 00911 g006
Figure 7. Experimental oblique images.
Figure 7. Experimental oblique images.
Sensors 17 00911 g007
Figure 8. Mesh simplification. (a) Pyramid simplified model; (b) Scene simplified instance.
Figure 8. Mesh simplification. (a) Pyramid simplified model; (b) Scene simplified instance.
Sensors 17 00911 g008
Figure 9. Contrast of defogging effect. (a) Original image; (b) Wallis processing result; (c) Dark channel processing result; (d) Depth map.
Figure 9. Contrast of defogging effect. (a) Original image; (b) Wallis processing result; (c) Dark channel processing result; (d) Depth map.
Sensors 17 00911 g009
Figure 10. Contrast of camera response curves.
Figure 10. Contrast of camera response curves.
Sensors 17 00911 g010
Figure 11. Contrast of texture radiation correction.
Figure 11. Contrast of texture radiation correction.
Sensors 17 00911 g011
Figure 12. Efficiency contrast of occlusion detection.
Figure 12. Efficiency contrast of occlusion detection.
Sensors 17 00911 g012
Figure 13. Results comparison of occlusion detection.
Figure 13. Results comparison of occlusion detection.
Sensors 17 00911 g013
Figure 14. Contrast of clustering optimization.
Figure 14. Contrast of clustering optimization.
Sensors 17 00911 g014
Figure 15. Comparison of the statistical results of clustering: (a) Before clustering; (b) After clustering.
Figure 15. Comparison of the statistical results of clustering: (a) Before clustering; (b) After clustering.
Sensors 17 00911 g015
Figure 16. Texture reconstruction results.
Figure 16. Texture reconstruction results.
Sensors 17 00911 g016
Figure 17. Contrast of local zoom views of texture optimization.
Figure 17. Contrast of local zoom views of texture optimization.
Sensors 17 00911 g017
Table 1. Efficiency comparison of energy minimization by Graph-Cuts.
Table 1. Efficiency comparison of energy minimization by Graph-Cuts.
The Number of Triangular FacetsConvergent Rate of Energy FunctionNumber of IterationsEnergy Difference of Adjacent Iteration
31177.678 s558.4, 5.8, 6.7, 4.2
764312.488 s697.2, 68.4, 30.0, 3.6, 1.0
32,52235.553 s5351.1, 66.8, 58.5, 20.2
136,834162.729 s41178.2, 278.3, 36.3

Share and Cite

MDPI and ACS Style

Zhang, W.; Li, M.; Guo, B.; Li, D.; Guo, G. Rapid Texture Optimization of Three-Dimensional Urban Model Based on Oblique Images. Sensors 2017, 17, 911. https://doi.org/10.3390/s17040911

AMA Style

Zhang W, Li M, Guo B, Li D, Guo G. Rapid Texture Optimization of Three-Dimensional Urban Model Based on Oblique Images. Sensors. 2017; 17(4):911. https://doi.org/10.3390/s17040911

Chicago/Turabian Style

Zhang, Weilong, Ming Li, Bingxuan Guo, Deren Li, and Ge Guo. 2017. "Rapid Texture Optimization of Three-Dimensional Urban Model Based on Oblique Images" Sensors 17, no. 4: 911. https://doi.org/10.3390/s17040911

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop