You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

11 May 2020

Identifying Informal Settlements Using Contourlet Assisted Deep Learning

,
and
1
Department of Environmental, Earth and Geospatial Sciences, North Carolina Central University, Durham, NC 27707, USA
2
Centre of Studies in Resources Engineering, Indian Institute of Technology Bombay, Mumbai 400076, India
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Artificial Intelligence for 3D Big Spatial Data Processing

Abstract

As the global urban population grows due to the influx of migrants from rural areas, many cities in developing countries face the emergence and proliferation of unplanned and informal settlements. However, even though the rise of unplanned development influences planning and management of residential land-use, reliable and detailed information about these areas is often scarce. While formal settlements in urban areas are easily mapped due to their distinct features, this does not hold true for informal settlements because of their microstructure, instability, and variability of shape and texture. Therefore, detecting and mapping these areas remains a challenging task. This research will contribute to the development of tools to identify such informal built-up areas by using an integrated approach of multiscale deep learning. The authors propose a composite architecture for semantic segmentation using the U-net architecture aided by information obtained from a multiscale contourlet transform. This work also analyzes the effects of wavelet and contourlet decompositions in the U-net architecture. The performance was evaluated in terms of precision, recall, F-score, mean intersection over union, and overall accuracy. It was found that the proposed method has better class-discriminating power as compared to existing methods and has an overall classification accuracy of 94.9–95.7%.

1. Introduction

Due to rapid urbanization and population migration, many cities in developing countries such as India have large areas of unplanned development interspaced with planned areas. A forecast of the United Nations (UN) estimated that the population of India will be about 1.44 billion in 2024 and will surpass 1.66 billion around 2050 [1]. This increase, coupled with migration from rural areas to urban centres, will lead to the growth of both informal and formal urban settlements that include low-, medium-, and upper-class housing and commercial developments. Urbanization is often not accompanied by adequate development of infrastructure, including housing, sanitation, and transportation corridors. This lack of planning when coupled with the large share of informal low-paid employment results in the growth of informal settlements in densely populated urban areas [2]. The identification and mapping of informal settlements plays an important role in many applications, such as urban analysis, updating geographical databases, land cover change assessment, disaster management, and extraction of thematic information. However, connecting reliable information for these areas with accurate detection and classification of informal settlements using remote sensing remains a challenge. Demarcating and differentiating urban structures is difficult due to the nature and intermingling of these classes as opposed to the cleaner separation of standard land cover classes in planned areas with clearer differentiations in classes. This requires the extraction and analysis of textural and spatial features and dissimilar vegetation classes, along with urban structures lacking unique and easily distinguishable spectral signatures [3]. In other words, different urban classes may present similar spectral values that make it more challenging to accurately classify pixels when identifying informal settlements.

3. Essential Concept—Contourlet Based MRA

The central focus of this work is the utility of directional features of contourlet-based MRA in a deep learning framework. A machine learning approach basically learns various features that manifest different objects or classes in an image. The contourlet features have been used in different applications, including textural segmentation. The authors are motivated to utilize the directional features of the contourlet transform to assist a deep learning algorithm.
The contourlet transform is implemented using a set of directional filters, which are designed using basis functions that have a choice of aspect ratios and directional orientations at multiple scales. In order to facilitate multiple scales, a Laplacian pyramid approach is combined with the directional filters [24]. The directional filter coefficients effectively capture the anisotropic relationship for curvilinear and disoriented edges.
The implementation of a contourlet transform facilitates any level of decomposition, a seamless transition from one scale to another, and faithful reconstruction. The numbers of directions and angular resolutions get doubled at every subsequent finer scale. Figure 1 shows a conceptual viewpoint of multiscale directional decomposition in terms of band-pass and low-pass filters and a down-sampler. The implementation details can be found in [24]. The low-pass filter outputs approximation level information, whereas the band-pass filter extracts the detailed information from a band. The process of decomposition can further be iterated in the low-pass filtered band to extract details of an approximation. These decomposed subbands are augmented with the layers of the U-net to provide multiscale learning, along with directional information.
Figure 1. Contourlet transform decomposition structure (adapted from [24]).

4. Proposed Method

In this paper, a modified version of the U-net [51] is used by augmenting contourlet masks of different sizes at different scales. During feature extraction, there are four steps, with the last three including several subbands. The feature maps in the same level have the same size, while the feature maps in the following level are half that of the previous level. For a 3-level decomposition, there are 16, 8, and 4 subbands respectively. The expansive part aims to extract feature maps for informal settlements using contourlet masks. The number of stages in contracting and expansive parts is the same. Having a convolutional layer followed by a max-pooling layer helps in gathering contextual information present at each level of decomposition in terms of generating activation functions. The decoder expands these activation functions with the help of the up-sampler and convolutional units to obtain the original size of a band. The central aim is to enhance the receptive field of the model using the down-sampler. The residual information in the process is fed to the up-sampler for faithful reconstruction. This is attained by the skip connections in the network, while the features learned during down-sampling are used in the up-sampling part. In turn, this mechanism provides smoother edges than other fully connected convolutional networks.
Informal settlement identification is considered as a binary classification. For training, logistic regression is used by optimizing the energy function. A gradient decent algorithm is used to minimize the error function. Both the softmax and cross-entropy functions are considered for the error function. The softmax layer outputs two lines as a probability indicator for informal settlements and rest of the classes. The last layer is a convolutional layer measuring 1 x 1, which is used to transform the features into two classes for the pixel under consideration. The concatenation in the expansive segment is able to learn the features at multiple scales. The feature learning at multiple scales enhances the ability to capture different properties of the classes and improves the classification accuracy. The proposed architecture is shown in Figure 2.
Figure 2. Proposed composite architecture (Conv: Convolution layer; ReLU: Rectified Linear Unit; MaxPooling: Maximum Pooling; Concat: Concatenation; Contourlet_x: Contourlet decomposed subband at level x)

5. Results

5.1. Dataset

The dataset comprises high-resolution Worldview-2 ( 2 m × 2 m ) images of parts of Mumbai and Pune cities from Western India. The study area is densely populated with a mixture of informal and formal built-up areas. Informal settlements in the region provide housing and livelihood for the lower economic strata population. In general, these informal settlements are very dense with clusters of row houses (called chawls in Mumbai). However, there are several discernible differences: areas that have been rehabilitated near high-rise colonies and towers; long-established localities, which have regular small-scale shops (called kirana in Mumbai) and roof structures alongside; highly congested pockets with only small lanes (parts of Govandi area in Mumbai); inner-city localities with high roof density (parts of PMG colony and Mankhurd) and small drainage lines (called nallahs in Mumbai).

5.2. Implementation and Results

The dataset comprises 1006 patches extracted from original images, including 878 patches for training, 38 patches for validation, and 90 patches for testing. These training patches were randomly sampled from the original images. For wavelet-based U-net, Daub (1, 2, 3, and 4) family and bi-orthogonal 9/7 basis functions were used to obtain the approximation and detailed subbands for three decomposition levels. The Adam optimizer with the second norm loss function was used in the training phase. The initial learning rate parameter was kept at 0.001 and decreased by a factor of one-tenth, with an epoch size of 20.
In semantic segmentation, the low-pass filtered output is important, which approximates the input band while keeping the band-pass-filtered detailed information intact for better results [51,53]. With a limited number of samples, training a deep network might be very difficult. The authors in [53] used a pre-trained network to solve this problem. A quantum-inspired differential evolution method is used to fine tune the network parameters while training and to overcome premature convergence [54]. In order to achieve a reduced classification rule set, a particle swam optimization technique based on rough set theory is used to train a back-propagation network [55].
In order to have a pre-trained network, the authors used the approximation subband of Haar wavelet instead of original images for the training phase in the first step. After training on the approximation subband, the network weights were saved and a new U-net was created and trained on the original images. This was the second phase of training, where the network was initialized using saved weights from the first phase instead of random initialization. Here, the mapping from the original images to the original ground truth was learned. The expected result was that the network would able to learn the mapping faster than it would if it were to be trained on the original images with random weight initialization. By means of this transfer learning from the approximation subband, the network was able to learn the important features for segmentation at a given resolution; some of those features would also be useful for higher resolution original images.
The most common metrics used to evaluate a two-class classification method are precision and recall. The precision is calculated as the fraction of predicted informal settlement area pixels being labeled as informal settlements (IS), and recall is computed as the fraction of all labeled IS pixels that are correctly predicted. The precision and recall are also termed as correctness and completeness, respectively. F-score, mean intersection over union (mIoU), and overall accuracy (OA) are also used for quantitative assessment. With true positives (TP), false positives (FP), and false negatives (FN), the metrics are as follows:
  • Correctness (C1) = TP/(TP + FP);
  • Completeness (C2) = TP/(TP + FN);
  • F-score = 2.C1.C2/(C1 + C2).
Figure 3 and Figure 4 show the informal settlement identification results for Pune and Mumbai city images, respectively. Figure 3a,b show the original Pune city image and reference image, respectively. Figure 3c,d show the result using the U-net and wavelet-based U-net methods, in which it is observed that informal settlement areas were not identified properly and classes were misclassified in many locations. The misclassified portions are circled in red. The wavelet was found to be unsuitable for orientation and anisotropic properties of the classes, however the contourlet efficiently captured the curved boundaries and represented the disoriented and anisotropic properties of the structures. Figure 3e and Figure 4e demonstrate the utility of the contourlet features from all three decomposition levels, which show detailed directional information in the image. It is observed that informal settlements were correctly identified with better accuracy and improved edge continuity. The class does not appear to be intermixed with partially built-up or formally built-up areas, as observed while using the plain U-net method. This is due to the fact that the contourlet features capture the directional and isotropic properties of these classes. The accuracy in the boundary shape and edge continuity is observed to be better especially in the middle (reverse L-shape) and top-right potions of the image in Figure 3. Similarly, deviation from the reference image (Figure 4b) boundaries can be observed in the top-left and bottom-right portions of Figure 4c–e. Figure 3f,g and Figure 4f,g show the identification results using contourlet- and wavelet-based texture features [20]. The work in [20] emphasizes the utility of texture-based moment and energy features computed from MRA coefficients and does not utilize deep learning approaches for classification. It is observed that contourlet-assisted U-net performs better than the texture-based classification in terms of accuracy and visual interpretation. The misclassified pixels are highlighted with red circles in Figure 3 and Figure 4.
Figure 3. Informal settlement identification for parts of Pune city: (a) original image; (b) reference; (c) using plain U-net, with misclassifications circled in red; (d) using wavelet-assisted U-net, with misclassifications circled in red; (e) using contourlet-assisted U-net; (f) using contourlet texture features [20]; (g) using wavelet texture features [20]; (h) Legend.
Figure 4. Informal settlement identification for parts of Mumbai city: (a) original image; (b) reference; (c) using plain U-net, with misclassifications circled in red; (d) using wavelet-assisted U-net, with misclassifications circled in red; (e) using contourlet-assisted U-net; (f) using contourlet texture features [20]; (g) using wavelet texture features [20]; (h) Legend.
The contourlet-assisted model performs better than the conventional U-net in terms of both pixel accuracy and mean intersection over union (mIoU). It is observed that Biorthogonal 9/7 is the best among all considered wavelet basis functions. However, the contourlet-assisted U-net outperforms all of the wavelet methods and the plain U-net method. The pixel accuracies and mIoU of the models for both the images using different methods are detailed in Table 1. As can be observed, the overall accuracy for Mumbai and Pune images improved by 3.04% and 3.76%, respectively. Additionally, precision and recall for informal settlements also improved compared to the results with plain and wavelet-based U-net methods.
Table 1. Performance comparison.

6. Discussion

As described in the methodology, a comparative analysis was carried out by considering augmentation applied to the MRA features in deep learning.
  • U-cnet_1: Original Image + contourlet subbands at first level of decomposition;
  • U-cnet_2: Original Image + contourlet subbands at first and second levels of decomposition;
  • U-cnet_3: Original Image + contourlet subbands at first, second, and third levels of decomposition.
The general trend was that the accuracies and mIoU for each image increased as the order of the decomposition increased. As observed in Figure 5 and Figure 6, U-cnet_3 (U-cnet of Figure 3 and Figure 4) outperforms other MRA methods, as it utilizes all 16 subbands containing directional details, which successfully capture the disoriented anisotropic features and irregular layouts of informal settlements. As observed from Figure 4, formal and informal settlements are mixed together in many places using U-net and wavelet-assisted U-net (U-wnet) models. This erroneous segmentation is due to the lack of finer details in plain U-net and wavelet-assisted U-net. This misclassification is also observed in U-cnet_1, as it does not incorporate all the subbands of the contourlet transform.
Figure 5. Effects of contourlet subbands for Pune image: (a) reference; (b) U-cnet_1; (c) U-cnet_2; (d) U-cnet_3.
Figure 6. Effects of contourlet subbands with U-net for Mumbai image: (a) reference; (b) U-cnet_1; (c) U-cnet_2; (d) U-cnet_3.
Figure 5 and Figure 6 compare the band-wise results for the contourlet-assisted U-net. As the level of aggregation of subbands increases, the ability to capture intrinsic geometrical details and directional selectivity also increases, which in turn improves the identification accuracy of informal settlements. The U-cnet_3 model can analyze and recognise very small features of different classes containing rich and dense detailed information. This is also demonstrated using band-wise overall accuracy in Figure 7. Figure 8 presents mIoU with different levels of wavelet- and contourlet-decomposed subbands. For both the images, overall accuracy and mIoU using contourlet subbands is higher than that of wavelet subbands. Even the first level of the contourlet subbands (U-cnet_1) shows better performance than those of third level of wavelet subbands (U-wnet_3) for both of the images. The trend of overall accuracy is in agreement with the mIoU for the number of MRA subbands utilized in deep learning.
Figure 7. Band-wise overall accuracy.
Figure 8. Band-wise mean intersection over union (mIoU).
The results of a comparison with other methods using the same training and validation datasets are reported in Table 1. The proposed contourlet-assisted U-net method shows improvements in identifying informal settlements in both the datasets. The network extracts contourlet features at multiple scales, which works well in contraction, and it facilitates clear separation of disoriented boundaries.

7. Conclusions

In this paper, a deep learning method utilizing contourlet MRA features for identification of informal settlements using remote sensing data was proposed and analyzed. The major change proposed is augmentation of the directional subbands of the contourlet transform with a deep learning model. The method progressively combines subbands at various scales to extract different disoriented details, which are key manifestations of informal subregions in remotely sensed images. The proposed algorithm is tested on Worldview-2 images of Mumbai and Pune (India) covering different regions of formal and informal urban settlements. The results were compared with plain U-net and wavelet-assisted U-net models. The performance was evaluated based on the visual interpretation, precision, recall, F-score, mIoU, and overall accuracy of these methods.
The results showed that multiscale contourlet subbands in the proposed U-net yielded better class discrimination for both the datasets. The improved performance was because of the ability of the contourlet transform to capture directional features of linear and nonlinear discontinuities when compared with plain U-net and wavelet-assisted U-net models. The roof tops, boundaries of small lanes (chawls), irregular areas, and structures manifest as detailed edges in the image at different scales. These details are efficiently captured by the subbands of the contourlet transform, which show directional sensitivity and anisotropy. The contourlet subbands are capable of identifying the essence of informal areas—regions with particularly dense and jammed houses-which were not identified efficiently by the plain U-net model. The results for the contourlet-assisted model showed robust performance in terms of both visual interpretation and class identification, and are sufficiently robust against random pixels while preserving spatial regularity. An overall classification accuracy of 94.9–95.7% was attained with proper boundary shapes and edge continuity. The proposed model would provide local bodies a better mechanism to identify informal settlements in order to carry out advanced analysis in urban planning. The proposed method provides an option to local municipal corporations to enhance the efficiency of their often limited resources, especially due to prohibitive software licensing cost in developing countries, and to target support and improvement measures. Other non-wavelet-based MRA, such as curvelet and shearlet transforms, will be explored in a deep learning framework in future studies.

Author Contributions

Conceptualization, R.A.A., R.M., and K.M.B.; methodology, R.A.A.; software, R.A.A.; formal analysis, R.A.A., R.M., and K.M.B.; writing—original draft preparation, R.A.A.; writing—review and editing, R.M. and K.M.B.; supervision, R.M. and K.M.B.; funding, R.M. and R.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors would like to thank all the anonymous reviewers and the editor for their valuable input and comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. DESA, UN. United Nations Department of Economic and Social Affairs/Population Division (2009b): World Population Prospects: The 2008 Revision. 2010. Available online: http://esa.un.org/unpp (accessed on 10 December 2019).
  2. UN. World Urbanization Prospects. The 2009 Revision; Population Division, Department of Economic and Social Affairs, United Nations Secretariat: New York, NY, USA, 2009.
  3. Mason, S.O.; Baltsavias, E.P.; Bishop, I. Spatial decision support systems for the management of informal settlements. Comput. Environ. Urban Syst. 1997, 21, 189–208. [Google Scholar] [CrossRef]
  4. Taubenböck, H.; Kraff, N.J. The physical face of slums: A structural comparison of slums in Mumbai, India, based on remotely sensed data. J. Hous. Built Environ. 2014, 29, 15–38. [Google Scholar] [CrossRef]
  5. Kuffer, M.; Barros, J.; Sliuzas, R.V. The development of a morphological unplanned settlement index using very-high-resolution (VHR) imagery. Comput. Environ. Urban Syst. 2014, 48, 138–152. [Google Scholar] [CrossRef]
  6. Owen, K.K.; Wong, D.W. An approach to differentiate informal settlements using spectral, texture, geomorphology and road accessibility metrics. Appl. Geogr. 2013, 38, 107–118. [Google Scholar] [CrossRef]
  7. Hofmann, P.; Strobl, J.; Blaschke, T.; Kux, H. Detecting informal settlements from QuickBird data in Rio de Janeiro using an object based approach. In Object-Based Image Analysis; Springer: Berlin/Heidelberg, Germany, 2008; pp. 531–553. [Google Scholar]
  8. Kohli, D.; Warwadekar, P.; Kerle, N.; Sliuzas, R.; Stein, A. Transferability of object-oriented image analysis methods for slum identification. Remote Sens. 2013, 5, 4209–4228. [Google Scholar] [CrossRef]
  9. Jain, S. Use of IKONOS satellite data to identify informal settlements in Dehradun, India. Int. J. Remote Sens. 2007, 28, 3227–3233. [Google Scholar] [CrossRef]
  10. Kit, O.; Lüdeke, M.; Reckien, D. Texture-based identification of urban slums in Hyderabad, India using remote sensing data. Appl. Geogr. 2012, 32, 660–667. [Google Scholar] [CrossRef]
  11. Konstantinidis, D.; Stathaki, T.; Argyriou, V.; Grammalidis, N. Building Detection Using Enhanced HOG–LBP Features and Region Refinement Processes. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 888–905. [Google Scholar] [CrossRef]
  12. Liu, F.; Jiao, L.; Hou, B.; Yang, S. POL-SAR Image Classification Based on Wishart DBN and Local Spatial Information. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3292–3308. [Google Scholar] [CrossRef]
  13. Munyati, C.; Motholo, G.L. Inferring urban household socioeconomic conditions in Mafikeng, South Africa, using high spatial resolution satellite imagery. Urban Plan. Transp. Res. 2014, 2, 57–71. [Google Scholar] [CrossRef]
  14. Ansari, R.A.; Buddhiraju, K.M.; Bhattacharya, A. Textural classification of remotely sensed images using multiresolution techniques. Geocarto Int. 2019, 1–23. [Google Scholar] [CrossRef]
  15. Vatsavai, R.R.; Bhaduri, B.; Graesser, J. Complex settlement pattern extraction with multi-instance learning. In Proceedings of the Joint Urban Remote Sensing Event 2013, Sao Paulo, Brazil, 21–23 April 2013; pp. 246–249. [Google Scholar]
  16. Engstrom, R.; Sandborn, A.; Yu, Q.; Burgdorfer, J.; Stow, D.; Weeks, J.; Graesser, J. Mapping slums using spatial features in Accra, Ghana. In Proceedings of the 2015 Joint Urban Remote Sensing Event (JURSE), Lausanne, Switzerland, 30 March–1 April 2015; pp. 1–4. [Google Scholar]
  17. Huang, X.; Liu, H.; Zhang, L. Spatiotemporal detection and analysis of urban villages in mega city regions of China using high-resolution remotely sensed imagery. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3639–3657. [Google Scholar] [CrossRef]
  18. Regniers, O.; Bombrun, L.; Lafon, V.; Germain, C. Supervised classification of very high resolution optical images using wavelet-based textural features. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3722–3735. [Google Scholar] [CrossRef]
  19. Huang, Y.; De Bortoli, V.; Zhou, F.; Gilles, J. Review of wavelet-based unsupervised texture segmentation, advantage of adaptive wavelets. IET Image Process. 2018, 12, 1626–1638. [Google Scholar] [CrossRef]
  20. Ansari, R.A.; Buddhiraju, K.M. Textural segmentation of remotely sensed images using multiresolution analysis for slum area identification. Eur. J. Remote Sens. 2019, 52 (Suppl. 2), 74–88. [Google Scholar] [CrossRef]
  21. Beyond Wavelets; Welland, G., Ed.; Academic Press: New York, NY, USA, 2003; Volume 10. [Google Scholar]
  22. Ansari, R.A.; Buddhiraju, K.M. Noise Filtering in High-Resolution Satellite Images Using Composite Multiresolution Transforms. PFG—J. Photogramm. Remote Sens. Geoinf. Sci. 2018, 86, 249–261. [Google Scholar] [CrossRef]
  23. Arivazhagan, S.; Ganesan, L. Texture segmentation using wavelet transform. Pattern Recognit. Lett. 2003, 24, 3197–3203. [Google Scholar] [CrossRef]
  24. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef]
  25. Zhang, Q.; Wang, Y.; Liu, Q.; Liu, X.; Wang, W. CNN based suburban building detection using monocular high resolution Google Earth images. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 661–664. [Google Scholar]
  26. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  27. Zhou, B.; Lapedriza, A.; Xiao, J.; Torralba, A.; Oliva, A. Learning deep features for scene recognition using places database. In Proceedings of the 2014 Neural Information Processing Systems Conference (NIPS), Montreal, QC, Canada, 8–11 December 2014; pp. 487–495. [Google Scholar]
  28. Wilkinson, G.G. Results and implications of a study of fifteen years of satellite image classification experiments. IEEE Trans. Geosci. Remote Sens. 2005, 43, 433–440. [Google Scholar] [CrossRef]
  29. Ren, Y.; Zhu, C.; Xiao, S. Deformable faster r-cnn with aggregating multi-layer features for partially occluded object detection in optical remote sensing images. Remote Sens. 2018, 10, 1470. [Google Scholar] [CrossRef]
  30. Zhang, W.; Wang, S.; Thachan, S.; Chen, J.; Qian, Y. Deconv R-CNN for small object detection on remote sensing images. In Proceedings of the IGARSS 2018-IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 2483–2486. [Google Scholar]
  31. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  32. Zhong, Y.; Fei, F.; Liu, Y.; Zhao, B.; Jiao, H.; Zhang, L. SatCNN: Satellite image dataset classification using agile convolutional neural networks. Remote Sens. Lett. 2017, 8, 136–145. [Google Scholar] [CrossRef]
  33. Liu, Q.; Hang, R.; Song, H.; Li, Z. Learning multiscale deep features for high-resolution satellite image scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 56, 117–126. [Google Scholar] [CrossRef]
  34. Pelletier, C.; Webb, G.I.; Petitjean, F. Temporal convolutional neural network for the classification of satellite image time series. Remote Sens. 2019, 11, 523. [Google Scholar] [CrossRef]
  35. Helber, P.; Bischke, B.; Dengel, A.; Borth, D. Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2019, 12, 2217–2226. [Google Scholar] [CrossRef]
  36. Vakalopoulou, M.; Karantzalos, K.; Komodakis, N.; Paragios, N. Building detection in very high resolution multispectral data with deep learning features. In Proceedings of the Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 1873–1876. [Google Scholar]
  37. Audebert, N.; Boulch, A.; Randrianarivo, H.; Le Saux, B.; Ferecatu, M.; Lefèvre, S.; Marlet, R. Deep learning for urban remote sensing. In Proceedings of the Urban Remote Sensing Event (JURSE), Dubai, UAE, 6–8 March 2017; pp. 1–4. [Google Scholar]
  38. Huang, H.; He, R.; Sun, Z.; Tan, T. Wavelet-srnet: A wavelet-based cnn for multi-scale face super resolution. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1689–1697. [Google Scholar]
  39. Liu, P.; Zhang, H.; Zhang, K.; Lin, L.; Zuo, W. Multi-level wavelet-CNN for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 19–21 June 2018; pp. 773–782. [Google Scholar]
  40. De Silva, D.D.N.; Fernando, S.; Piyatilake, I.T.S.; Karunarathne, A.V.S. Wavelet based edge feature enhancement for convolutional neural networks. In Eleventh International Conference on Machine Vision (ICMV 2018); International Society for Optics and Photonics: Munich, Germany, 2019; Volume 11041, p. 110412R. [Google Scholar]
  41. Laban, N.; Abdellatif, B.; Ebied, H.M.; Shedeed, H.A.; Tolba, M.F. Multiscale Satellite Image Classification Using Deep Learning Approach. In Machine Learning and Data Mining in Aerospace Technology; Springer: Cham, Switzerland, 2020; pp. 165–186. [Google Scholar]
  42. Farabet, C.; Couprie, C.; Najman, L.; Lecun, Y. Learning Hierarchical Features for Scene Labeling. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1915–1929. [Google Scholar] [CrossRef]
  43. Mullissa, A.G.; Persello, C.; Tolpekin, V. Fully Convolutional Networks for Multi-Temporal SAR Image Classification. In Proceedings of the IGARSS 2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 6635–6638. [Google Scholar]
  44. Maggiori, E.; Tarabalka, Y.; Charpiat, G.; Alliez, P. Convolutional Neural Networks for Large-Scale Remote-Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 645–657. [Google Scholar] [CrossRef]
  45. Audebert, N.; Saux, B.L.; Lefèvre, S. Semantic Segmentation of Earth Observation Data Using Multimodal and Multi-scale Deep Networks. In Proceedings of the Computer Vision—ACCV 2016, Taipei, Taiwan, 20–24 November 2016; pp. 180–196. [Google Scholar]
  46. Marmanis, D.; Schindler, K.; Wegner, J.D.; Galliani, S.; Datcu, M.; Stilla, U. Classification With an Edge: Improving Semantic Image Segmentation with Boundary Detection. arXiv 2016, arXiv:1612.01337. [Google Scholar] [CrossRef]
  47. Dosovitskiy, A.; Fischer, P.; Ilg, E.; Hausser, P.; Hazirbas, C.; Golkov, V.; van der Smagt, P.; Cremers, D.; Brox, T. Flownet: Learning optical flow with convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2758–2766. [Google Scholar]
  48. Marmanis, D.; Wegner, J.D.; Galliani, S.; Schindler, K.; Datcu, M.; Stilla, U. Semantic Segmentation of Aerial Images with an Ensemble of CNSS. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 473–480. [Google Scholar] [CrossRef]
  49. Sherrah, J. Fully Convolutional Networks for Dense Semantic Labelling of High-Resolution Aerial Imagery. arXiv 2016, arXiv:1606.02585. [Google Scholar]
  50. Fu, G.; Liu, C.; Zhou, R.; Sun, T.; Zhang, Q. Classification for High Resolution Remote Sensing Imagery Using a Fully Convolutional Network. Remote Sens. 2017, 9, 498. [Google Scholar] [CrossRef]
  51. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  53. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2013; pp. 3431–3440. [Google Scholar]
  54. Deng, W.; Liu, H.; Xu, J.; Zhao, H.; Song, Y. An improved quantum-inspired differential evolution algorithm for deep belief network. IEEE Trans. Instrum. Meas. 2020. [Google Scholar] [CrossRef]
  55. Deng, W.; Li, W.; Yang, X.H. A novel hybrid optimization algorithm of computational intelligence techniques for highway passenger volume prediction. Expert Syst. Appl. 2011, 38, 4198–4205. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.