You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

30 November 2019

Deep Green Diagnostics: Urban Green Space Analysis Using Deep Learning and Drone Images

,
,
,
,
and
1
Instituto Politécnico Nacional, Centro de Investigación en Computación, Av. Juan de Dios Bátiz s/n, Ciudad de México 07738, Mexico
2
Cátedra CONACyT, Instituto Politécnico Nacional, Centro de Investigación en Computación, Av. Juan de Dios Bátiz s/n, Ciudad de México 07738, Mexico
3
Escuela Superior de Cómputo, Instituto Politécnico Nacional, Av. Juan de Dios Bátiz s/n, Col. Lindavista, Ciudad de México 07738, Mexico
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Intelligent Sensors for Smart City

Abstract

Nowadays, more than half of the world’s population lives in urban areas, and this number continues increasing. Consequently, there are more and more scientific publications that analyze health problems of people associated with living in these highly urbanized locations. In particular, some of the recent work has focused on relating people’s health to the quality and quantity of urban green areas. In this context, and considering the huge amount of land area in large cities that must be supervised, our work seeks to develop a deep learning-based solution capable of determining the level of health of the land and to assess whether it is contaminated. The main purpose is to provide health institutions with software capable of creating updated maps that indicate where these phenomena are presented, as this information could be very useful to guide public health goals in large cities. Our software is released as open source code, and the data used for the experiments presented in this paper are also freely available.

1. Introduction

Grass is perhaps the most extended plant on earth. Except for rainforests, we may find grass everywhere on the planet [1]; it grows even in saltwater and arid areas. Grass plays a fundamental role in our ecosystem and has a significant impact on agriculture and economy [2]. For instance, a very substantial percentage of meat production comes from ruminant animals that are fed with grass [1]; the grass system roots help to keep the soil stable and prevent erosion [2]. For those of us who live in urban zones, grass is fundamental to our well-being [3,4]. Some studies have shown that natural landscape has a positive influence on mood, stress levels, cognitive function, and impulse control [5]. Furthermore, having access to green areas can help to some extent to prevent serious health problems such as obesity, heart disease, and diabetes, because these environments are more attractive for people to exercise and do outdoor activities [4].
Most studies of urban green space used the Normalized Difference Vegetation Index (NDVI) derived from satellite imagery [6]. For instance, in [7] the authors found, using data from the U.S.-based Nurses’ Health Study prospective cohort and the NDVI, that exposure to greenness is associated with decreased mortality and improves health. In [8], the NDVI was used in Barcelona, Spain, to conclude that exposure to outdoor surrounding greenness was associated with a beneficial impact on cognitive development in schoolchildren in the second to fourth grades (7–10 y.o.). In [9], the authors used both a cohort study, consisting of approximately 1.3 million adults, and the Cox proportional hazards models in order to, respectively, assign estimates of exposure to greenness derived from remotely sensed NDVI (NASA’s Aqua satellite), and estimate associations between residential greenness and mortality. They concluded that increased amounts of residential greenness were associated with reduced risks of dying from several common causes of death among urban Canadians. Finally, we mention that in [10], using different tools, the authors also found evidence that supports that, among other factors like food quality and exercise, being exposed to greenness improves health— particularly, the risk of heart disease can be reduced. Although the NDVI is useful for quantifying vegetation, because of its resolution, it is not able to detect smaller details, such as the presence of cans, plastic bags, bottles, etc., so is not suitable for our purposes.
Identifying contamination is important because contaminated locations may represent places with health risks and possible proliferation of harmful fauna. Automatic detection of contaminated land in the city can be very useful information to design forecast models for solid waste generation, as in [11], where the authors use four different Artificial Intelligence techniques to model this serious problem. In [12], the pollution problem is considered part of the Sustainable Development Goals for Smart Cities (SC). These cities seek to promote economic development, improve the welfare of their citizens, and help in the ability of people to use technologies to build sustainable services.

3. Deep Green Diagnostics (DGD)

Our proposal consists of the modules shown in Figure 2, explained in detail below:
Figure 2. Deep Green Diagnostics general architecture.
  • Geotagged image dataset. Users must fly a UAV about 30 m above ground level to obtain the image of the terrain to be analyzed; these images include the georeference in their metadata and will constitute the dataset used for training the Deep Neural Network. See Section 3.1 for more details.
  • Deep Neural Network. Users enter the previously obtained dataset into the system, which in turn will process it using the architecture described in Section 3.2. As output, the system will classify each image according to the eight classes shown in Figure 3.
    Figure 3. The eight proposed classes of terrain.
  • Visualization and retraining. In this stage, the users are able to visualize the classification results and, if desired, re-train the pre-loaded DNN with a new dataset. Details are presented in Section 3.3.

3.1. Geotagged Image Dataset

The process starts by capturing a set of aerial images, recently taken from ground-level, within a distance of 20 to 30 m and with natural light, ideally on a sunny day between 7 a.m. and 5 p.m. The images must be georeferenced and saved into the RGB format. For this purpose, we use a UAV, the DJI Phantom 4 drone [23], to obtain a dataset that meets the required characteristics.
After several flights of the following landscapes, a park (Bicentenario Park), a university campus (ESCOM-IPN), suburban neighborhood (Bonito Ecatepec), and a forested area were selected for designing our solution. The obtained images (img_drone) are divided into a regular grid to obtain images of 200 × 200 pixels each (img_200 × 200), thus obtaining more than 135,000 images of this size. In order to carry out the labeling, 9901 images were randomly taken. Labeling consists of assigning a tag to indicate the state of health of the soil; if the soil is also contaminated, another tag is added. The ground health classes are: healthy vegetation (H), dry vegetation (D), unhealthy vegetation (UNH), and no vegetation (NV). The latter refers to any other type of objects, such as pavement, buildings, cars, among others. Then, contaminated images are marked, adding the letter C to the previous classes: HC, DC, UNHC, and NVC, which, along with the ground health condition, indicate the presence of garbage contamination, obtaining eight classes of terrain. Examples of these classes are shown in Figure 3 and the statistics in Figure 4. This dataset is available in [24].
Figure 4. Number of data in each of the classes for training.

3.2. Deep Neural Network

Before the software can be used, it is necessary to design a DNN capable of properly classifying an image of the terrain in one of the eight classes defined in Figure 3.

Model Training and Validation

To this end, we split our dataset into 9001 images for training and 901 for testing. Then, we used the grid search method to obtain a suitable DNN for our goal, and this architecture is shown in Figure 5. It is composed of a convolutional neural network (CNN) [25] which automatically extracts features from the land images and a Multilayer Perceptron (MLP) as a classifier.
Figure 5. Architecture of the designed Deep Neural Network (convolutional neural network [CNN]+Multilayer Perceptron [MLP]).
The final parameters [26] used to train this DNN are shown in Table 1. During the experiments, we noticed that regularization does not help in the learning process, so we decided not to use it.
Table 1. Parameters used to train the Deep Neural Network (DNN).
To evaluate the DNN’s behavior, we used the well-known and widely used metrics to evaluate the performance of this kind of networks—precision, recall, and F-score [27]—which are calculated in this context using the following equations. t p means true positives, t n true negatives, f p false positives, and f n false negatives.
A c c u r a c y = i = 1 l t p i + t n i t p i + f n i + f p i + t n i l
P r e c i s i o n = i = 1 l t p i i = 1 l t p i + f p i
R e c a l l = i = 1 l t p i i = 1 l t p i + f n i
F 1 _ s c o r e = 2 × ( P r e c i s i o n × R e c a l l ) P r e c i s i o n + R e c a l l
Using the confusion matrix values shown in Figure 6a with l = 8 , we obtained the following results: training accuracy: 100%, testing accuracy: 72%, precision (micro avg.): 72%, recall (micro avg.): 72% and F 1 score (micro avg.): 0.71%. Additionally, we show in Figure 6b the normalized confusion matrix, which allows to examine in detail the classifier’s behavior. In this figure, it can be seen that the Healthy (H) and the No-Vegetation (NV) classes obtained the scores 0.94 and 0.87, respectively.
Figure 6. Confusion matrices for the test dataset.
It is important to notice that, since we acquire images at an altitude of 30 m, resolution is 1.31 cm/px. Images taken at low altitude have the advantage of having high quality, and allowing to easily spot small details, like garbage—pet bottles, organic waste, plastic bags, etc.
From each of the four selected terrains, we report the number of labels of each class for these fields (see Table 2). In Table 3 we show the ground surface covered by the drone, measured in square meters.
Table 2. Number of labels of each class for all maps.
Table 3. Number of square meters analyzed by the drone.
Finally, our open source software is available at [28], and a tutorial video about how to use it can be found at [29].

3.3. Visualization and Retraining

The output of the system is represented as cells arranged in a matrix. Each cell is colored according to the color code in Table Figure 3, as shown in Figure 7. Cells correspond to a portion of the analyzed terrain. To be able to see in detail the results of classification, it is possible to click on any cell of interest, then the system informs about the percentages of healthy terrain, percentage of biomass, etc. See Figure 8.
Figure 7. Visualization of results as matrix cells.
Figure 8. Visualization of detail and biomass percentages.
Also, there is a button labelled as retrain. This button allows the user to load a new dataset from a different terrain, and retrain the DNN. When the training is finished, the user can load data from this new terrain and obtain new classification results.
As an example, consider the map shown in Figure A1. For this map we applied the process described in Section 3. We obtained 600 images (img_200× 200) for retraining and 120 images (img_200 × 200) for testing. We ran the experiment 10 times and an average accuracy of 68.33% was obtained on the test set. Additionally, Figure 9b shows that the Healthy (H) and No-Vegetation (NV) classes obtained scores of 0.84 and 0.73, respectively, which is consistent with the results previously shown in Figure 6b. This is an example of how our method is able to reasonably analyze new fields without the need to change its architecture or having to manually adjust further parameters. This is a convenient feature because this way it is possible for the final user to easily apply this method to obtain information from their own maps.
Figure 9. Confusion matrices for the test dataset of Figure A1.

3.4. Full Map Display

In addition, our software allows users to examine in detail each piece of classification results on a map. In order to illustrate this, we manually built a complete map of the four locations to show that is possible to precisely identify zones where the terrain is dry, unhealthy, and/or contaminated, see Figure A3, Figure A4 and Figure A5.

4. Discussion

4.1. Model’s Performance

In this study, the terrain classifier model obtains a testing accuracy of 72 % . To our knowledge, there are no other works that perform exactly the same tasks, therefore, we are unable to present a straightforward comparison; however, there are several studies that use closely related approaches. For instance, in [30] the authors use the pre-trained network Inception-v3 to extract the characteristics used to obtain land cover types, while in [31] a DCNN is used to extract from images a characteristic vector, to subsequently introduce it into an SVM classifier to generate the human perception mapping. To improve the knowledge capability of our system, we considered four terrains with very different or diverse landscapes, as shown in Figure A2, Figure A3, Figure A4 and Figure A5. From these terrains, we obtained a total of 49,800 image ( img_200 × 200) labels. The drone flights covered 341,822.18 m 2 . The obtained results support the evidence that our model can be used for knowledge transfer [32,33], i.e., it can be re-trained to add functionality into our system.
Additionally, in Table 4, we show the biomass percentage found in each terrain, and the percentage of the polluted area in all four reported maps, where we can see that the forested area has the greatest amount of biomass ( 78.69 % ), while Bonito Ecatepec is the most polluted ( 16.73 % ) and has the least amount of biomass ( 9.37 % ).
Table 4. Contamination percentages obtained by our system.
Terrain classification through modern technology has been carried out for several decades, because there are countless situations where the obtained information helps to solve or to avoid several problems. Initially, satellite images were used, whereas recently cameras mounted on unmanned aerial or ground vehicles are used as the primary source of images. Even when a full review of this topic is beyond the scope of this work, we mention in Table 5 some interesting studies related to terrain classification, with different image sources and applications, to emphasize its relevance. In [34], the authors analyze the health of tropical forests and one of their most important results is that the use of NDVI was less precise due to the saturation effect. Mainly due to this effect, we did not consider NDVI to be useful for our application.
Table 5. Related works using the phrase “terrain classification” as search criteria.
Even when satellite information has been a reliable source of vegetation quantification, it is not an affordable source for most people. Conversely, images obtained from drones are of relatively easy acquisition, and they have the requisite quality to be used in several classification processes.

4.2. Software

Our software does not require measurements or adjustments of any kind, requiring only the images obtained by the drone flying over the terrain of interest to carry out the corresponding analysis. As far as we know, there is no other software that performs the same task, the closest software being SIPAU-CDMX [15], which we introduced in Section 1, identifying its differences to our system. Another similar software is called Atfarm [43], which, using satellite images, is able to analyze terrain using the Adapt-N algorithm. This algorithm indicates which parts of the farmland need to be fertilized. This software is exclusively oriented to agriculture precision, and consequently, it is not able to find pollution in the terrain, if it exists.

5. Conclusions

In this work we have shown that, using images taken by a drone at an altitude of around 30 m, it is possible to carry out detailed health and pollution analysis of a terrain. The successful design of the deep learning architecture allowed the system to perform a proper classification of previously unseen terrain images.
As we are using a well-established architecture, our main focus is on the proposal of a practical approach to apply deep learning in order to carry out environmental impact studies with high resolution ( 1.31 cm/px). Our main contribution consists of showing how using an unmanned aerial vehicle and deep learning can aid to perform automatic terrain health and contamination level analysis. We have introduced open source software useful to researchers or government agencies that require this kind of analysis. Related to future research, new images from other terrains could be obtained, particularly for highly polluted areas. Having more samples of each of the four classes that show polluted terrain parts may help to improve the DNN’s performance presented in this work in general. In summary, our results are trustworthy, but it is still necessary to expand the research and undertake analysis with more images of contaminated terrains.

Author Contributions

Conceptualization, methodology, M.A.M.-A. and C.A.D.; investigation and resources, M.A.M.-A., C.A.D., A.P.L.-J. and I.A.V.-M.; software, visualization and data curation, A.P.L.-J. and I.A.V.-M.; validation H.C.; formal analysis, M.A.M.-A., C.A.D., H.C. and M.S.S.-C.; writing—original draft preparation, M.A.M.-A., M.S.S.-C. and H.C.; writing—review and editing, H.C. and M.A.M.-A.; supervision, project administration and funding acquisition, M.A.M.-A.

Funding

This work has been possible thank to the support of the Mexican government through the FOINS program of Consejo Nacional de Ciencia y Tecnología (CONACYT) under grants Problemas Nacionales 5241, Cátedras CONACYT 556; and the SIP-IPN research grants 20190077, 20195886, and 20196405.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Full Maps

Figure A1. Map used for retrain the DNN.
Figure A2. Classification results for Bonito Ecatepec.
Figure A3. Classification results for ESCOM-IPN.
Figure A4. Classification results for Bicentenario Park.
Figure A5. Classification results for a forested area.

References

  1. O’Mara, F.P. The role of grasslands in food security and climate change. Ann. Bot. 2012, 110, 1263–1270. [Google Scholar] [CrossRef]
  2. Jacobs, S.W.; Whalley, R.; Wheeler, D.J. Grasses of New South Wales; University of New England Botany: Armidale, Australia, 2008. [Google Scholar]
  3. Angold, P.; Sadler, J.P.; Hill, M.O.; Pullin, A.; Rushton, S.; Austin, K.; Small, E.; Wood, B.; Wadsworth, R.; Sanderson, R.; et al. Biodiversity in urban habitat patches. Sci. Total Environ. 2006, 360, 196–204. [Google Scholar] [CrossRef]
  4. Wolch, J.R.; Byrne, J.; Newell, J.P. Urban green space, public health, and environmental justice: The challenge of making cities ‘just green enough’. Landscape Urban Plann. 2014, 125, 234–244. [Google Scholar] [CrossRef]
  5. Abraham, A.; Sommerhalder, K.; Abel, T. Landscape and well-being: A scoping study on the health-promoting impact of outdoor environments. Int. J. Publ. Health 2010, 55, 59–69. [Google Scholar] [CrossRef] [PubMed]
  6. Kondo, M.; Fluehr, J.; McKeon, T.; Branas, C. Urban green space and its impact on human health. Int. J. Environ. Res. Publ. Health 2018, 15, 445. [Google Scholar] [CrossRef] [PubMed]
  7. James, P.; Hart, J.E.; Banay, R.F.; Laden, F. Exposure to greenness and mortality in a nationwide prospective cohort study of women. Environ. Health Perspect. 2016, 124, 1344–1352. [Google Scholar] [CrossRef] [PubMed]
  8. Dadvand, P.; Nieuwenhuijsen, M.J.; Esnaola, M.; Forns, J.; Basagaña, X.; Alvarez-Pedrerol, M.; Rivas, I.; López-Vicente, M.; Pascual, M.D.C.; Su, J.; et al. Green spaces and cognitive development in primary schoolchildren. Proc. Natl. Acad. Sci. USA 2015, 112, 7937–7942. [Google Scholar] [CrossRef]
  9. Crouse, D.L.; Pinault, L.; Balram, A.; Hystad, P.; Peters, P.A.; Chen, H.; van Donkelaar, A.; Martin, R.V.; Ménard, R.; Robichaud, A.; et al. Urban greenness and mortality in Canada’s largest cities: A national cohort study. Lancet Planet. Health 2017, 1, e289–e297. [Google Scholar] [CrossRef]
  10. Paquet, C.; Coffee, N.T.; Haren, M.T.; Howard, N.J.; Adams, R.J.; Taylor, A.W.; Daniel, M. Food environment, walkability, and public open spaces are associated with incident development of cardio-metabolic risk factors in a biomedical cohort. Health Place 2014, 28, 173–176. [Google Scholar] [CrossRef]
  11. Abbasi, M.; El Hanandeh, A. Forecasting municipal solid waste generation using artificial intelligence modelling approaches. Waste Manag. 2016, 56, 13–22. [Google Scholar] [CrossRef]
  12. Souza, J.T.D.; Francisco, A.C.D.; Piekarski, C.M.; Prado, G.F.D. Data mining and machine learning to promote smart cities: A systematic review from 2000 to 2018. Sustainability 2019, 11, 1077. [Google Scholar] [CrossRef]
  13. Rodriguez Lopez, J.M.; Rosso, P.; Scheffran, J.; Delgado Ramos, G.C. Remote sensing of sustainable rural-urban land use in Mexico City: A qualitative analysis for reliability and validity. Interdisciplina 2015, 3, 321–342. [Google Scholar] [CrossRef][Green Version]
  14. Trimble. eCognition Software. 2019. Available online: http://www.ecognition.com/ (accessed on 9 February 2019).
  15. De la CDMX, P.A.O.T. Sistema de Información del Patrimonio Ambiental y Urbano de la CDMX. 2014. Available online: http://200.38.34.15:8008/mapguide/sig/siginterno.php (accessed on 9 February 2019).
  16. Ševo, I.; Avramović, A. Convolutional neural network based automatic object detection on aerial images. IEEE Geosci. Remote Sens. Lett. 2016, 13, 740–744. [Google Scholar] [CrossRef]
  17. Lu, H.; Fu, X.; Liu, C.; Li, L.G.; He, Y.X.; Li, N.W. Cultivated land information extraction in UAV imagery based on deep convolutional neural network and transfer learning. J. Mt. Sci. 2017, 14, 731–741. [Google Scholar] [CrossRef]
  18. Do, D.; Pham, F.; Raheja, A.; Bhandari, S. Machine learning techniques for the assessment of citrus plant health using UAV-based digital images. In Proceedings of the Autonomous Air and Ground Sensing Systems for Agricultural Optimization and Phenotyping III, Orlando, FL, USA, 15–19 April 2018; p. 106640O. [Google Scholar]
  19. Phan, C.; Raheja, A.; Bhandari, S.; Green, R.L.; Do, D. A predictive model for turfgrass color and quality evaluation using deep learning and UAV imageries. In Proceedings of the Autonomous Air and Ground Sensing Systems for Agricultural Optimization and Phenotyping II, Anaheim, CA, USA, 9–13 April 2017; p. 102180H. [Google Scholar]
  20. Ding, K.; Raheja, A.; Bhandari, S.; Green, R.L. Application of machine learning for the evaluation of turfgrass plots using aerial images. In Proceedings of the Autonomous Air and Ground Sensing Systems for Agricultural Optimization and Phenotyping, Baltimore, MD, USA, 17–21 April 2016; p. 98660I. [Google Scholar]
  21. Feng, Q.; Liu, J.; Gong, J. UAV remote sensing for urban vegetation mapping using random forest and texture analysis. Remote Sens. 2015, 7, 1074–1094. [Google Scholar] [CrossRef]
  22. Dos Santos Ferreira, A.; Freitas, D.M.; da Silva, G.G.; Pistori, H.; Folhes, M.T. Weed detection in soybean crops using ConvNets. Comput. Electron. Agric. 2017, 143, 314–324. [Google Scholar] [CrossRef]
  23. DJI. PHANTOM 4.2019. Available online: https://www.dji.com/mx/phantom-4 (accessed on 28 August 2019).
  24. Vargas, I. Deep Green Diagnostics, Mendeley Data. 2019. Available online: http://dx.doi.org/10.17632/dn8rj26kzm.4 (accessed on 26 March 2019).
  25. LeCun, Y.; Bengio, Y. Convolutional networks for images, speech, and time series. Handb. Brain Theor. Neural Netw. 1995, 3361, 1995. [Google Scholar]
  26. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  27. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  28. Lopez, A. Deep Green Diagnostics, Github. 2019. Available online: https://github.com/AnayantzinPao/DeepGreenDiagnostics (accessed on 10 September 2019).
  29. Moreno-Armendariz, M.A. Deep Green Diagnostics Video, Youtube. 2019. Available online: https://youtu.be/OnOQ8g0cAfc (accessed on 10 September 2019).
  30. Xu, G.; Zhu, X.; Fu, D.; Dong, J.; Xiao, X. Automatic land cover classification of geo-tagged field photos by deep learning. Environ. Modell. Softw. 2017, 91, 127–134. [Google Scholar] [CrossRef]
  31. Zhang, F.; Zhou, B.; Liu, L.; Liu, Y.; Fung, H.H.; Lin, H.; Ratti, C. Measuring human perceptions of a large-scale urban region using machine learning. Landscape Urban Plann. 2018, 180, 148–160. [Google Scholar] [CrossRef]
  32. Bengio, Y. Deep learning of representations for unsupervised and transfer learning. In Proceedings of the 2011 International Conference on Unsupervised and Transfer Learning Workshop, Bellevue, WA, USA, 2 July 2011; pp. 17–36. [Google Scholar]
  33. Ren, R.; Hung, G.G.; Tan, K.C. A Generic Deep-Learning-Based Approach for Automated Surface Inspection. IEEE Trans. Cybern. 2018, 48, 929–940. [Google Scholar] [CrossRef] [PubMed]
  34. González-Jaramillo, V.; Fries, A.; Bendix, J. AGB Estimation in a Tropical Mountain Forest (TMF) by Means of RGB and Multispectral Images Using an Unmanned Aerial Vehicle (UAV). Remote Sens. 2019, 11, 1413. [Google Scholar] [CrossRef]
  35. Xu, G.; Zhu, X.; Tapper, N.; Bechtel, B. Urban climate zone classification using convolutional neural network and ground-level images. Prog. Phys. Geogr. Earth Environ. 2019. [Google Scholar] [CrossRef]
  36. Kitano, B.T.; Mendes, C.C.; Geus, A.R.; Oliveira, H.C.; Souza, J.R. Corn Plant Counting Using Deep Learning and UAV Images. IEEE Geosci. Remote Sens. Lett. 2019. [Google Scholar] [CrossRef]
  37. Kinaneva, D.; Hristov, G.; Raychev, J.; Zahariev, P. Early Forest Fire Detection Using Drones and Artificial Intelligence. In Proceedings of the 2019 42nd International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 20–24 May 2019; pp. 1060–1065. [Google Scholar]
  38. Ojeda, L.; Borenstein, J.; Witus, G.; Karlsen, R. Terrain characterization and classification with a mobile robot. J. Field Rob. 2006, 23, 103–122. [Google Scholar] [CrossRef]
  39. Pellenz, J.; Lang, D.; Neuhaus, F.; Paulus, D. Real-time 3d mapping of rough terrain: A field report from disaster city. In Proceedings of the 2010 IEEE Safety Security and Rescue Robotics, Bremen, Germany, 26–30 July 2010; pp. 1–6. [Google Scholar]
  40. Jin, Y.Q.; Wang, D. Automatic detection of terrain surface changes after Wenchuan earthquake, May 2008, from ALOS SAR images using 2EM-MRF method. IEEE Geosci. Remote Sens. Lett. 2009, 6, 344–348. [Google Scholar]
  41. Sumfleth, K.; Duttmann, R. Prediction of soil property distribution in paddy soil landscapes using terrain data and satellite information as indicators. Ecol. Indic. 2008, 8, 485–501. [Google Scholar] [CrossRef]
  42. Middel, A.; Lukasczyk, J.; Zakrzewski, S.; Arnold, M.; Maciejewski, R. Urban form and composition of street canyons: A human-centric big data and deep learning approach. Landscape Urban Plann. 2019, 183, 122–132. [Google Scholar] [CrossRef]
  43. Atfarm. Precise Fertilisation Made Simple. 2019. Available online: https://www.at.farm/ (accessed on 10 September 2019).

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.