Next Article in Journal
Remote Sensing of Near-Real-Time Heavy Precipitation Using Observations from GPM and MFG over India and Nearby Oceanic Regions
Previous Article in Journal
Performance Analysis of Detector Algorithms Using Drone-Based Radar Systems for Oil Spill Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Habitat Mapping of Ma-le’l Dunes Coupling with UAV and NAIP Imagery †

1
Department of Forestry & Wildland Resources, Humboldt State University, 1st Harpst Street, Arcata, CA 95521, USA
2
Environmental Science program, Humboldt State University, 1st Harpst Street, Arcata, CA 95521, USA
3
Environmental Science & Management program, Humboldt State University, 1st Harpst Street, Arcata, CA 95521, USA
*
Author to whom correspondence should be addressed.
Presented at the 2nd International Electronic Conference on Remote Sensing, 22 March–5 April 2018; Available online: https://sciforum.net/conference/ecrs-2.
Proceedings 2018, 2(7), 368; https://doi.org/10.3390/ecrs-2-05182
Published: 23 March 2018
(This article belongs to the Proceedings of The 2nd International Electronic Conference on Remote Sensing)

Abstract

:
The Ma-le’l Dunes are located at the upper end of the North Spit of Humboldt Bay, California and are home to a range of plant and animal species. The goal of this study was to determine which classification method was the most accurate in identifying dune features when performed on a large, diverse area. The data sources used for this study were an orthomosaic image (2017) with 14 cm spatial resolution and NAIP images (2012, 2014, and 2016) with 1 m spatial resolution. A DJI Mavic Pro Unmanned Aerial Vehicle (UAV) was used to fly a 31 acre plot of the Ma-le’l Dunes at a height of about 80 m. The images from this flight were used to create an orthomosaic image in AgisoftPhotoScan. The dune feature classes were compared with two images using supervised, unsupervised, and feature extraction classification methods and accuracy assessments were performed using 100 ground control points. The classified feature classes were beach grass, shore pine, sand, other vegetation, and water. Overall, the NAIP classified maps showed a higher accuracy for all classification methods than UAV classified maps, with 86% overall accuracy for the supervised classification. A feature extraction method showed a low accuracy for both NAIP (46%) and UAV ortho classified images (30%). Of the classified methods for the orthomosaic image, the unsupervised classification showed a high accuracy (44%). The Ma-le’l dune habitats are more heterogeneous and some classes were overlapping (i.e., beach grass and sand) due to high microtopographic variation of the dune, resulting in lower accuracy for the feature extraction method. Monitoring dune habitats and geomorphic changes over time with UAV images is important for implementing suitable management practices for species conservation and mitigating coastal vulnerabilities.

1. Introduction

Coastal ecosystems are complex and dynamic systems that are influenced by varying micro-climatic, biotic, and abiotic factors. For example, coastal habitat erosion and mobile dunes alternate coastal communities that can potentially facilitate the invasion of exotic species [1,2]. Coastal vegetation can act as barriers for storm tides and sea level rise vulnerabilities [2,3]. Mapping coastal habitats is challenging with traditional mapping and ground surveying methods because of the complexity of the landforms and the dynamic micro-topographical features of the habitat [4]. Aerial photography is a popular, cost-effective method for obtaining and analyzing remotely sensed data and also a useful tool for determining the characteristics of dune features remotely. Unmanned aerial vehicles (UAV) have become a popular and cost-effective remote sensing technology, composed of aerial platforms capable of carrying small-sized and lightweight sensors [5]. Discerning features from one another can be difficult when viewing the image with the naked eye, but when using remote sensing software, the task can be accomplished with greater accuracy.
Light-weight UAVs loaded with multispectral sensors are useful for data acquisition in diverse ecosystems and have the advantages of being low-cost, having high spatial and temporal resolution, and having minimum risk [6,7]. UAV imagery can be utilized to obtain a high degree of classification accuracy for heterogeneous landscapes, while solving classification problems for the analysis of informal settlements of topographic features such as irregular buildings and sloped terrain vegetation classes [8]. UAV-based hyperspectral images and digital surface models (DSM) have been used to derive 13 species classification in the wetland area of Hong Kong [6], estimating percent cover of emergent vegetation in a wetland [9] and discriminating invasive species within coastal dune vegetation [10]. Similar processes were integrated into this study, utilizing multispectral images taken from a UAV for classification analysis on a larger, more diverse study area.
The Ma-le’l Dunes are located at the upper end of the North Spit of Humboldt Bay, and are home to a range of both plant and animal species. A preliminary study on identifying Ma-le’l Dune features in a selected plot of the UAV orthomosaic image has given the baseline information for this study [10]. The study compared the accuracy of feature extraction and supervised/unsupervised classifications for land-use classes in the dune habitat. The accuracy assessments showed that the supervised classification had an accuracy of 50%, while feature extraction had an accuracy of 65%. However, the selected plot for the feature extraction only captured an area of the beach and foredunes, while the entire orthomosaic image shows the ocean, beach, foredunes, and middunes. Our goal for this study was to determine which analysis is most accurate in identifying dune features when performed on a larger, more diverse area. The study area was expanded to the entire orthomosaic plot, which is approximately 31 acres. We perform the same classification analyses as in the previous study on the orthomosaic image and National Agriculture Imagery Program (NAIP) images of the dunes taken in 2012, 2014, and 2016. Methodologies that were used and incorporated included remote sensing, geographic information systems, and cartography. The analysis contributes theoretically and empirically to the ongoing research in the field of geospatial science.

2. Experiments

A DJI Mavic Pro UAV was used to fly a 31 acre plot of the Ma-le’l Dunes in Humboldt County, California at a height of about 80 m. Mission planning was conducted utilizing the DJI Ground Station Pro application on a tablet. There were a total of two missions flown within the same area, one in a west and east direction with the sensor pointed directly down while the other mission was flown in a north and south direction with the sensor pointed at a 15° upward angle. The images were georeferenced into a dense point cloud by taking estimates of the camera positions and calculating elevation.
The UAV images were processed and ortho-mosaicked using Agisoftphotoscan. The Align Photos Tool was used to find matching points between overlapping images, estimate camera positions of each photo, and build a sparse point cloud model. Place markers were utilized to optimize the camera position and orientation of the data. This was done by using the Build Mesh Workflow to reconstruct the geometry. Based on the estimated camera positions, the program calculates depth information for each camera to be combined into a single dense point cloud with the Build Dense Cloud Workflow. A textured polygonal model was created from the dense point cloud to assist with precise marker placement. The Build DEM Workflow was then used to generate a digital elevation model (DEM). The coordinate system was set to WGS 84 UTM Zone 10N. The DEM was then used to build an orthomosaic with the pixel size set to 14 cm. The DEM and orthomosaic were exported as tif files.
NAIP images from 2012, 2014, and 2016 were downloaded from United States Geological Survey’s Earth Explorer and clipped to the orthomosaic extent. The four images were opened in ENVI 64 bit and used in an Example-Based Feature Extraction Workflow, Supervised Classification Workflow, and Unsupervised Classification Workflow. Four classifications were set when classifying the three NAIP images: European beachgrass, shore pines, sand, and other vegetation. A fifth classification for water was added when analyzing the orthomosaic image.
The Example-Based Feature Extraction Workflow Tool was used, and the entire image was set as the spatial subset. The object creation segment and merge was left as the default, and the segmentation and computing attributes process was run. After each NAIP image was selected in the workflow, examples of each feature were selected. These features included European beachgrass, shore pine, sand, and other vegetation. The fifth class of water had to be added to the orthographic, photographs because water was included in the orthographic images, whereas it was not in any of the three NAIP images. After each feature had examples selected, a raster was created showing the distribution of each feature class.
The Classification Workflow Tool was used to perform supervised and unsupervised classification on the images. For the supervised classification, approximately 15 training sites for each of the five classes were selected. Spectral Angle Mapper was selected as the classification method in the algorithm tab. The resulting classifications were then refined using the Cleanup Tool to smooth and aggregate the image. The final classified image and statistics were then exported as .img files. For the unsupervised classifications, the same steps were followed, except no training data was used.
The classified images were opened in ArcMap, and 100 stratified random points were created using the Spatial Analyst Tool: Create Accuracy Assessment Points. The Update Accuracy Assessment Points Tool was used to update the classification of each point to correspond with each of the classified images. The high resolution of each image and previous knowledge from site visits allowed for the true classifications of each point to be identified on each image visually. The classified data and the true class data were then used to create accuracy assessment pivot tables in excel. The pivot tables displayed how the assessment points were classified and how they compared to their true classifications, and allowed for the overall accuracy to be calculated. The final maps of the classified images were designed and exported using ArcMap and Adobe Illustrator. These maps were made in a process of laying the NAIP or orthographic image under the classification raster, and then lowering the transparency of the classification raster to 60% opacity.

3. Results

The results of the analysis and accuracy assessments showed that supervised classifications had the highest overall accuracy, followed by unsupervised classifications and then feature extraction (Figure 1, Figure 2 and Figure 3, Table 1), when classifying European beachgrass, shore pine, sand, and other vegetation. In addition, the NAIP imagery classifications yielded a higher overall accuracy than the orthomosaic image classifications across all three methods. The orthomosaic image showed a higher topographical variation than NAIP images, in which supervised and unsupervised classification resulted in overlapping classes. However, feature extraction did not show a higher variation in image classification, and the accuracy assessment showed only 30% (Figure 3, Table 1).
The overall accuracy assessment for NAIP supervised classification images is ≥80 for each time step, and therefore we utilized those images for time-series analysis to see the land-use/cover change between 2012–2014, 2014–2016, and 2012–2016 (Table 2). The coefficient of agreement (Kappa) for classified images for all three time steps was 0.60. Of the land-use/cover classes, sand, shore pine, and other vegetation showed a ≥86 producer’s accuracy, while sand and shore pine showed a ≥87 user’s accuracy between time steps (Table 2). The transition matrix between 2012 and 2014 showed that sand was mainly misclassified to other vegetation, and other vegetation to beach grass. In 2014–2016 period, misclassification was high between other vegetation and sand, and shore pine and beach grass. Finally, the entire time period (2012–2016) showed a comparatively high misclassification, sand-other vegetation, and other vegetation-beach grass (Table 2).

4. Discussion

The results of this analysis indicated that when using multispectral imagery on dune ecology, it is most accurate to use a supervised classification when putting the ecosystem features into separate classes. The orthomosaic image was at a resolution that was too high for the feature extraction to accurately assign the correct class. This can be said for the unsupervised classification as well.
Increasing the number of training sites for the supervised classification and feature extraction could aid in improving accuracy for future research. Supervised classification worked well in this study, contributing moderate agreement or accuracy of kappa analysis (60%) between classified and reference image [11]. Additionally, using assessments points with an equalized stratified random distribution, which would place an equal number of points in each class, could give more insight on the classification accuracy of each method based on each selected class. Classifications could also be improved by including topographic data with visual reflectance. Studies showed that including this could have a greater influence on classification accuracy than increasing resolution [12]. Images acquired through another source, such as kite aerial photography, could additionally be added and analyzed to observe different color composites.
Feature extraction is a valuable classification technique for meter and/or sub-meter resolution images, for which our results for NAIP imagery reported moderately accuracy (40%–64%). The UAV orthomosaic image captured micro-topographic features, which makes it difficult to collect training sites due to high spectral and spatial resolution (ca. 14 cm). For example, a single pixel of NAIP image represents 51 pixels for the orthomosaic image. Alternative methods, such as superpixel segmentation with angular difference feature classification method [13], could be useful in yielding higher accuracy for the feature extraction technique. As Cortenbach [10] reported, a subset of the study area with a small plot of 0.72 acres worked very well for a feature extraction that contributed to 65% accuracy. The preliminary mapping at Ma-le’l Dunes with an orthomosaicimage captured linear features like social trails and coastal shorelines, and developed sea level rise and digital elevation models [10].
Compared to the UAV orthomosaic image, NAIP images have four bands, including visible and NIR, which can be used to create NDVI to discriminate vegetation with other features. As a supplement to UAV, we attempted to use a kite with two cameras including RGB and NIR filters to derive NDVI [2,3]. However, due to low wind conditions we could not take the Picavet rig up to the air, which was a limitation of this study.
The dune habitat management is necessary to understand the past land-use changes, anthropogenic activities (i.e., social trails), ongoing ecological alterations like invasion of exotic species, and natural environmental issues, such as sea level rise and dune movement to make better management plans to conserve the dunes. The UAV orthoimage with high spatial and temporal resolution is useful for monitoring ongoing changes periodically to implement best management practices to mitigate future climatic and/or anthropogenic impacts.

5. Conclusions

The Ma-le’l dune habitats are dynamic and fragile; this study overviewed how land-use/cover classes changed from 2012–2017 with high resolution images using three classification techniques: supervised, unsupervised, and feature extraction. Of the classification techniques, supervised classification gave the highest overall accuracy for land-use classes (≥80), in which sand and shore pine classified well among all other land-uses. The UAV orthomosaic image captured a high topographic variability, which produced numerous fragments resulting in lower accuracy for the classifications. Therefore, future feature extractions for high-resolution orthomage may require additional steps, such as texture analysis [10] and the angular difference feature classification method [13], to obtain better results. The study can be expanded upon or furthered to support the detailed mapping of European beach grass, a predominant invasive species in the dune, using a high-resolution ortho imagery.

Author Contributions

In this article: J.L., an FAA Certified Remote Pilot, flew the UAV under FAA regulations and with permission from Friends of the Dunes; J.D. and J.C. flew a kite prior to the UAV study; S.F., M.C., P.C. and K.T. participated in data analysis and the manuscript write-up. B.M. supervised the project by teaching and conducting field labs for the intermediate remote sensing class. In addition, B.M. contributed to data analysis and manuscript write-up. All authors have read and agreed to the published version of the manuscript.

Acknowledgments

We greatly acknowledge the non-profit organization Friends of the Dunes for giving us permission to conduct this research at the Ma-le’l Dunes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lubke, R.A. Vegetation Dynamics and Succession on Sand Dunes of the Eastern Coasts of Africa. In Coastal Dunes; Martínez, M.L., Psuty, N.P., Eds.; Ecological Studies; Springer: Berlin/Heidelberg, Germany, 2008; Volume 171. [Google Scholar]
  2. Madurapperuma, B.D.; Dellysse, J.E.; Kuruppuarachchi, K.A.J.M.; Dissanayake, K. Mapping topographic and vegetation bioshield mass recovery along the shoreline using Kite Aerial Photography. In Proceedings of the 22nd International Forestry and Environment Symposium, Department of Forestry and Environmental Science, University of Sri Jayewardenepura, Tangerine Beach Hotel, Waskaduwa, Sri Lanka, 10–11 November 2017; p. 78. [Google Scholar]
  3. Madurapperuma, B.D.; Dellysse, J.E.; Zahir, I.L.M.; Aathamlebbe, I. Mapping shoreline vulnerabilities using kite aerial photographs at OluvilHarbour in Ampara. In Proceedings of the 7th International Conference of South Eastern University of Sri Lanka, South Eastern University of Sri Lanka, Oluvil, Sri Lanka, 7–8 December 2017; pp. 197–204. [Google Scholar]
  4. Shanmugam, S.; Lucas, N.; Phipps, P.; Richards, A.; Barnsley, M. Assessment of Remote Sensing Techniques for Habitat Mapping in Coastal Dune Ecosystems. J. Coast. Res. 2003, 19, 64–75. [Google Scholar]
  5. Adão, T.; Hruška, J.; Pádua, L.; Bessa, J.; Peres, E.; Morais, R.; Sousa, J. Hyperspectral Imaging: A Review on UAV-Based Sensors, Data Processing and Applications for Agriculture and Forestry. Remote Sens. 2017, 9, 1110. [Google Scholar] [CrossRef]
  6. Li, Q.S.; Wong, F.K.K.; Fung, T. Assessing the utility of UAV-borne hyperspectral image and photogrammetry derived 3D data for wetland species distribution quick mapping. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 209–215. [Google Scholar] [CrossRef]
  7. Anderson, K.; Gaston, K.J. Lightweight unmanned aerial vehicles will revolutionize spatial ecology. Front. Ecol. Environ. 2013, 11, 138–146. [Google Scholar] [CrossRef]
  8. Gevaert, C.; Persello, C.; Vosselman, G. Optimizing Multiple Kernel Learning for the Classification of UAV Data. Remote Sens. 2016, 8, 1025. [Google Scholar] [CrossRef]
  9. Marcaccio, J.V.; Markle, C.E.; Chow-Fraser, P. Unmanned aerial vehicles produce high-resolution, seasonally-relevant imagery for classifying wetland vegetation. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2015, 40, 249. [Google Scholar] [CrossRef]
  10. Cortenbach, J.W.; Arthur, T.; Johnson, G.; Trockey, E.; Borrowdale, C.; Close, P.; Fleming, S.; Collin, M.; Luckens, E.; Thuresson, K.; et al. Mapping Ma-le’l Dune Coastal Ecosystem: A multiscale approach using UAV images. In Proceedings of the INRSEP/CNRS UG Scientific Research Symposium, Humboldt State University Library, 2 October 2017; Available online: https://engineering.humboldt.edu/news/cnrsinrsep-undergraduate-scientific-research-symposium-2017-october-13-12-pm.
  11. Jensen, J.R. Introductory Digital Image Processing: A Remote Sensing Perspective, 3rd ed.; Prentice-Hall, Inc.: Upper Saddle River, NJ, USA, 2005; pp. 506–513. [Google Scholar]
  12. Sturdivant, E.; Lentz, E.; Thieler, E.R.; Farris, A.; Weber, K.; Remsen, D.; Henderson, R. UAS-SfM for Coastal Research: Geomorphic Feature Extraction and Land Cover Classification from High-Resolution Elevation and Optical Imagery. Remote Sens. 2017, 9, 1020. [Google Scholar] [CrossRef]
  13. Huang, X.; Chen, H.; Gong, J. Angular difference feature extraction for urban scene classification using ZY-3 multi-angle high-resolution satellite imagery. ISPRS J. Photogramm. Remote Sens. 2018, 135, 127–141. [Google Scholar] [CrossRef]
Figure 1. Results of the unsupervised classifications. Top left: 2012 NAIP. Top right: 2014 NAIP image. Bottom left: 2016 NAIP image. Bottom right: 2017 Orthomosaic image.
Figure 1. Results of the unsupervised classifications. Top left: 2012 NAIP. Top right: 2014 NAIP image. Bottom left: 2016 NAIP image. Bottom right: 2017 Orthomosaic image.
Proceedings 02 00368 g001
Figure 2. Results of the supervised classifications. Top left: 2012 NAIP. Top right: 2014 NAIP image. Bottom left: 2016 NAIP image. Bottom right: 2017 Orthomosaic image.
Figure 2. Results of the supervised classifications. Top left: 2012 NAIP. Top right: 2014 NAIP image. Bottom left: 2016 NAIP image. Bottom right: 2017 Orthomosaic image.
Proceedings 02 00368 g002
Figure 3. Results of the feature extraction. Top left: 2012 NAIP. Top right: 2014 NAIP image. Bottom left: 2016 NAIP image. Bottom right: 2017 Orthomosaic image.
Figure 3. Results of the feature extraction. Top left: 2012 NAIP. Top right: 2014 NAIP image. Bottom left: 2016 NAIP image. Bottom right: 2017 Orthomosaic image.
Proceedings 02 00368 g003
Table 1. Comparison of overall accuracy of three classification methods: supervised, unsupervised, and feature extraction.
Table 1. Comparison of overall accuracy of three classification methods: supervised, unsupervised, and feature extraction.
YearImageClassification Overall Accuracy (%)
UnsupervisedSupervised Feature Extraction
2012NAIP829064
2014NAIP628040
2016NAIP608646
2017Orthomoasic444030
Table 2. Land-use/cover transitional acreage from NAIP supervised classification of each sub period and the full period 2012–2016.
Table 2. Land-use/cover transitional acreage from NAIP supervised classification of each sub period and the full period 2012–2016.
CategoriesSandOther VegetationShore PineBeach GrassTotal% Accuracy
ProducersUsers
2012—2014
Sand5.65942.80760.00420.16198.63319866
Other vegetation0.08434.69850.45861.18046.42185973
Shore pine0.00350.04302.99390.23303.27348691
Beach grass0.00570.39020.01660.06650.4789414
2014—2016
Sand5.63570.0642--0.02055.72056799
Other vegetation2.62854.86450.01060.40537.90888762
Shore pine0.00420.03631.95341.47233.46619856
Beach grass0.10180.63060.02920.87721.63883254
2012—2016
Sand7.46501.0356--0.08578.58648987
Other vegetation0.88364.15330.13991.22376.40057465
Shore pine--0.03511.85251.38133.26909357
Beach grass0.02150.37160.00070.08450.4784318
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Madurapperuma, B.; Close, P.; Fleming, S.; Collin, M.; Thuresson, K.; Lamping, J.; Dellysse, J.; Cortenbach, J. Habitat Mapping of Ma-le’l Dunes Coupling with UAV and NAIP Imagery. Proceedings 2018, 2, 368. https://doi.org/10.3390/ecrs-2-05182

AMA Style

Madurapperuma B, Close P, Fleming S, Collin M, Thuresson K, Lamping J, Dellysse J, Cortenbach J. Habitat Mapping of Ma-le’l Dunes Coupling with UAV and NAIP Imagery. Proceedings. 2018; 2(7):368. https://doi.org/10.3390/ecrs-2-05182

Chicago/Turabian Style

Madurapperuma, Buddhika, Paulina Close, Sean Fleming, Melissa Collin, Kevin Thuresson, James Lamping, John Dellysse, and John Cortenbach. 2018. "Habitat Mapping of Ma-le’l Dunes Coupling with UAV and NAIP Imagery" Proceedings 2, no. 7: 368. https://doi.org/10.3390/ecrs-2-05182

APA Style

Madurapperuma, B., Close, P., Fleming, S., Collin, M., Thuresson, K., Lamping, J., Dellysse, J., & Cortenbach, J. (2018). Habitat Mapping of Ma-le’l Dunes Coupling with UAV and NAIP Imagery. Proceedings, 2(7), 368. https://doi.org/10.3390/ecrs-2-05182

Article Metrics

Back to TopTop