Automatic Ship Classification from Optical Aerial Images with Convolutional Neural Networks
Abstract
:1. Introduction
- A learning-driven architecture for detecting if an aerial image contains a ship or not, trained to detect these type of objects in a generic way. In order to improve the performance, the proposed architecture uses a CNN for extracting features (neural codes) which are eventually classified using a kNN algorithm. Different fine-tuning strategies have also been evaluated for training several CNN models.
- A comparison between the proposed architecture, classical and state-of-the-art works, including some of the most common CNN topologies. The proposed method outperforms the performance of previous approaches, both using a reference database and the dataset compiled in this work.
- A dataset (MASATI) freely available for download with more than 6000 optical aerial images properly labeled.
2. Related Work
3. Deep Learning Background
4. Methodology
4.1. Network Topologies
- VGG-16 and VGG-19 [38]. VGG-16 has 13 convolutional and three fully-connected layers, whereas VGG-19 is composed of 16 convolutional and three fully-connected layers. Both topologies use Dropout and Max-pooling techniques and ReLU activation functions.
- Inception V3 [39]. This architecture has six convolutional layers followed by three Inception modules and a last fully connected layer. It has fewer parameters than other similar models thanks to the Inception modules whose design is based on two main ideas: the approximation of a sparse structure with spatially repeated dense components, and the use of dimensionality reduction to keep the computational complexity in bounds.
- REsNet [42]. The deep REsidual learning Network learns residual functions with reference to the layer inputs rather than learning unreferenced functions. This technique enables the use of a large number of layers. We have used the 50-layer version for our experimental tests.
- Xception [40]. This model has 36 convolutional layers with a redesigned version of Inception modules which enable the depthwise separable convolution operation. This architecture outperforms the Inception results using the same number of parameters.
4.2. Proposed Architecture
4.3. MASATI Dataset
5. Experiments
5.1. Dataset Configuration
5.2. Evaluation Metrics
5.3. Hyperparameters Evaluation
5.4. Results with MASATI Dataset
- “Features + NN”: An approach that is similar to the method described in [7], which is based on hand-crafted features, was evaluated. This algorithm applies an adaptive threshold and a morphological opening (with a kernel of 2 × 2) to remove noise. The candidate ships are located by means of a region-growing process. These objects are characterized by a set of features which are used to train a neural network (a fully-connected network with three hidden layers and four nodes in each layer).
- “HOG + SVM”: An approach with which to extract local features from the input images that is based on the methods proposed in [9,10,11]. This algorithm calculates the Histogram of Oriented Gradients (HOG) [56] (with a 8 × 8 cell size, a 16 × 16 block size, and nine histogram channels) and classifies them using SVM (using the C-Support Vector implementation with a penalty parameter of ). HOG is based on the counting occurrences of gradient orientation in localized portions of an input image.
- “ORB + aNN”: This method uses the Oriented FAST and rotated BRIEF (ORB) [57] (with an edge threshold of 10, a patch size of 31, a scale factor of 1.2, and eight levels in the scale pyramid) to extract local features that are paired using an approximate Nearest Neighbors (aNN) algorithm. For this step we evaluated different values of k (up to ), finally determining that obtained the best results. ORB is a fast robust local feature detector based on the FAST keypoint detector and on the BRIEF (Binary Robust Independent Elementary Features) visual descriptor, and includes some modifications to enhance the performance.
5.5. Results with MWPU VHR-10 Dataset
- “BOW-SVM” of Xu et al. [58] based on Bag-Of-Words (BOW) feature and SVM classifier.
- “SSCBOW” of Sun et al. [59] based on Spatial Sparse Coding (SSC) and BOW.
- “Exemplar-SVMs” of Malisiewicz et al. [60] based on a set of exemplar-based SVMs.
- “FDDL” of Han et al. [61] based on visual saliency modeling and Fisher Discrimination Dictionary Learning.
- “COPD” of Cheng et al. [49] based on a collection of part detectors, in which each detector is a linear SVM classifier specialized in the classification of objects or recurring spatial patterns within a certain range of orientation.
6. Conclusions and Future Work
Acknowledgments
Author Contributions
Conflicts of Interest
References
- Crisp, D. The State-of-the-Art in Ship Detection in Synthetic Aperture Radar Imagery; Australian Government, Department of Defense: Canberra, Australia, 2004; p. 115. [Google Scholar]
- Greidanus, H.; Kourti, N. Findings of the DECLIMS project—Detection and Classification of Marine Traffic from Space. In Proceedings of the SEASAR 2006: Advances in SAR Oceanography from ENVISAT and ERS Missions, Frascati, Italy, 23–26 January 2006. [Google Scholar]
- Marino, A.; Sanjuan-Ferrer, M.J.; Hajnsek, I.; Ouchi, K. Ship Detection with Spectral Analysis of Synthetic Aperture Radar: A Comparison of New and Well-Known Algorithms. Remote Sens. 2015, 7, 5416. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Lure, F.Y.M.; Rau, Y.C. Detection of Ship Tracks in AVHRR Cloud Imagery with Neural Networks. In Proceedings of the 1994 IEEE International Geoscience and Remote Sensing Symposium (IGARSS ’94), Pasadena, CA, USA, 8–12 August 1994; pp. 1401–1403. [Google Scholar] [CrossRef]
- Weiss, J.; Luo, R.; Welch, R. Automatic detection of ship tracks in satellite imagery. In Proceedings of the IEEE International Geoscience and Remote Sensing (IGARSS ’97), Remote Sensing—A Scientific Vision for Sustainable Development, Singapore, 3–8 August 1997; Volume 1, pp. 160–162. [Google Scholar] [CrossRef]
- Corbane, C.; Marre, F.; Petit, M. Using SPOT-5 HRG Data in Panchromatic Mode for Operational Detection of Small Ships in Tropical Area. Sensors 2008, 8, 2959–2973. [Google Scholar] [CrossRef] [PubMed]
- Corbane, C.; Najman, L.; Pecoul, E.; Demagistri, L.; Petit, M. A Complete Processing Chain for Ship Detection Using Optical Satellite Imagery. Int. J. Remote Sens. 2010, 31, 5837–5854. [Google Scholar] [CrossRef]
- Zhu, C.; Zhou, H.; Wang, R.; Guo, J. A Novel Hierarchical Method of Ship Detection from Spaceborne Optical Image Based on Shape and Texture Features. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3446–3456. [Google Scholar] [CrossRef]
- Bi, F.; Liu, F.; Gao, L. A hierarchical salient-region based algorithm for ship detection in remote sensing images. In Lecture Notes in Electrical Engineering; Springer: Berlin/Heidelberg, Germany, 2010; Volume 67, pp. 729–738. [Google Scholar] [CrossRef]
- Xia, Y.; Wan, S.; Jin, P.; Yue, L. A Novel Sea-Land Segmentation Algorithm Based on Local Binary Patterns for Ship Detection. Int. J. Signal Process. Image Process. Pattern Recognit. 2014, 7, 237–246. [Google Scholar] [CrossRef]
- Yang, G.; Li, B.; Ji, S.; Gao, F.; Xu, Q. Ship Detection From Optical Satellite Images Based on Sea Surface Analysis. IEEE Geosci. Remote Sens. Lett. 2014, 11, 641–645. [Google Scholar] [CrossRef]
- Marques, J.S.; Bernardino, A.; Cruz, G.; Bento, M. An algorithm for the detection of vessels in aerial images. In Proceedings of the 2014 11th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Seoul, Korea, 26–29 August 2014; pp. 295–300. [Google Scholar] [CrossRef]
- Yang, F.; Xu, Q.; Gao, F.; Hu, L. Ship detection from optical satellite images based on visual search mechanism. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 3679–3682. [Google Scholar] [CrossRef]
- Tang, J.; Deng, C.; Huang, G.B.; Zhao, B. Compressed-Domain Ship Detection on Spaceborne Optical Image Using Deep Neural Network and Extreme Learning Machine. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1174–1185. [Google Scholar] [CrossRef]
- Yao, Y.; Jiang, Z.; Zhang, H.; Zhao, D.; Cai, B. Ship detection in optical remote sensing images based on deep convolutional neural networks. J. Appl. Remote Sens. 2017, 11. [Google Scholar] [CrossRef]
- Hu, F.; Xia, G.S.; Wang, Z.; Huang, X.; Zhang, L.; Sun, H. Unsupervised Feature Learning Via Spectral Clustering of Multidimensional Patches for Remotely Sensed Scene Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2015–2030. [Google Scholar] [CrossRef]
- Zou, Z.; Shi, Z. Ship Detection in Spaceborne Optical Image With SVD Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 5832–5845. [Google Scholar] [CrossRef]
- Zhang, R.; Yao, J.; Zhang, K.; Feng, C.; Zhang, J. S-CNN Ship Detection from High-Resolution Remote Sensing Images. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 423–430. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J. A survey on object detection in optical remote sensing images. SPRS J. Photogramm. Remote Sens. 2016, 117, 11–28. [Google Scholar] [CrossRef]
- Bengio, Y.; Courville, A.; Vincent, P. Representation Learning: A Review and New Perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [PubMed]
- Lin, H.; Shi, Z.; Zou, Z. Fully Convolutional Network With Task Partitioning for Inshore Ship Detection in Optical Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1665–1669. [Google Scholar] [CrossRef]
- Yang, W.; Dai, D.; Triggs, B.; Xia, G.S. SAR-Based Terrain Classification Using Weakly Supervised Hierarchical Markov Aspect Models. IEEE Trans. Image Process. 2012, 21, 4232–4243. [Google Scholar] [CrossRef] [PubMed]
- Arel, I.; Rose, D.C.; Karnowski, T.P. Deep Machine Learning—A New Frontier in Artificial Intelligence Research. IEEE Comput. Intell. Mag. 2010, 5, 13–18. [Google Scholar] [CrossRef]
- Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed]
- Martinez, H.P.; Bengio, Y.; Yannakakis, G.N. Learning deep physiological models of affect. IEEE Comput. Intell. Mag. 2013, 8, 20–33. [Google Scholar] [CrossRef]
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
- LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2323. [Google Scholar] [CrossRef]
- LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the The Twenty-sixth Annual Conference on Neural Information Processing Systems (NIPS), Stateline, NV, USA, 3–8 December 2012. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
- Hu, F.; Xia, G.S.; Hu, J.; Zhang, L. Transferring Deep Convolutional Neural Networks for the Scene Classification of High-Resolution Remote Sensing Imagery. Remote Sens. 2015, 7, 14680–14707. [Google Scholar] [CrossRef]
- Zhang, L.; Zhang, L.; Du, B. Deep Learning for Remote Sensing Data: A Technical Tutorial on the State of the Art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
- Ball, J.E.; Anderson, D.T.; Chan, C.S. Comprehensive survey of deep learning in remote sensing: Theories, tools, and challenges for the community. J. Appl. Remote Sens. 2017, 11, 042609. [Google Scholar] [CrossRef]
- Pascanu, R.; Montufar, G.; Bengio, Y. On the number of inference regions of deep feed forward networks with piece-wise linear activations. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. arXiv, 2015; arXiv:abs/1512.00567. [Google Scholar]
- Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. arXiv, 2016; arXiv:abs/1610.02357. [Google Scholar]
- Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the 14th International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 11–13 April 2011; Volume 15, pp. 315–323. [Google Scholar]
- Azizpour, H.; Razavian, A.S.; Sullivan, J. Factors of Transferability for a Generic ConvNet Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38. [Google Scholar] [CrossRef] [PubMed]
- Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Advances in Neural Information Processing Systems (NIPS); Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger, K.Q., Eds.; MIT Press Ltd.: Cambridge, MA, USA, 2014; pp. 3320–3328. [Google Scholar]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
- Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the 19th International Conference on Computational Statistics (COMPSTAT 2010), Paris, France, 22–27 August 2010; Springer: Berlin, Germany, 2010; pp. 177–186. [Google Scholar]
- Zeiler, M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv, 2012; arXiv:abs/1604.00133. [Google Scholar]
- Sharif Razavian, A.; Azizpour, H.; Sullivan, J.; Carlsson, S. CNN Features Off-the-Shelf: An Astounding Baseline for Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
- Zheng, L.; Zhao, Y.; Wang, S.; Wang, J.; Tian, Q. Good Practice in CNN Feature Transfer. arXiv, 2016; arXiv:abs/1604.00133. [Google Scholar]
- Cheng, G.; Han, J.; Zhou, P.; Guo, L. Multi-class geospatial object detection and geographic image classification based on collection of part detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
- Kohavi, R. A Study of Cross-validation and Bootstrap for Accuracy Estimation and Model Selection. In Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI ’95), Montreal, QC, Canada, 20–25 August 1995; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1995; Volume 2, pp. 1137–1143. [Google Scholar]
- Shalabi, L.A.; Shaaban, Z.; Kasasbeh, B. Data Mining: A Preprocessing Engine. J. Comput. Sci. 2006, 2, 735–739. [Google Scholar] [CrossRef]
- Chatfield, K.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Return of the Devil in the Details: Delving Deep into Convolutional Nets. In Proceedings of the British Machine Vision Conference, Nottingham, UK, 1–5 September 2014; pp. 1–11. [Google Scholar] [CrossRef]
- Kang, L.; Ye, P.; Li, Y.; Doermann, D. Convolutional Neural Networks for No-Reference Image Quality Assessment. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1733–1740. [Google Scholar] [CrossRef]
- Dodge, S.; Karam, L. Understanding how image quality affects deep neural networks. In Proceedings of the 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal, 6–8 June 2016; pp. 1–6. [Google Scholar] [CrossRef]
- Demsar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR ’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar] [CrossRef]
- Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An Efficient Alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision (ICCV ’11), Barcelona, Spain, 6–13 November 2011; IEEE Computer Society: Washington, DC, USA, 2011; pp. 2564–2571. [Google Scholar] [CrossRef]
- Xu, S.; Fang, T.; Li, D.; Wang, S. Object Classification of Aerial Images With Bag-of-Visual Words. IEEE Geosci. Remote Sens. Lett. 2010, 7, 366–370. [Google Scholar] [CrossRef]
- Sun, H.; Sun, X.; Wang, H.; Li, Y.; Li, X. Automatic Target Detection in High-Resolution Remote Sensing Images Using Spatial Sparse Coding Bag-of-Words Model. IEEE Geosci. Remote Sens. Lett. 2012, 9, 109–113. [Google Scholar] [CrossRef]
- Malisiewicz, T.; Gupta, A.; Efros, A.A. Ensemble of exemplar-SVMs for Object Detection and Beyond. In Proceedings of the 2011 International Conference on Computer Vision (ICCV ’11), 6–13 November 2011; IEEE Computer Society: Washington, DC, USA, 2011; pp. 89–96. [Google Scholar] [CrossRef]
- Han, J.; Zhou, P.; Zhang, D.; Cheng, G.; Guo, L.; Liu, Z.; Bu, S.; Wu, J. Efficient, simultaneous detection of multi-class geospatial targets based on visual saliency modeling and discriminative learning of sparse coding. ISPRS J. Photogramm. Remote Sens. 2014, 89, 37–48. [Google Scholar] [CrossRef]
- Yang, W.; Yin, X.; Xia, G.S. Learning High-level Features for Satellite Image Classification With Limited Labeled Samples. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4472–4482. [Google Scholar] [CrossRef]
# | Layer | F | K | Output Size | # Parameters |
---|---|---|---|---|---|
1 | Convolution | 64 | 5 | 64 × 60 × 60 | 1664 |
Activation ReLU | |||||
Max-Pooling | 2 | 64 × 30 × 30 | |||
Dropout 0.2 | |||||
2 | Convolution | 128 | 5 | 128 × 26 × 26 | 204,928 |
Activation ReLU | |||||
Max-Pooling | 2 | 128 × 13 × 13 | |||
Dropout 0.2 | |||||
3 | Fully-connected | 256 | 1 × 256 | 5,538,048 | |
Activation ReLU | |||||
4 | Fully-connected | 2 | 1 × 2 | 514 | |
SoftMax |
# | Layer | Output Size |
---|---|---|
1 | Global Average Pooling | 1 × 2048 |
Fully-connected | ||
Activation ReLU | ||
Dropout 0.2 | ||
2 | Fully-connected | 1 × 1256 |
Activation ReLU | ||
Dropout 0.2 | ||
3 | Fully-connected | 1 × (#classes) |
SoftMax |
Main Class | Sub-Class | #Samples | Description |
---|---|---|---|
Ship | Ship | 1015 | Sea with a ship (no coast) |
Detail | 1789 | Ship details | |
Multi | 188 | Multiple ships | |
Coast & ship | 121 | Coast with ships | |
Non-ship | Sea | 1010 | Sea (no ships) |
Coast | 1054 | Coast (no ships) | |
Land | 1035 | Land (no sea) |
Model | Set | NC + Softmax | NC + + kNN | ||||||
---|---|---|---|---|---|---|---|---|---|
Final | Mid | Full | Final | Mid | Full | ||||
Baseline | 1 | – | – | 64.59 | – | – | 65.12 | ||
2 | – | – | 61.76 | – | – | 62.28 | |||
3 | – | – | 84.76 | – | – | 84.81 | |||
VGG-16 | 1 | 85.66 | 73.39 | 73.28 | 87.40 | 73.48 | 74.02 | ||
2 | 85.25 | 79.18 | 80.01 | 85.40 | 79.95 | 80.24 | |||
3 | 92.53 | 98.01 | 98.09 | 93.52 | 98.12 | 98.42 | |||
VGG-19 | 1 | 94.57 | 74.91 | 97.53 | 94.62 | 75.09 | 98.02 | ||
2 | 91.64 | 94.33 | 97.18 | 91.64 | 94.50 | 97.20 | |||
3 | 95.27 | 97.67 | 97.01 | 95.51 | 97.92 | 97.02 | |||
Inception | 1 | 79.26 | 96.29 | 98.02 | 79.97 | 96.54 | 98.02 | ||
2 | 79.46 | 94.32 | 95.28 | 80.22 | 94.49 | 96.26 | |||
3 | 90.21 | 98.09 | 98.50 | 90.95 | 98.17 | 98.67 | |||
REsNet | 1 | 85.81 | 92.09 | 96.79 | 86.35 | 93.33 | 96.80 | ||
2 | 77.90 | 92.58 | 92.78 | 84.25 | 93.00 | 93.08 | |||
3 | 92.91 | 97.76 | 97.76 | 93.92 | 98.17 | 98.32 | |||
Xception | 1 | 60.14 | 95.55 | 98.27 | 62.40 | 95.56 | 98.32 | ||
2 | 55.33 | 93.74 | 96.75 | 61.38 | 93.85 | 96.92 | |||
3 | 65.33 | 96.92 | 98.92 | 65.82 | 97.09 | 99.05 | |||
Average | 1 | 81.09 | 86.45 | 92.78 | 82.15 | 86.80 | 93.04 | ||
2 | 77.92 | 90.83 | 92.40 | 80.58 | 91.16 | 92.74 | |||
3 | 87.25 | 97.69 | 98.06 | 87.94 | 97.89 | 98.30 | |||
All | 82.08 | 91.66 | 94.41 | 83.56 | 91.95 | 94.69 |
Model | Set | NC + Softmax | NC + + kNN | ||||||
---|---|---|---|---|---|---|---|---|---|
Final | Mid | Full | Final | Mid | Full | ||||
Baseline | 2 | – | – | 56.84 | – | – | 58.92 | ||
3 | – | – | 65.81 | – | – | 68.31 | |||
VGG-16 | 2 | 94.51 | 97.31 | 98.47 | 94.91 | 97.86 | 98.59 | ||
3 | 97.46 | 97.15 | 99.67 | 97.74 | 98.60 | 99.75 | |||
VGG-19 | 2 | 77.50 | 93.50 | 92.13 | 77.72 | 93.50 | 92.13 | ||
3 | 83.94 | 93.32 | 93.71 | 85.28 | 94.28 | 94.15 | |||
Inception | 2 | 71.69 | 91.40 | 92.75 | 71.77 | 91.40 | 93.26 | ||
3 | 75.62 | 93.80 | 94.38 | 75.62 | 94.04 | 94.39 | |||
REsNet | 2 | 75.78 | 94.51 | 95.01 | 78.85 | 95.20 | 95.82 | ||
3 | 82.74 | 95.27 | 94.85 | 84.66 | 96.13 | 94.88 | |||
Xception | 2 | 40.85 | 74.80 | 99.06 | 40.88 | 75.03 | 99.27 | ||
3 | 34.35 | 95.32 | 99.75 | 35.82 | 95.58 | 99.76 | |||
Average | 2 | 72.07 | 90.30 | 95.48 | 72.83 | 90.60 | 95.81 | ||
3 | 74.82 | 94.97 | 96.47 | 75.82 | 95.73 | 96.59 |
Method | Sets | ||
---|---|---|---|
Set 1 | Set 2 | Set 3 | |
Features + NN | 57.89 | 53.21 | 48.67 |
HOG + SVM | 63.00 | 60.56 | 79.27 |
ORB + aNN | 35.13 | 43.50 | 39.14 |
Our approach | 98.32 | 96.92 | 99.05 |
© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gallego, A.-J.; Pertusa, A.; Gil, P. Automatic Ship Classification from Optical Aerial Images with Convolutional Neural Networks. Remote Sens. 2018, 10, 511. https://doi.org/10.3390/rs10040511
Gallego A-J, Pertusa A, Gil P. Automatic Ship Classification from Optical Aerial Images with Convolutional Neural Networks. Remote Sensing. 2018; 10(4):511. https://doi.org/10.3390/rs10040511
Chicago/Turabian StyleGallego, Antonio-Javier, Antonio Pertusa, and Pablo Gil. 2018. "Automatic Ship Classification from Optical Aerial Images with Convolutional Neural Networks" Remote Sensing 10, no. 4: 511. https://doi.org/10.3390/rs10040511
APA StyleGallego, A. -J., Pertusa, A., & Gil, P. (2018). Automatic Ship Classification from Optical Aerial Images with Convolutional Neural Networks. Remote Sensing, 10(4), 511. https://doi.org/10.3390/rs10040511