Enhanced Ship/Iceberg Classification in SAR Images Using Feature Extraction and the Fusion of Machine Learning Algorithms
Abstract
:1. Introduction
2. Materials and Methods
2.1. Datasets
- The main dataset for this project was sourced from a Kaggle competition [36] and comprised 1604 images, depicting 851 ships and 753 icebergs. These images were SAR images captured by the Sentinel-1 satellite, with a 10-m pixel spacing. This enhanced resolution was crucial for detecting smaller ships accurately. The Sentinel-1 SAR operates with two dual polarizations, featuring HH (horizontal transmit, horizontal receive) and HV (horizontal transmit, vertical receive) bands within each scene. Human experts and geographical knowledge contributed to the labeling of the dataset. The image size for each band was 75 × 75 pixels. Notably, the intensity values for both HH and HV bands were represented in decibels (dB) and were flattened into one-dimensional vectors. Figure 1 depicts four sample images of the dataset.
- The SAR Ship Detection Dataset (SSDD) stands as the pioneering, openly available dataset extensively employed for exploring state-of-the-art ship detection technology using DL from the SAR images [37]. The dataset had 39,729 images, including 17,648 HH and 14,713 HV images, where 4911 of them had both HH and HV images that were used in our research, which required both HH and HV bands. Since this dataset only provided ship images, in order to keep our data relatively balanced, we selected and combined 1000 of these pairs of HH and HV images with the SAR data. Figure 2 illustrates two sample images from this dataset. A summary of both datasets are presented in Table 1.
2.2. Models
- Visual Geometry Group 16 (VGG16) [38] is a CNN architecture that gained significant attention and recognition for its role in image classification and deep learning research. VGG16 is characterized by its deep architecture, consisting of 16 layers, including 13 convolutional layers and 3 fully connected layers utilizing small 3 × 3 convolutional filters throughout the network, which contributes to its depth and effectiveness in capturing hierarchical features from images. VGG16 has been known to have the best classification accuracy, but since its architecture consists of approximately 138 million weights, it might not be the most computationally efficient model. In this project, the features were extracted from the output of the second Fully Connected (FC) layer, which was the last layer before the classification layer and included 4096 nodes.
- Residual Network 50 (ResNet-50) [18] is a sophisticated CNN structure that utilizes residual connections to address the vanishing gradient problem. With a total of 50 layers, it includes convolutional layers, batch normalization layers, ReLU activation functions, and fully connected layers. A noteworthy feature of ResNet-50 is its incorporation of skip connections, allowing the network to effectively capture both low- and high-level features by bypassing certain layers in the architecture. In this project, the features were extracted from the output of average pooling layer, which was the last layer before the classification layer with 2048 nodes.
- ConvNeXt [39] is a purely convolutional model designed to enhance the capabilities of convolutional neural networks by utilizing parallel branches for capturing distinct and supplementary features. This architecture, known for its varying number of layers, demonstrates a remarkable performance across a range of computer vision tasks, including image classification, object detection, and semantic segmentation. In this project, we used ConvNeXtSmall, which had 50 million parameters, and the features were extracted from the output of average pooling layer, which was the last layer before the classification layer and included 768 nodes.
3. Proposed Method
3.1. Preprocessing
3.2. Feature Extraction
3.2.1. CNN Features
3.2.2. Quantitative (Statistical/Spatial) Features
3.2.3. Incident Angle
3.3. Feature Concatenation
3.4. Feature Classification
- NN: in the NN method, we used two fully connected layers with a hidden layer, including 96 neurons and a single-neuron classification layer. To minimize the overfitting problem, we used the L2 norm kernel regularization with a factor of 0.01 and a drop-out layer with a parameter of 0.2. For the classification layer, we used a sigmoid activation function that provided the probability of the target being an iceberg.
- CatBoost: a robust gradient boosting classifier that employs a blend of ordered boosting and oblivious trees, alongside a novel strategy known as “ordered boosting.” It also incorporates innovations, like the “Bayesian bootstrap” technique, to enhance generalization [43]. Our efforts to enhance the model’s performance involved optimizing specific parameters: iterations = 500, depth = 6, learning rate = 0.2, border count = 125, and using log loss as the loss function. We selected the best values of depth and border count in order to minimize the overfitting problem. Default settings were retained for the remaining parameters, streamlining the process by reducing the requirement for extensive hyperparameter tuning.
- LightGBM: an advanced gradient boosting framework recognized for its effectiveness and adaptability in managing extensive datasets. Its distinguishing features include a novel histogram-based binning strategy, leading to faster training and decreased memory consumption [44]. To enhance our model’s effectiveness, we optimized specific parameters, such as the log loss metric, utilizing 31 leaves, a learning rate of 0.05, feature fraction of 0.9, and 200 boost rounds. Furthermore, for regularization, we set min child samples = 20 and depth = 6. Default settings were retained for the remaining parameters.
- Decision tree: after calculating the ship/iceberg probabilities using the three machine learning methods mentioned above, we utilized a single decision tree to fuse them and determine the final log loss. The decision tree classifier is a core classification algorithm that constructs a tree-like structure to classify the data based on their features [45]. In this study, as the tree involved only three inputs, we adopted the default parameter settings of the decision tree. These settings entailed using the Gini impurity as the splitting criterion, employing the ‘best’ strategy as the splitter, not imposing a maximum depth limit, and requiring a minimum of two samples for splitting the internal nodes.
4. Results and Discussion
4.1. Performance Evaluation of the Proposed Method for Different Models and Classifiers
4.2. Comparison of the Proposed Method with Existing Methods
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Amani, M.; Mehravar, S.; Asiyabi, R.M.; Moghimi, A.; Ghorbanian, A.; Ahmadi, S.A.; Ebrahimy, H.; Moghaddam, S.H.; Naboureh, A.; Ranjgar, B.; et al. Ocean remote sensing techniques and applications: A review (part ii). Water 2022, 14, 3401. [Google Scholar] [CrossRef]
- Chan, Y.K.; Koo, V. An introduction to synthetic aperture radar (SAR). Prog. Electromagn. Res. B 2008, 2, 27–60. [Google Scholar] [CrossRef]
- Jawak, S.D.; Bidawe, T.G.; Luis, A.J. A review on applications of imaging synthetic aperture radar with a special focus on cryospheric studies. Adv. Remote Sens. 2015, 4, 163. [Google Scholar] [CrossRef]
- Howell, C.; Power, D.; Lynch, M.; Dodge, K.; Bobby, P.; Randell, C.; Vachon, P.; Staples, G. Dual polarization detection of ships and icebergs-recent results with ENVISAT ASAR and data simulations of RADARSAT-2. In Proceedings of the InIGARSS 2008-2008 IEEE International Geoscience and Remote Sensing Symposium 2008, Boston, MA, USA, 7–11 July 2008; IEEE: New York, NY, USA, 2008; Volume 3, p. III-206. [Google Scholar]
- Howell, C.; Bobby, P.; Power, D.; Randell, C.; Parsons, L. Detecting icebergs in sea ice using dual polarized satellite radar imagery. In Proceedings of the InSNAME International Conference and Exhibition on Performance of Ships and Structures in Ice 2012, Banff, AB, Canada, 17–20 September 2012. [Google Scholar]
- Zhang, Y.; Hao, Y. A Survey of SAR Image Target Detection Based on Convolutional Neural Networks. Remote Sens. 2022, 14, 6240. [Google Scholar] [CrossRef]
- Fracastoro, G.; Magli, E.; Poggi, G.; Scarpa, G.; Valsesia, D.; Verdoliva, L. Deep learning methods for synthetic aperture radar image despeckling: An overview of trends and perspectives. IEEE Geosci. Remote Sens. Mag. 2021, 9, 29–51. [Google Scholar] [CrossRef]
- El Housseini, A.; Toumi, A.; Khenchaf, A. Deep Learning for target recognition from SAR images. In Proceedings of the 2017 Seminar on Detection Systems Architectures and Technologies (DAT) 2017, Algiers, Algeria, 20–22 February 2017; IEEE: New York, NY, USA, 2017; pp. 1–5. [Google Scholar]
- Zhong, C.; Mu, X.; He, X.; Wang, J.; Zhu, M. SAR target image classification based on transfer learning and model compression. IEEE Geosci. Remote Sens. Lett. 2018, 16, 412–416. [Google Scholar] [CrossRef]
- Malmgren-Hansen, D.; Kusk, A.; Dall, J.; Nielsen, A.A.; Engholm, R.; Skriver, H. Improving SAR automatic target recognition models with transfer learning from simulated data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1484–1488. [Google Scholar] [CrossRef]
- Geng, J.; Deng, X.; Ma, X.; Jiang, W. Transfer learning for SAR image classification via deep joint distribution adaptation networks. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5377–5392. [Google Scholar] [CrossRef]
- Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef]
- Bhatt, D.; Patel, C.; Talsania, H.; Patel, J.; Vaghela, R.; Pandya, S.; Modi, K.; Ghayvat, H. CNN variants for computer vision: History, architecture, application, challenges and future scope. Electronics 2021, 10, 2470. [Google Scholar] [CrossRef]
- Wang, P.; Zhang, H.; Patel, V.M. SAR image despeckling using a convolutional neural network. IEEE Signal Process. Lett. 2017, 24, 1763–1767. [Google Scholar] [CrossRef]
- Zhou, Y.N.; Luo, J.; Feng, L.; Zhou, X. DCN-based spatial features for improving parcel-based crop classification using high-resolution optical images and multi-temporal SAR data. Remote Sens. 2019, 11, 1619. [Google Scholar]
- Scarpa, G.; Gargiulo, M.; Mazza, A.; Gaetano, R. A CNN-based fusion method for feature extraction from sentinel data. Remote Sens. 2018, 10, 236. [Google Scholar] [CrossRef]
- Li, Y.; Ding, Z.; Zhang, C.; Wang, Y.; Chen, J. SAR ship detection based on resnet and transfer learning. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium 2019, Yokohama, Japan, 28 July–2 August 2019; IEEE: New York, NY, USA, 2019; pp. 1188–1191. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Wang, R.; Li, J.; Duan, Y.; Cao, H.; Zhao, Y. Study on the combined application of CFAR and deep learning in ship detection. J. Indian Soc. Remote Sens. 2018, 46, 1413–1421. [Google Scholar] [CrossRef]
- Wang, Y.; Wang, C.; Zhang, H. Combining a single shot multibox detector with transfer learning for ship detection using sentinel-1 SAR images. Remote Sens. Lett. 2018, 9, 780–788. [Google Scholar] [CrossRef]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
- Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; IEEE: New York, NY, USA, 2017; pp. 1–6. [Google Scholar]
- Kang, M.; Leng, X.; Lin, Z.; Ji, K. A modified faster R-CNN based on CFAR algorithm for SAR ship detection. In Proceedings of the 2017 International Workshop on Remote Sensing with Intelligent Processing (RSIP) 2017, Shanghai, China, 18–21 May 2017; IEEE: New York, NY, USA, 2017; pp. 1–4. [Google Scholar]
- Ge, J.; Zhang, B.; Wang, C.; Xu, C.; Tian, Z.; Xu, L. Azimuth-Sensitive Object Detection in Sar Images Using Improved Yolo V5 Model. In Proceedings of the IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium 2022, Kuala Lumpur, Malaysia, 17–22 July 2022; IEEE: New York, NY, USA, 2022; pp. 2171–2174. [Google Scholar]
- Sun, Z.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. BiFA-YOLO: A novel YOLO-based method for arbitrary-oriented ship detection in high-resolution SAR images. Remote Sens. 2021, 13, 4209. [Google Scholar] [CrossRef]
- Zhu, H.; Xie, Y.; Huang, H.; Jing, C.; Rong, Y.; Wang, C. DB-YOLO: A duplicate bilateral YOLO network for multi-scale ship detection in SAR images. Sensors 2021, 21, 8146. [Google Scholar] [CrossRef]
- Randell, C.; Freeman, R.; Power, D.; Stuckey, P. SS: Canadian: Atlantic development; technological advances to assess, manage and reduce ice risk in northern developments. In Proceedings of the Offshore Technology Conference 2009, Houston, TX, USA, 4–7 May 2009; p. OTC-20264. [Google Scholar]
- Heiselberg, H. Ship-iceberg classification in SAR and multispectral satellite images with neural networks. Remote Sens. 2020, 12, 2353. [Google Scholar] [CrossRef]
- Joachims, T. Making Large-Scale SVM Learning Practical; Technical Report; Universität Dortmund: Dortmund, Germany, 1998. [Google Scholar]
- Heiselberg, P.; Sørensen, K.A.; Heiselberg, H.; Andersen, O.B. SAR Ship–Iceberg Discrimination in Arctic Conditions Using Deep Learning. Remote Sens. 2022, 14, 2236. [Google Scholar] [CrossRef]
- Yang, X.; Ding, J. A computational framework for iceberg and ship discrimination: Case study on Kaggle competition. IEEE Access 2020, 8, 82320–82327. [Google Scholar] [CrossRef]
- Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining 2016, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
- Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 1–9. [Google Scholar]
- Gong, M.; Yang, H.; Zhang, P. Feature learning and change feature classification based on deep learning for ternary change detection in SAR images. ISPRS J. Photogramm. Remote Sens. 2017, 129, 212–225. [Google Scholar] [CrossRef]
- Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; pp. 1–11. [Google Scholar]
- Statoil/C-CORE. Statoil/C-CORE Iceberg Classifier Challenge 2018. Available online: https://www.kaggle.com/c/statoiliceberg-classifier-challenge (accessed on 1 April 2023).
- Lyu, H.; Huang, W.; Mahdianpari, M. Eastern arctic sea ice sensing: First results from the RADARSAT constellation mission data. Remote Sens. 2022, 14, 1165. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA, 19–20 June 2022; pp. 11976–11986. [Google Scholar]
- Jafari, Z.; Karami, E. Breast Cancer Detection in Mammography Images: A CNN-Based Approach with Feature Selection. Information 2023, 14, 410. [Google Scholar] [CrossRef]
- Ross, B.C. Mutual Information between Discrete and Continuous Data Sets. PLoS ONE 2014, 9, e87357. [Google Scholar] [CrossRef] [PubMed]
- Žunić, J.; Hirota, K.; Rosin, P.L. A Hu moment invariant as a shape circularity measure. Pattern Recognit. 2010, 43, 47–57. [Google Scholar] [CrossRef]
- Demir, S.; Sahin, E.K. Predicting occurrence of liquefaction-induced lateral spreading using gradient boosting algorithms integrated with particle swarm optimization: PSO-XGBoost, PSO-LightGBM, and PSO-CatBoost. Acta Geotech. 2023, 18, 3403–3419. [Google Scholar] [CrossRef]
- Nayak, J.; Naik, B.; Dash, P.B.; Souri, A.; Shanmuganathan, V. Hyper-parameter tuned light gradient boosting machine using memetic firefly algorithm for hand gesture recognition. Appl. Soft Comput. 2021, 107, 107478. [Google Scholar] [CrossRef]
- Shanthi, J.; Rani, D.G.; Rajaram, S. A C4. 5 decision tree classifier based floorplanning algorithm for System-on-Chip design. Microelectron. J. 2022, 121, 105361. [Google Scholar] [CrossRef]
Dataset | Number of Images | Polarizations | Image Size |
---|---|---|---|
Sentinel-1 | 1604 × 2 1 | HH, HV | 75 × 75 |
SSDD | 39,729 2 | HH, HV, VV, VH | 256 × 256 |
CNN Models | Layer Name 1 | Number of Feature |
---|---|---|
VGG16 | Fc2 | 4096 |
ResNet50 | Avg_pool | 2048 |
ConvNeXtSmall | Head_layer | 768 |
Bands | Sk | Ku | Ent | Min | Max | Ave | std | Area | Per | Rat | Ecc | Hu1 | Hu2 | Hu3 | Hu4 | Hu5 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HH | 0 | 0.04 | 0.15 | 0.06 | 0.14 | 0.11 | 0.01 | 0.01 | 0.05 | 0.04 | 0.11 | 0.11 | 0.09 | 0.09 | 0.09 | 0.07 |
HV | 0.01 | 0.12 | 0.10 | 0.07 | 0.17 | 0.13 | 0.01 | 0.01 | 0.02 | 0.02 | 0.09 | 0.09 | 0.08 | 0.08 | 0.08 | 0.05 |
HH − HV | 0.01 | 0.01 | 0.06 | 0.05 | 0.03 | 0.04 | 0 | 0.03 | 0.03 | 0.01 | 0.03 | 0.03 | 0 | 0 | 0 | 0.01 |
HH + HV | 0 | 0.11 | 0.15 | 0.08 | 0.18 | 0.12 | 0.01 | 0.02 | 0.05 | 0.02 | 0.10 | 0.10 | 0.09 | 0.09 | 0.09 | 0.05 |
Models | NN | LightGBM | CatBoost | Fused Classifier (Decision Tree) |
---|---|---|---|---|
VGG16 | 99% | 100% | 100% | 100% |
Resnet50 | 99.6% | 100% | 100% | 100% |
ConvNeXtSmall | 99.8% | 100% | 100% | 100% |
Concat.Model | 99.9% | 100% | 100% | 100% |
Models | NN | LightGBM | CatBoost | Fused Classifier (Decision Tree) |
---|---|---|---|---|
VGG16 | 91.3%/0.21 | 89.4%/0.26 | 90.2%/0.23 | 91.8%/0.20 |
Resnet50 | 92.2%/0.19 | 90.7%/0.22 | 91.1%/0.21 | 92.5%/0.18 |
ConvNeXtSmall | 90.7%/0.22 | 88.8%/0.27 | 89.7%/0.26 | 91.2%/0.21 |
Concat. CNNs | 93.1%/0.17 | 89.2%/0.26 | 91.8%/0.21 | 93.9%/0.16 |
Modified Method | NN | LightGBM | CatBoost | Fused Classifier |
---|---|---|---|---|
Proposed method with quantitative features only (49 features) | 91.8%/0.20 | 90.3%/0.23 | 90.7%/0.22 | 93.1%/0.17 |
Proposed method with comb. of quantitative and concat. CNN features (349 features) | 93.8%/0.16 | 91.8%/0.20 | 92.6%/0.18 | 95.4%/0.11 |
Method | Dataset | Accuracy/Log Loss |
---|---|---|
Method in [32] | SAR 1 | 88%/0.28 |
Method in [34] | SAR 2 | 93.4%/0.12 |
Method in [35] | SAR 1 | NA/0.12 |
Proposed method | SAR 1 | 95.4%/0.11 |
Proposed method | SAR 1 + SSDD | 96.1%/0.09 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jafari, Z.; Karami, E.; Taylor, R.; Bobby, P. Enhanced Ship/Iceberg Classification in SAR Images Using Feature Extraction and the Fusion of Machine Learning Algorithms. Remote Sens. 2023, 15, 5202. https://doi.org/10.3390/rs15215202
Jafari Z, Karami E, Taylor R, Bobby P. Enhanced Ship/Iceberg Classification in SAR Images Using Feature Extraction and the Fusion of Machine Learning Algorithms. Remote Sensing. 2023; 15(21):5202. https://doi.org/10.3390/rs15215202
Chicago/Turabian StyleJafari, Zahra, Ebrahim Karami, Rocky Taylor, and Pradeep Bobby. 2023. "Enhanced Ship/Iceberg Classification in SAR Images Using Feature Extraction and the Fusion of Machine Learning Algorithms" Remote Sensing 15, no. 21: 5202. https://doi.org/10.3390/rs15215202
APA StyleJafari, Z., Karami, E., Taylor, R., & Bobby, P. (2023). Enhanced Ship/Iceberg Classification in SAR Images Using Feature Extraction and the Fusion of Machine Learning Algorithms. Remote Sensing, 15(21), 5202. https://doi.org/10.3390/rs15215202