Next Article in Journal
A Dynamic Multi-Layer Steganography Approach Based on Arabic Letters’ Diacritics and Image Layers
Previous Article in Journal
Do Oral Antiseptics Affect the Force Degradation of Elastomeric Chains?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Aggregate Channel Features and Fast Regions CNN Approach for Classification of Ship and Iceberg

by
Sivapriya Sethu Ramasubiramanian
1,*,
Suresh Sivasubramaniyan
1 and
Mohamed Fathimal Peer Mohamed
2
1
Department of Computer Science and Engineering, SRM Institute of Science and Technology, Ramapuram, Chennai 600089, India
2
Department of Computer Science and Engineering, SRM Institute of Science and Technology, Vadapalani, Chennai 600026, India
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(12), 7292; https://doi.org/10.3390/app13127292
Submission received: 20 April 2023 / Revised: 2 June 2023 / Accepted: 17 June 2023 / Published: 19 June 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Detection and classification of icebergs and ships in synthetic aperture radar (SAR) images play a vital role in marine surveillance systems even though available adaptive threshold methods give satisfying results on detection and classification for ships and icebergs, including techniques of convolutional neural networks (CNNs), but need more accuracy and precision. An efficient and accurate method was developed to detect and classify the ship and icebergs. Hence, the research method proposed locating and classifying both ships and icebergs in a given SAR image with the help of deep learning (DL) and non-DL methods. A non-DL method utilized here was the aggregate channel features (ACF) detector, which extracts region proposals from huge SAR images. The DL object detector called fast regions CNN (FRCNN) detects objects accurately from the result of ACF since the ACF method avoids unwanted regions. The novelty of this study was that ACF-FRCNN concentrates only on accurately classifying ships and icebergs. The proposed ACF-FRCNN method gave a better performance in terms of loss (18.32%), accuracy (96.34%), recall (98.32%), precision (95.97%), and the F1 score (97.13%). Compared to other conventional methods, the combined effect of ACF and FRCNN increased the speed and quality of the detection of ships and icebergs. Thus, the ACF-FRCNN method is considered a novel method for over 75 × 75 resolution ship and iceberg SAR images.

1. Introduction

Remote sensing is a technique used to acquire data such as temperature, photographs, growth, pressure, and fire without direct contact with the target material. Remote sensing methodology has been increasingly applied in forest fire detection, iceberg detection, ship detection, floating object detection, agriculture monitoring, gathering pictures of the earth, etc. [1]. Many methods and tools such as radars are commonly used to capture remotely sensed image datasets [2], of which SAR image datasets have been playing a vital role. SARs can be used to construct two-dimensional images, which can be further reconstructed into multidimensional images if necessary. When the SAR focuses on the target application, such as biomass detection in the forest, ship detection in the ocean, and vehicle count on land, it comprises many small objects [3]. Detecting these small images is less effortless than general object detection in typical images. Specifically, observing ships or any biomass-detecting objects in applications such as marine surveillance has been challenging. The reason behind the difficulty is the number of pixels available in the small objects in a large image [4].
Over the past decade, new techniques have been developed for small object detection in remote sensing images. Due to image processing and definition complexity, detecting icebergs and ships still needs improvement. This detection technique has been studied for a long time. Deep learning has made rapid progress in the computer vision field.

2. Related Works on DL Methods of Small Object Detection

The CNN, a deep feed-forward artificial neural network, shows its different characteristics depending on the problem types it is applied to. CNN plays many roles in detecting ships but can detect only small objects. Unsupervised machine learning (ML) approaches such as principle component analysis (PCA) and the k-method have been used to discriminate between ships and icebergs. However, ML has to be extended to DL when a large dataset volume is there. Many remote sensing researchers have used CNN differently for their problems. A research-implemented solution was introduced for marine surveillance systems to discriminate between ships and icebergs [1]. They utilized four Conv2D layers and three dense layers with a 3 × 3 convolutional filter. They had a fully connected layer with 256 fully connected neurons. The authors used the pseudo-labeling approach since, in their dataset, only a small amount of labels were available. The transfer learning approach tackled the same problem, transferring the knowledge gained through a labeled SAR dataset to an unlabeled SAR dataset with future, present, and past information [2].
To classify icebergs and ships in multispectral satellite images, a C-Core dataset was used [3]. Two different CNNs were implemented with 561,217 and 1,134,081 parameters, respectively. Finally, those results were compared with support vector machine (SVM) results, and it was concluded that CNN had superior results in detecting ships. Iceberg detection was accomplished with the help of CNN, and it showed improved efficiency with the assistance of the Region of Interest (ROI), Sift, Surf, Threshold, and Transfer Learning [4]. Sidelobes were also included since the reflection of sidelobes may suppress large ships [5]. A constant false alarm rate was used to detect objects in ROI, followed by a parallel CNN, and in this approach, the parallel algorithm trained the model [6]. Moreover, a constant false alarm rate was also used after despeckling SAR images in supervised reinforcement learning [7].
SAR images is where CNN brought dominant accuracy results. The accuracy was improved by improving the features of objects in the training process [8]. The single-shot detector (SSD) is also used for object detection. Two major types of algorithms are widely available for object detection. Some algorithms such as RCNN, FRCNN, and Mask RCNN come under the first technique. There is a variety of CNN processing. Two detection stages occur. In the first stage, the region type of objects is expected to detect the object, and then objects are detected in those regions. The second technique is the fully convolutional approach. The you only look once (YOLO) approach and SSD are examples of this technique. These algorithms’ networks are capable of detecting images in a single pass. SSD works with the help of two components. SSD components are the backbone model and the single shot (SS) head.
VGG 16 was used as the backbone architecture [9], and the SSD method was treated as a baseline detector with a set of layers for the detection of objects. In this work, layers of SSD are responsible for specific scale objects. Large objects can be detected in the deeper layer, and small objects can be spotted in the shallow layer with lightweight objects [10]. Since SSD cannot detect tiny objects, the image pyramid network helped the model improve SSD’s performance. Since there was a need for oriented object information, in place of a conventional region proposal, network-oriented candidate region networks were used. At the same time, the feature fusion network (FFN) enhanced spatial information. This FFN is used for fusing IPN layers with SSD layers.
YOLO is a deep learning method that does not require dividing the images into blocks; the entire image can be looked at once. This speedy network is specifically designed for real-time object detection such as traffic signals, locating persons, finding animals, etc. It shows high accuracy with the help of three essential components: residual blocks, bounding box regression, and intersection over union, which are responsible for detecting, predicting, and locating, respectively. In [11], YOLO v2 was implemented to detect ships in marine environments with feature separation and feature alignment, which outperforms conventional neural networks with deeper layers. However, the number of parameters handled was still increased. To avoid this hyperparameter issue, they reduced the number of layers.
Tiny YOLO V3 [12] was used to detect small objects in aerial imagery, unlike conventional YOLO, which needs a large memory volume for better accuracy. However, tiny YOLO v3 needs a small amount of memory, even though there is a tradeoff between accuracy and memory. Another study [13] used conventional YOLO based on Darknet-53. Three-level pyramids were utilized for feature map creation. The Adam optimizer was implemented since the study used a small-size dataset. Finally, the model produced high precision. The generative adversarial network is a deep learning architecture that has the potential to work in both supervised and semi-supervised modes.
GAN is composed of two models: generative and discriminative models. The generator model uses the network, which generates images by adding random objects [14] and produces an input image that is similar to the actual image. The discriminator attempts to classify whether the generated image the generator generates looks like the actual input image. A super-resolution network was integrated with a cycle model based on GAN residuals [15]. Feature pyramid networks are networks [16] designed explicitly as feature extractors. They create high-quality multi-scale feature maps, which are significantly better than conventional feature extractors utilized in the CNN. Feature pyramid networks use two different pathways: bottom-up and top-down pathways [17].
More ship detection research solutions were recently proposed over the SAR imagery dataset. A method was developed with the DL algorithm for a SAR image dataset [18]. In this work, various proactive solutions and their mitigated risk were analyzed. An improved method was developed for detecting ships with SAR images [19]. A pre-training technique was adopted with DL, especially for scarcely labeled SAR image datasets. A novelty was achieved with a coarse-to-fine method for detecting ships with an optical remote sensing (OSR) image dataset [20]. In this work, ResNet was used alongside discrete wavelet transform (DWT) to achieve higher accuracy for detecting ships as a state-of-the-art method. Yet another approach was initiated to detect small ship-like objects in complex SAR datasets, and a hybrid method model was created alongside intersection over union loss prediction [21]. Here, shape similarity was identified for small-ship object detection. A ship detection technique was developed with a spatial SAR dataset with a shuffle group for a large scale [22]. This work introduced enhancement with the DL algorithm for ship detection.
A refinement network with feature balancing was used for multi-scale ship detection, and it was developed as an anchor-free method [23]. The DL algorithm used a SAR dataset for feature balancing in ship identification. A high-resolution ship detection mechanism was designed for a SAR dataset with large quantities [24]. This work developed a scattering and critical point-guided network with a core DL algorithm. Dense attention feature aggregation with CNN detected the ship over a SAR image dataset [25]. In this work, CNN was used anchor-free and featured dense attention. The modified mechanism with DL was used to develop a model for ship detection, and it was called a CenterNet++ mode [26]. A SAR image dataset was used for the seamless detection of the ship. A lightweight ship detection mechanism was introduced with YOLO v4, achieving high-speed detection of the ship over a SAR image dataset [27]. This work used three channels of Red, Green, and Blue (RGB) SAR images and achieved precise results. Recently, a target signature-based ship indicator was developed for complex signal kurtosis alongside a SAR image dataset [28]. This work introduced signals for complex image detection. A model was used to identify the ships on a high-resolution SAR image dataset tested with the DL algorithm [29]. This method was equipped with adaptive hierarchical detection. A lightweight neural network (NN)-based solution was recently introduced to detect ships over SAR image datasets [30]. This work tested a multi-scale ship identification mechanism for space-borne SAR images. Yet another work was developed for SAR-image-based ship detection, and it was technically named U-net [31]. In this work, the efficacy was achieved and tested for a low-cost ship detection model.
A few more works have introduced ship detection techniques with neural-network-based models [32,33,34,35]. A work discussed multiple downsampling for small objects with weaker features of small objects. Here, YOLOv5 was introduced to overcome irrelevant context. Feature enhancement was achieved with YOLOv5 with discriminative features of smaller objects [36]. Recently, GAN was used with ResNet 50 architecture with a feature pyramid network (FPN) to detect multi-class objects alongside the image enhancement process for a SAR dataset [37]. GAN was used to detect small, as well as medium-sized, objects. Yet another work used FPN with CNN for scale sequence feature detection over small objects. Scale-sequence-based FPN introduced considerable accuracy in detecting small objects [38].
In the bottom-up pathway, spatial resolution is half-reduced when it raises one level. It can be achieved by doubling the stride at each level. Each convolutional module produces the output that the top-down pathway will use. Let us take this output as Bi. The output of each module in the bottom-up pathway will be fed to 1 × 1 filters to reduce the depth of the Bi channel and to generate Ti. In the top-down path, which contrasts the bottom-up pathway, each layer will be upsampled by 2 when it goes down. The output of each module will be added element-wise before going to the next level. A problem arises when it is combined with an upsampled layer which is called aliasing. To recover ship images from the aliasing problem, 3 × 3 convolutions images are applied after merging [17].
The output of this 3 × 3 convolution layer is the final extracted feature that can be given to the object detector. If we take as a set of the production of the bottom-up pathway {B1, B2, B3, B4} and {T1, T2, T3, T4} as input to the top-down pathway, then {F1, F2, F3, F4} is a set of extracted features. The sample feature pyramid of a SAR image is given in Figure 1a,b, a top-down and bottom-up path.

3. Materials and Methods

Dataset: This work was carried out at SRM Institute of Science and Technology, Chennai, India. The implementation was carried out with MATLAB with the proposed ACF-FRCNN method and later compared the proposed method’s performance with existing methods for ship and iceberg detection and classification. The SAR dataset was collected, and it had images of a ship and iceberg with a resolution of 75 × 75 pixel values, each image with two bands [3]. A total of 1604 images, of both ships and icebergs, were taken for validation and verification with training and testing. Among dataset sizes, 80% were trained, and 20% were tested. Table 1 describes the dataset information and implementation details of the detection of classification of icebergs and ships with the SAR dataset.
Recent trending methodologies such as GAN, ACF Detector, and FRCNN were used for better object detection performance in the proposed system. The first step of the proposed method is generating multi-transformed images of a single image using GAN architecture from the SAR image dataset [3]. The GAN architecture augments different images with another direction at a distant object location. Then, all generated images undergo an ACF detector to generate ROI in each image. Then, FRCNN generates a feature vector of all region proposals nominated by ACF, followed by object detection and object classification. The proposed research’s methods are explained in detail in the following segments.

3.1. GAN Architecture Training

While performing a technical analysis in GAN, a winning model was discovered between the generator and discriminator. The generator tends to confuse the discriminator by producing data that look realistic but are not actually where the discriminator tries to distinguish images generated by generator and real ones. Both generator and discriminator have their loss function [2,3] for the initial process of dataset validation.

3.2. Loss Function of Generator

As mentioned earlier in the race of generator and discriminator, generator tries to confuse the discriminator by producing fake images. Let ȥ be a latent vector and g(ȥ) be data generated by the generator. Let d(g(ȥ)) be discriminator’s valuation of data generated by the generator. The loss function of the generator could be expressed as Equation (1).
LOSSg   =   ERR d g ȥ , 1
It is known that the model always tends to minimize the loss function. In this scenario, the generator also wants to minimize its loss function by reducing the difference between 1 and the discriminator valuation mark. There is a reason why the one is taken. Number 1 is given as a label for “true”.

3.3. Loss Function of Discriminator

The discriminator’s motive is to correctly identify the generated images of the generator as fake. Let d(r) be the valuation of the discriminator on accurate data. The loss function of the generator could be expressed as Equation (2).
LOSSd   =   ERR d r , 1 +   ERR d g ȥ , 0
Now take the generic loss function for binary classification, binary cross-entropy. Apply binary cross-entropy on the loss functions of the generator and discriminator. Express them, respectively, as Equations (3) and (4).
LOSSd   = r R , ȥ Zlog d r +   log 1 d g ȥ
LOSSg   = ȥ Zlog ( d g ȥ
The generator’s loss function will always be low because log (1) is zero. The next step is to optimize the model using loss functions. It is always better to train one model at a time. This model is suitable for considering the generator is fixed while training the discriminator, and the discriminator is set while training the generator.

3.4. Discriminator Training

Let the quantity of interest of the generator and discriminator be Qg and Qd. The value function in the amount of interest of the generator and discriminator is V(Qg Qd), expressed as Equation (5).
V(Qg Qd) = Er ﬥ pdata [log(d(r)) +Eȥ ﬥ pz log(1 − d(g(ȥ)))
Let y be g(ȥ); it is modified as Equations (6) and (7), respectively. The discriminator aims to maximize this value function.
V Qg   Qd =   E ř     pdata   log d r + E   pz   log 1 d y
V Qg   Qd = p   data   x log d r + p   g x   log 1 d y

3.5. Generator Training

As discussed earlier, while training, the generator discriminator remains fixed, and the value function is calculated as for the discriminator and expressed as Equation (8). However, the difference is generator now tends to minimize this value function.
V Qg   Qd =   Er     pdata   [ log Qd x + E ȥ     pz   log 1 Qd x

3.6. ACF Model

The next step in iceberg and ship detection is obtaining regional proposals. The proposed system uses an aggregate channel feature detector as a region proposal extraction method. Before exploiting the influence of aggregated channel features in extracting region proposals, there is a need to have a vision of channels. Beyond generally known color channels, such as gray scale, RGB, hue saturation value (HSV), and LUV, two more channels are available: gradient magnitude and gradient histograms. The gradient of magnitude and histograms is known as the histogram of oriented gradient (HoG) features. In addition, with color channels, HoG channels also play a vital role in extracting features in the computer vision world. For extracting features, color information alone will not give better results because in the case of edge detection, with the use of a color channel, it can be identified whether there is a pixel in the edge, but it is incapable of recognizing the direction of edges. It is also a must to know information about the object’s movement. It can be possible with the help of HoG features.
The idea in the proposed system is to aggregate all channels’ features using an ACF detector for region proposal extraction. If I is an input image, then H = Ω(I) is all channels of an image. After finding all channels, all blocks of pixels have to be summed up. As a result, features are single-pixel lookups in the aggregated channels. The role of ACF is set as RoI since ACF extracts ROI as the output of the ACF architecture model.
Thus, ACF gives fast feature pyramids with the set n number of layers, that is, Ƥ = { Ƥ1, Ƥ2, Ƥ3, Ƥ4, …… Ƥn,}. The proposed system used 3 LUV color channels, 6 HoG channels, and 1 normalized gradient channel. So a total of ten channels were used. The proposed ACF detector is shown in Figure 2. ACF can accelerate the model’s detection speed by extracting the right region proposal where the object could be present from the focus of a large SAR image. The extracted region proposals will be given as input to any detector. The proposed system uses FRCNN as an object detector network.

3.7. Object Detector—ResNet 50

The ResNet 50 is a residual network deep with 50 layers. ResNet avoids the vanishing gradient problem, which occurs in classic CNN, by using skip connections. The architecture diagram of ResNet 50 is shown in Figure 3. Colour varies depending on output size.

3.8. The FRCNN Method

Region-based convolutional neural network is specifically designed for object detection. The most important focus of a region-based convolutional network is generating region proposals using selective search. The developed region proposals will be resized and converted into a feature vector. A typical CNN will treat these feature vectors where pre-trained SVM will be a classifier to classify the objects into specific classes. The R-CNN has some drawbacks such as heavy storage requirements, a rigid structure that could be more flexible for customization, and time complexity due to selective search. Due to these disadvantages, it is extended to the next level: FRCNN. Basic units of FRCNN are pre-trained CNN, ROI pooling layer, and two fully connected layers for Softmax layer and bounding box regression. The pre-trained convolutional neural network used in the proposed system is ResNet 50 which is explained in the last part of the work.
The proposed fast RCNN takes a full input image and proposals generated by ACF. Since our input image is a vast SAR image where the iceberg and ship are microscopic objects in an image whose shape and size are very complex to predict, ACF output is used as region proposals. The initial step of the proposed network is generating a feature map for the entire input image with the help of multiple convolutional layers and a max pooling layer. The next step is creating a feature vector for each proposal that ACF produces. This model can be applied by extracting features from the feature map for each proposal and then converting them into fixed-length feature vectors with the help of the ROI pooling layer.
With the help of the max pooling layer, the ROI pooling layer picks all features of the region proposal and converts it into a fixed-size vector, which could be in the size of height × width where height and width are hyperparameters. If the ROI window is in the size of height × width, then a grid of sub-windows can be created by dividing the ROI window into sub-window sizes. For example, if their m × n grid of sub-windows is there, each sub-window or cell size could be height/m × width/n. The third step is feeding these feature vectors to a set of fully connected layers, which ends with an output layer of two branches. The first branch is the softmax layer for predicting object classes: background, iceberg, and ship. Another branch has four values as a bounding box for locating predicted objects.
If there are k ROIs, the first branch calculates the discrete probability for all k numbers of ROI as P = {p0, p1, p2, p3,… pk,}. In the second branch bounding box, regression will be performed for calculating bb = {bbx, bby, bbw, bbh} where bbx is the x coordinate, bby is the y coordinate, bbw is the width, and bbh is the height of the bounding box. The bounding box will be created for all ROIs. For calculating loss for RCNN, two different losses, each for one branch, are used for weight updating. The first loss is for class prediction, which can be calculated as in Equations (9)–(11).
LOSSclass P , T = log   PT
where log is a loss for true class T, and the second loss is for bounding box regression, as expressed in Equation (10).
Lloc bbpred , T = SMOOTHloc bbpredi Ti
where the notation “smoothloc” is represented as Equation (11).
SMOOTHloc   x = 0.5 x x   i f   x < 1 x 0.5   o t h e r w i s e
where bbpred is a predicted set of bounding boxes such as {bbpredx, bbpredy, bbpredw, bbpredh}; the proposed FRCNN architecture is given in Figure 4.
The algorithm of the proposed method FRCNN is structured below. It has procedure description, input, and output, and it is presented as Algorithm 1. Initially, the SAR input image is taken from the dataset and output such as class probability and bounding box set. Algorithm 1 procedure consists of computational functions such as RGB calculation of GAN and conversion of color space followed by LUV channel. The final LUV format image calculates magnitude and orientation. Then, the final image LUV is checked for ships or icebergs with a confidence score of region of proposal (ROP). This FRCNN must not be dropped to detect even small images, and the ROP confidence score is set at greater than 50% onwards with aggregate channel [21]. The next step of this FRCNN is to generate a feature map with ResNet with its appropriate layers. Next, the feature vectors are extracted for the image of LUV format with ROP of ACF, which underwent ResNet 50. Finally, it can find the probability and bounding box for inputting SAR images to detect ships and icebergs.
Algorithm 1: FRCNN method
Input: SAR Image I, Dataset
Output: Class Probability, Bounding Box
Procedure:
For g = 1 to Ng where Ng is Number of image generated by GAN
Calculate RGB channel for GAN[g]
Convert RGB channel into Colors space
Convert color space into LUV channel
Calculate Magnitude and Orientation of LUV image of m x n
    Mag(i,j) = √((∂GAN[g](i,j)/∂x)2 + (∂GAN[g](i,j)/∂y)2)
                        Ori(i,j) = tan-1((∂GAN[g](i,j)/∂x)/(∂GAN[g](i,j)/∂x))
                  Do Convolution of Magnitude Image
                  Do Normalization of Convoluted Image
        Compute HoG and Aggregate with LUV image
        If (Confidence > 50) extract ROP from Aggregate channel
// FRCNN working mechanism
Generate Feature Map for Input Image I using RESNET convolutional and Maxpool layer
Extract feature vector for each ROP of ACF
Input fixed size vector to fully connected layers of RESNET 50
Calculate discrete probabilities of all region proposals
P = {p0, p1, p2, p3,.. pk,}.
bb= {bbx, bby, bbw, bbh}
Calculate Lossclass (P, T) = -log PT which is log loss for true class T.
Lloc (bbpred,T) = ∑ smoothloc (bbpredi-Ti)
smooth loc   ( x ) = 0.5 x x   i f   x < 1 x 0.5   o t h e r w i s e   and
bbpred is predicted set of bounding box like {bbpredx, bbpredy, bbpredw, bbpredh
End for
For all ROI of I
Ibb = {IbbxI, Ibby, Ibbw, Ibbh} where Ibb is bounding boxes of ROIs of Input image
Pc = ∑ PPf-CNN/N where Pc is class probability of ROI of input image and PPf-CNN is output of fast CNN of all GAN images
End for
The combination and usage of all the technologies in the proposed framework are explained in the architecture diagram and depicted in Figure 5.

4. Experimental Result

The proposed algorithm was implemented in MATLAB. The C-core dataset was used in the implementation [3]. The images of the ship and iceberg were stored as JSON files. These four fields are in a JSON file. The ID of the image, band one and band 2 data flattened as a numerical value, incidence angle, and the last filed refer to whether the given ROI is an iceberg. If the field contains 0, it shows it is a ship; if it has 1, it is an iceberg. The images in the dataset are in 75 × 75 sizes and have two bands. The expert team gives the labels. The implementation started with augmenting images by using GAN architecture. Figure 6, left side, demonstrates the sample output of augmented images. Figure 6 on the right side shows the training fashion and the generator and discriminator scores.
Then, the experiment was continued with an ACF detector where each augmented image color channel and histogram of the gradient had to be calculated. Figure 7 shows the sample output of the HoG of the SAR image.
After calculating the aggregation of all channels, the ACF detector finds the region proposal with the help of the confidence score. Figure 8 and Figure 9 show the output of the ACF detector. The inference was that a ship or iceberg was detected if the confidence score gained more than 50% as per the computation of Algorithm 1 with the guidance of ACF.
Then, fast R-CNN started with the region proposals of ACF detectors with the help of ResNet 50 backbone architecture. Finally, the bounding box located the object with classifications labels, whether an iceberg or a ship. Figure 9 shows an example of iceberg classification, and Figure 10 shows an example of ship classification.

5. Performance Analysis of Results

Table 2 depicts the loss performance numerical illustration for four different methods: ACF, CNN, FRCNN, and ACF-FRCNN. This experimental loss performance was tested for the SAR dataset with an iterative method, and the number of data records was increased and tested with all methods as mentioned for the loss parameter. The inferences are that the loss gain is decreased for expanded records. The minimal loss for iteration 9 with 31.98%, 32.12%, 29.32%, and 18.32% was gained for ACF, CNN, FRCNN, and ACF-FRCNN. Among loss parameter performances, ACF-FRCNN had the best performance.
The proposed algorithm was compared for loss performance with CNN and ACF without fast RCNN and FRCNN without aggregate channel features. The analysis of loss is explained in the graph depicted in Figure 11.
Table 3 depicts the accuracy performance numerical illustration for four different methods ACF, CNN, FRCNN, and ACF-FRCNN. This experimental accuracy performance was tested for the SAR dataset with iteration, and the number of data records was increased and tested with all methods as mentioned for the accuracy parameter. The inference is that the accuracy gain is increased for records in the dataset. The maximum accuracy for iteration 8 with 86.12%, 87.92%, 89.14%, and 96.34% was achieved for ACF, CNN, FRCNN, and ACF-FRCNN, respectively. Among accuracy parameter performances, ACF-FRCNN had the best performance.
The proposed algorithm was compared for accuracy with CNN and ACF without FRCNN and FRCNN without ACF. The accuracy analysis is illustrated in the graph depicted in Figure 12.
The performance metrics for the proposed system are the F1 score, recall, and precision. Here, the metrics are calculated based on the is_iceberg attribute of the given JSON dataset file. The formula used to calculate precision, recall, and the F1 score is provided in Equations (12)–(14).
P r e c i s i o n = TP TP + FP
R e c a l l = TP TP + FN
F 1 s c o r e = 2 Precision Recall Precision Recall
where TP is true positive, FP is false positive, and FN is false negative. Table 4 shows the F1 score, recall, and precision performances’ numerical illustration for four different methods: ACF, CNN, FRCNN, and ACF-FRCNN. This experimental F1 score, recall, and precision performance was tested for the SAR dataset. The number of data records was increased and tried with all of the methods mentioned: F1 score, recall, and precision parameters. The inference is that the F1 score, recall, and precision were increased when the number of records increased. The maximum F1 score, recall, and precision were 97.13%, 98.32%, and 95.97%, respectively, for the proposed ACF-FRCNN method. Among the F1 score, recall, and precision parameter performances of all four methods, ACF-FRCNN had the best performance.
F1 score is treated as a better evaluation metric than accuracy since it is a harmonic mean of precision and recall. The proposed method analysis is compared with other methods and shown in Figure 13.

6. Discussion and Comparison

The conducted experiment used MATLAB for ship and iceberg detection. The observation was made after inference with the proposed ACF-FRCNN algorithm. The performance parameters accuracy, recall, precision, and F1 score were measured and analyzed. These performance parameters were also compared with the ACF, CNN, and FRCNN methods. The dataset used 1604 SAR images with a pixel resolution of 75 × 75. From Figure 11 to Figure 13, it was observed that ACF combined with the FRCNN method outperformed 20% of the dataset in detecting ships and icebergs. The reason for including ACF with FRCNN was that FRCNN is only an RGB channel and is considered only a region proposal. However, ACF with FRCNN can measure the gradient magnitude and histogram. Thereby, the proposed ACF-FRCNN can detect ships and icebergs speedily and with high quality. Hence the novelty was achieved by combining ACF with the FRCNN algorithm for ship and iceberg detection and classification.
The previous methods were analyzed to state the proposed method’s performance. It was noticed that YOLO v3 gained an accuracy of 51.16% in detecting ship and iceberg discrimination [13]. The faster RCNN algorithm achieved accuracy for ship detection with 89.8% [32] with the SAR image dataset. The SSD introduced small ship detection accuracy of 92% and comparatively lesser accuracy than the proposed ACF-FRCNN method [33] over the SAR image dataset. Yet another CenterNet method achieved 83.71% accuracy for the SAR image dataset for detecting ships [34]. Work was introduced to detect ship detection, and it was a detection transformer (DETR) [35]. DETR achieved accuracy in detecting ships with a SAR dataset of 57.8%. After the detailed comparison, it was observed that ACF-FRCNN achieved higher performance. Table 5 lists the accuracy performance value comparison of various methods, including the proposed ACF-FRCNN method.
The proposed ACF-FRCNN logic is a novelty since it produced optimal accuracy and faster detection of ships and icebergs over a SAR image dataset. It was also noticed that when the FRCNN was used to detect ships and icebergs directly, the FRCNN considered only color channels and missed some ROIs or added unwanted ROIs. However, when the proposed experiment initially fed ACF followed by FRCNN, the following outcome was introduced since ACF processed the region of the proposal with the variation of the gradient and the variation of the histograms. Thus, the outcome was optimum concerning accuracy, precision, etc., and detection speed was also increased. The YOLO method is often used for moving video frame applications such as traffic on roads, surveillance video systems, etc., for SAR image datasets, and the best effect to predict objects is the combined effect of the ACF and FRCNN methods, which is a novelty. In the future, the same proposed research work will be enhanced by using a high-resolution SAR image dataset, and we will also try to implement real-time automation for detecting ships and icebergs over the SAR image dataset.

7. Conclusions and Future Work

Small object detection in remote sensing images such as SAR images always plays a hectic role. For the past decade, various ML and DL algorithms have been applied to detect remotely sensed objects such as ships and icebergs. Even though some non-DL and DL-based algorithms are used to see ships and icebergs, performance measures still need improvement and efficiency. The proposed research reduces this complexity in detecting and classifying icebergs and ships with DL and non-DL combinations. The ACF detectors offer regions, and the FRCNN detector works on the proposals of ACF and produces accurate classification with a small amount of complexity. The proposed method achieved better numerical performance values: loss (18.32%), accuracy (96.34%), recall (98.32%), precision (95.97%), and the F1 score (97.13%). It was finally compared with other conventional methods, and it was claimed that the ACF and FRCNN methods outperformed and reduced the detector’s complexity. This framework can be extended with advanced pre-trained CNN and object detectors to include additional classes such as marine life. In the future, this proposed ACF and FRCNN algorithm will be developed with reinforcement concepts to train and automate remote sensing SAR-image-based submerged human body detection.

Author Contributions

Conceptualization, S.S.R. and S.S.; methodology, S.S.R.; software, S.S.; validation, M.F.P.M. and S.S.; formal analysis, S.S.R. and S.S.; investigation, S.S. and M.F.P.M.; writing—original draft preparation, S.S.R. and S.S.; writing—review and editing, M.F.P.M. and S.S.; supervision, S.S.; project administration, M.F.P.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received Synergy Facade funding with grant number SFC1023. I worked with a deep learning algorithm for a ship and iceberg detection mechanism.

Data Availability Statement

Data is unavailable due to privacy or ethical restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rane, A.; Sangili, V. Implementation of Improved Ship-IcebergClassifier Using Deep Learning. J. Intell. Syst. 2020, 29, 1514–1522. [Google Scholar] [CrossRef]
  2. Li, J.; Xu, C.; Su, H.; Gao, L.; Wang, T. Deep Learning for SAR Ship Detection: Past, Present and Future. Remote Sens. 2022, 14, 2712. [Google Scholar] [CrossRef]
  3. Heiselberg, H. Ship-Iceberg Classification in SAR and Multispectral Satellite Images with Neural Networks. Remote Sens. 2020, 12, 2353. [Google Scholar] [CrossRef]
  4. Pandey, K.S.; Tirthkar, S.; Gaikwad, S. Deep Learning for Iceberg Detection in Satellite Images. Int. Res. J. Eng. Technol. 2021, 8, 2395. [Google Scholar]
  5. Heiselberg, H. Ship-Iceberg Detection & Classification in Sentinel-1 SAR Images. Int. J. Mar. Navig. Saf. Sea Transp. 2020, 14, 235–241. [Google Scholar] [CrossRef]
  6. Song, L.; Peters, D.K.; Huang, W.; Power, D. Ship-iceberg discrimination from Sentinel-1 synthetic aperture radar data using parallel convolutional neural network. Concurr. Comput. Pract. Exp. 2021, 33, e6297. [Google Scholar] [CrossRef]
  7. Si, L.; Li, G.; Zheng, C.; Xu, F. Self-supervised Representation Learning for the Object Detection of Marine Radar. In Proceedings of the 8th International Conference on Computing and Artificial Intelligence, Tianjin, China, 18–21 March 2022; pp. 751–760. [Google Scholar] [CrossRef]
  8. Cheng, G.; Lang, C.; Wu, M.; Xie, X.; Yao, X.; Han, J. Feature Enhancement Network for Object Detection in Optical Remote Sensing Images. J. Remote Sens. 2021, 8, 9805389. [Google Scholar] [CrossRef]
  9. Shamsolmoali, P.; Zareapoor, M.; Granger, E.; Chanussot, J.; Yang, J. Enhanced single-shot detector for small object detection in remote sensing images. arXiv 2022, arXiv:2205.05927. [Google Scholar]
  10. Gao, Z.; Zhang, Y.; Wang, S. Lightweight Small Ship Detection Algorithm Combined with Infrared Characteristic Analysis for Autonomous Navigation. J. Mar. Sci. Eng. 2023, 11, 1114. [Google Scholar] [CrossRef]
  11. Zhu, M.; Hu, G.; Li, S.; Zhou, H.; Wang, S. FSFADet: Arbitrary-Oriented Ship Detection for SAR Images Based on Feature Separation and Feature Alignment. Neural Process. Lett. 2022, 54, 1995–2005. [Google Scholar] [CrossRef]
  12. Ajaz, A.; Salar, A.; Jamal, T.; Khan, A.U. Computer Vision and Pattern Recognition. arXiv 2022, arXiv:2203.04799v1. [Google Scholar]
  13. Hass, F.S.; Arsanjani, J.J. Deep Learning for Detecting and Classifying OceanObjects: Application of YoloV3 for Iceberg–Ship Discrimination. ISPRS Int. J. Geo-Inf. 2020, 9, 758. [Google Scholar] [CrossRef]
  14. Samanta, S.; Panda, M.; Ramasubbareddy, S.; Sankar, S.; Burgos, D. Spatial-Resolution Independent Object Detection Framework for Aerial Imagery. Comput. Mater. Contin. 2021, 68, 1937–1948. [Google Scholar] [CrossRef]
  15. Courtrai, L.; Pham, M.-T.; Lefevr, S. Small Object Detection in Remote Sensing Images Based on Super-Resolution with Auxiliary Generative Adversarial Networks. Remote Sens. 2020, 12, 3152. [Google Scholar] [CrossRef]
  16. Shamsolmoali, P.; Chanussot, J.; Zareapoor, M.; Zhou, H.; Yang, J. Multipatch Feature Pyramid Network for Weakly Supervised Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5610113. [Google Scholar] [CrossRef]
  17. Tian, Y.; Liu, J.; Zhu, S.; Xu, F.; Bai, G.; Liu, C. Ship Detection in Visible Remote Sensing Image Based on Saliency Extraction and Modified Channel Features. Remote Sens. 2022, 14, 3347. [Google Scholar] [CrossRef]
  18. Muhammad, Y.; Wan, J.; Xu, M.; Sheng, H.; Zeng, Z.; Liu, S.; Arife, T.I.C.; Sakaouth, H. Ship detection based on deep learning using SAR imagery: A systematic literature review. Soft Comput. 2023, 27, 63–84. [Google Scholar] [CrossRef]
  19. Bao, W.; Huang, M.; Zhang, Y.; Xu, Y.; Liu, X.; Xiang, X. Boosting ship detection in SAR images with complementary pretraining techniques. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8941–8954. [Google Scholar] [CrossRef]
  20. Chen, L.; Shi, W.; Fan, C.; Zou, L.; Deng, D. A novel coarse-to-fine method of ship detection in optical remote sensing images based on a deep residual dense network. Remote Sens. 2020, 12, 3115. [Google Scholar] [CrossRef]
  21. Chen, P.; Zhou, H.; Li, Y.; Liu, B.; Liu, P. Shape similarity intersection-over-union loss hybrid model for detection of synthetic aperture radar small ship objects in complex scenes. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9518–9529. [Google Scholar] [CrossRef]
  22. Cui, Z.; Wang, X.; Liu, N.; Cao, Z.; Yang, J. Ship detection in large-scale SAR images via spatial shuffle-group enhance attention. IEEE Trans. Geosci. Remote Sens. 2020, 59, 379–391. [Google Scholar] [CrossRef]
  23. Fu, J.; Sun, X.; Wang, Z.; Fu, K. An anchor-free method based on feature balancing and refinement network for multiscale ship detection in SAR images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 1331–1344. [Google Scholar] [CrossRef]
  24. Fu, K.; Fu, J.; Wang, Z.; Sun, X. Scattering-keypoint-guided network for oriented ship detection in high-resolution and large-scale SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 11162–11178. [Google Scholar] [CrossRef]
  25. Gao, F.; He, Y.; Wang, J.; Hussain, A.; Zhou, H. Anchor-free convolutional network with dense attention feature aggregation for ship detection in SAR images. Remote Sens. 2020, 12, 2619. [Google Scholar] [CrossRef]
  26. Guo, H.; Yang, X.; Wang, N.; Gao, X. A CenterNet++ model for ship detection in SAR images. Pattern Recogn. 2021, 112, 107787. [Google Scholar] [CrossRef]
  27. Jiang, J.; Fu, X.; Qin, R.; Wang, X.; Ma, Z. High-speed lightweight ship detection algorithm based on YOLO-v4 for three-channels RGB SAR image. Remote Sens. 2021, 13, 1909. [Google Scholar] [CrossRef]
  28. Leng, X.; Ji, K.; Xiong, B.; Kuang, G. Complex signal kurtosis—Indicator of ship target signature in SAR images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5208312. [Google Scholar] [CrossRef]
  29. Liang, Y.; Sun, K.; Zeng, Y.; Li, G.; Xing, M. An adaptive hierarchical detection method for ship targets in high-resolution SAR images. Remote Sens. 2020, 12, 303. [Google Scholar] [CrossRef] [Green Version]
  30. Liu, S.; Kong, W.; Chen, X.; Xu, M.; Yasir, M.; Zhao, L.; Li, J. Multi-scale ship detection algorithm based on a lightweight neural network for spaceborne SAR images. Remote Sens. 2022, 14, 1149. [Google Scholar] [CrossRef]
  31. Mao, Y.; Yang, Y.; Ma, Z.; Li, M.; Su, H.; Zhang, J. Efficient low-cost ship detection for SAR imagery based on simplified U-net. IEEE Access 2020, 8, 69742–69753. [Google Scholar] [CrossRef]
  32. Li, Y.; Zhang, S.; Wang, W.-Q. A Lightweight Faster R-CNN for Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4006105. [Google Scholar] [CrossRef]
  33. Zou, Y.; Zhao, L.; Qin, S.; Pan, M.; Li, Z. Ship target detection and identification based on SSD_MobilenetV2. In Proceedings of the IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 1676–1680. [Google Scholar] [CrossRef]
  34. Jiang, Y.; Li, W.; Liu, L. R-CenterNet+: Anchor-Free Detector for Ship Detection in SAR Images. Sensors 2021, 21, 5693. [Google Scholar] [CrossRef] [PubMed]
  35. Xing, Z.; Ren, J.; Fan, X.; Zhang, Y. S-DETR: A Transformer Model for Real-Time Detection of Marine Ships. J. Mar. Sci. Eng. 2023, 11, 696. [Google Scholar] [CrossRef]
  36. Wang, M.; Yang, W.; Wang, L.; Chen, D.; Wei, F.; KeZiErBieKe, H.; Liao, Y. FE-YOLOv5: Feature enhancement network based on YOLOv5 for small object detection. J. Vis. Commun. Image Represent. 2023, 90, 103752. [Google Scholar] [CrossRef]
  37. Ahmad, T.; Chen, X.; Saqlain, A.S.; Ma, Y. FPN-GAN: Multi-class Small Object Detection in Remote Sensing Images. In Proceedings of the IEEE 6th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), Chengdu, China, 24–26 April 2021; pp. 478–482. [Google Scholar] [CrossRef]
  38. Park, H.-J.; Kang, J.-W.; Kim, B.-G. ssFPN: Scale Sequence (S2) Feature-Based Feature Pyramid Network for Object Detection. Sensors 2023, 23, 4432. [Google Scholar] [CrossRef]
Figure 1. (a) Top-down pathway, (b) bottom-up pathway.
Figure 1. (a) Top-down pathway, (b) bottom-up pathway.
Applsci 13 07292 g001aApplsci 13 07292 g001b
Figure 2. Proposed ACF detector model.
Figure 2. Proposed ACF detector model.
Applsci 13 07292 g002
Figure 3. ResNet 50 architecture model.
Figure 3. ResNet 50 architecture model.
Applsci 13 07292 g003
Figure 4. Proposed FRCNN model.
Figure 4. Proposed FRCNN model.
Applsci 13 07292 g004
Figure 5. Proposed Framework’s Architecture Diagram.
Figure 5. Proposed Framework’s Architecture Diagram.
Applsci 13 07292 g005
Figure 6. Generated Image and Training Fashion.
Figure 6. Generated Image and Training Fashion.
Applsci 13 07292 g006
Figure 7. HoG of SAR image.
Figure 7. HoG of SAR image.
Applsci 13 07292 g007
Figure 8. Output of ACF detectors.
Figure 8. Output of ACF detectors.
Applsci 13 07292 g008
Figure 9. Output of FRCNN with class of iceberg.
Figure 9. Output of FRCNN with class of iceberg.
Applsci 13 07292 g009
Figure 10. Output of FRCNN with class of ship.
Figure 10. Output of FRCNN with class of ship.
Applsci 13 07292 g010
Figure 11. Percentage of loss parameter comparison among proposed and existing methods.
Figure 11. Percentage of loss parameter comparison among proposed and existing methods.
Applsci 13 07292 g011
Figure 12. Percentage of accuracy parameter comparison among proposed and existing methods.
Figure 12. Percentage of accuracy parameter comparison among proposed and existing methods.
Applsci 13 07292 g012
Figure 13. Performance metric graph for detection.
Figure 13. Performance metric graph for detection.
Applsci 13 07292 g013
Table 1. Accuracy comparison for ship detection over the SAR dataset with different methods.
Table 1. Accuracy comparison for ship detection over the SAR dataset with different methods.
ContentDescription
Method usedACF-FRCNN
DatasetSAR images of ships and icebergs
Volume of dataset1604
ImplementationMATLAB
Image resolution in pixels75 × 75
Training dataset size1283
Testing dataset size321
Table 2. Loss performance comparison for different methods for SAR dataset to detect ships and icebergs.
Table 2. Loss performance comparison for different methods for SAR dataset to detect ships and icebergs.
MethodIter.1Iter.2Iter.3Iter.4Iter.5Iter.6Iter.7Iter.8Iter.9
ACF [17]55.1653.8248.1242.0639.2637.4635.6634.8631.98
CNN [1]79.2376.2973.6570.1361.4352.5443.3134.2332.12
FRCNN [8] 69.4661.9753.2745.8737.2535.4333.9230.7629.32
ACF-FRCNN49.7945.2541.8737.8233.6529.4325.7321.4518.32
Table 3. Accuracy comparison for ship detection over the SAR dataset with different methods.
Table 3. Accuracy comparison for ship detection over the SAR dataset with different methods.
MethodIter.1Iter.2Iter.3Iter.4Iter.5Iter.6Iter.7Iter.8
ACF [17]44.2150.2956.9262.3468.7874.9780.6386.12
CNN [1]49.2358.3262.2367.3970.9275.7881.2187.92
FRCNN [8]56.1259.7667.1369.3274.8976.5485.7289.14
ACF-FRCNN65.1476.7879.5484.5687.8788.4389.1396.34
Table 4. Comparison of precision, recall, and F1 score performances for different methods.
Table 4. Comparison of precision, recall, and F1 score performances for different methods.
MethodPrecisionRecallF1 Score
ACF [17]84.7683.7684.26
CNN [1]88.1587.8387.99
FRCNN [8] 94.6595.3294.98
ACF-FRCNN95.9798.3297.13
Table 5. Various DL accuracy measure comparison with the proposed EDRL algorithm for ship and iceberg detection.
Table 5. Various DL accuracy measure comparison with the proposed EDRL algorithm for ship and iceberg detection.
Used MethodAccuracy
YOLO v3 [13]51.16%
Faster RCNN [32]89.8%
SSD [33]92%
CenterNet [34]83.71%
DETR [35]57.8%
Proposed ACF-FRCNN96.34%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sethu Ramasubiramanian, S.; Sivasubramaniyan, S.; Peer Mohamed, M.F. Aggregate Channel Features and Fast Regions CNN Approach for Classification of Ship and Iceberg. Appl. Sci. 2023, 13, 7292. https://doi.org/10.3390/app13127292

AMA Style

Sethu Ramasubiramanian S, Sivasubramaniyan S, Peer Mohamed MF. Aggregate Channel Features and Fast Regions CNN Approach for Classification of Ship and Iceberg. Applied Sciences. 2023; 13(12):7292. https://doi.org/10.3390/app13127292

Chicago/Turabian Style

Sethu Ramasubiramanian, Sivapriya, Suresh Sivasubramaniyan, and Mohamed Fathimal Peer Mohamed. 2023. "Aggregate Channel Features and Fast Regions CNN Approach for Classification of Ship and Iceberg" Applied Sciences 13, no. 12: 7292. https://doi.org/10.3390/app13127292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop