Next Article in Journal
Strong Modulation of Short Wind Waves and Ka-Band Radar Return Due to Internal Waves in the Presence of Surface Films. Theory and Experiment
Next Article in Special Issue
Classification of Sea Ice Types in the Arctic by Radar Echoes from SARAL/AltiKa
Previous Article in Journal
A Lightweight Keypoint-Based Oriented Object Detection of Remote Sensing Images
Previous Article in Special Issue
Hyperspectral Sea Ice Image Classification Based on the Spectral-Spatial-Joint Feature with the PCA Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RUF: Effective Sea Ice Floe Segmentation Using End-to-End RES-UNET-CRF with Dual Loss

1
Department of Systems Design Engineering, University of Waterloo, Waterloo, ON N2L G31, Canada
2
School of Medicine, Stanford University, Stanford, CA 94305, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(13), 2460; https://doi.org/10.3390/rs13132460
Submission received: 1 May 2021 / Revised: 3 June 2021 / Accepted: 8 June 2021 / Published: 24 June 2021
(This article belongs to the Special Issue Remote Sensing of Sea Ice and Icebergs)

Abstract

:
Sea ice observations through satellite imaging have led to advancements in environmental research, ship navigation, and ice hazard forecasting in cold regions. Machine learning and, recently, deep learning techniques are being explored by various researchers to process vast amounts of Synthetic Aperture Radar (SAR) data for detecting potential hazards in navigational routes. Detection of hazards such as sea ice floes in Marginal Ice Zones (MIZs) is quite challenging as the floes are often embedded in a multiscale ice cover composed of ice filaments and eddies in addition to floes. This study proposes a segmentation model tailored for detecting ice floes in SAR images. The model exploits the advantages of both convolutional neural networks and convolutional conditional random field (Conv-CRF) in a combined manner. The residual UNET (RES-UNET) computes expressive features to generate coarse segmentation maps while the Conv-CRF exploits the spatial co-occurrence pairwise potentials along with the RES-UNET unary/segmentation maps to generate final predictions. The whole pipeline is trained end-to-end using a dual loss function. This dual loss function is composed of a weighted average of binary cross entropy and soft dice loss. The comparison of experimental results with the conventional segmentation networks such as UNET, DeepLabV3, and FCN-8 demonstrates the effectiveness of the proposed architecture.

1. Introduction

Sea ice is one of the greatest physical constraints for shipping activities in the Arctic. Due to the lengthening of the open-water season, maritime traffic in the Arctic has increased three-fold over the past few years [1]. The story is similar in the Canadian Arctic as well, with Hudson Strait being the most traffic prone area [2]. However, although the ice extent and thickness are reducing across the Arctic, the risks and hazards involved in sailing through the Arctic are even more significant than in the past.
As the ice melts, the ice pack becomes more mobile, allowing hazards such as ice floes to break away. These ice floes move at high speeds in a dynamic fashion and can cause damage to vessels and man-made structures [3]. Ice charting is performed by the Canadian Ice Services (CIS) for estimating sea ice concentration and stage of development (which includes floes size information) to make management decisions for ensuring safety and efficient maritime activities in the Canadian Arctic [4]. CIS uses data from various data sources for the production of ice charts and the development of guidelines for mariners. One of the prominent data sources used for the task of sea ice monitoring is Synthetic Aperture Radar (SAR), which provides high spatial resolution images irrespective of the daylight conditions [5]. This data is manually interpreted, for which highly skilled personnel are required for the job.
To process the vast volumes of available data in real-time, there is a need for automated methods for ice floe detection. Previous studies [6,7,8,9,10,11] used image processing techniques coupled with traditional machine learning on SAR images for the task of ice–water segmentation and ice floe separation. SAR images usually have intrinsic speckle noise due to the coherent nature of the imaging process [12]. The presence of this speckle-noise has been identified as a limitation in classification accuracy [6,11]. To circumvent this issue, instead of using SAR images, some studies [13,14,15,16,17] used data from vessel mounted cameras for identifying ice floes. This approach solves the issue of speckle-noise contamination. However, geometric error compensation is required to tackle the problem of underestimation/overestimation of ice cover caused due to oblique sensor placement [14]. Information from shipboard sensors is also biased in the sense that the ships prefer to transit through regions with lower ice concentration.
In this paper, we present an end-to-end ResUnet-CRF (RUF)-architecture-based model with a dual loss for ice floe segmentation in SAR images. From the perspective of deep learning, the proposed RUF architecture integrates three main modules: encoder–decoder framework [18], deep residual connections [19], and a probabilistic graphical model [20]. The encoder–decoder framework allows the network to learn the latent space representation of the data. Such networks have been shown to work with image noise for tasks such as image deblurring [21] and super-resolution [22]. Hence, such a network was chosen for the present study to aid in dealing with speckle noise typically present in SAR images. Residual connections ease the network training and the conditional random field aids in the refinement of segmentation boundaries. We train our proposed network architecture in an end-to-end manner using a dual loss function. The dual loss [23] is a combination of BCE and Dice loss.
Our main contributions are as follows:
  • We propose a novel encoder–decoder-based deep residual network embedded with a dense probabilistic graphical model for sea ice floe segmentation. To the best of our knowledge, this is the first time such a network has been successfully implemented in the domain of sea ice segmentation.
  • Passive microwave data does not provide precise information about sea ice concentration (SIC) in low SIC areas with small ice floes due to a coarse resolution and low instrument sensitivity. Our method successfully detects ice floes in SAR images, especially in the regions with less than 20% SIC, which could be important for marine hazard monitoring and wildlife management.
  • The proposed approach, RUF, is able to achieve higher metric scores along with visually superior results when compared with standard state of the art segmentation backbone models such as FCN-8 and DeepLabV3. These results have been achieved with fewer weights than other leading approaches, as our approach uses 26 M parameters, while FCN-8 and DeepLabV3 use 54 M and 60 M parameters, respectively.
This paper is organized as follows: after a brief literature review (Section 2), the description of the study area, image database, and data annotation are provided (Section 3). Next, detailed information regarding the various components of our proposed network architecture is presented in Section 4, followed by the description of evaluation metrics (Section 5) used in this paper. Section 6 provides information regarding the experimental setup, conducted experiments, and obtained results. Finally, the paper ends with the conclusion and future improvements in Section 7.

2. Background

Sea ice charting is typically conducted by national ice services to identify the boundaries between ice and open water, and to identify the dominant ice types and ice concentration for a given region. In the past years, due to the improvement in both aerial and remote sensing sensors, numerous studies using sea ice data have emerged [24]. These studies cover ice–water segmentation [25,26,27], ice concentration estimation [28,29,30], ice thickness estimation [31,32], ice type classification [33,34], and sea ice feature detection [35,36]. Methods using superpixel segmentation [37,38], watershed segmentation [39,40], and active contours [41,42] have been actively employed in SAR image segmentation. There are several related studies in the area of ship detection in ice-covered waters [43,44,45]. The study in [43] used a novel approach, combining depthwise convolution and pointwise convolution to enable a lightweight and efficient network. This may be interesting to explore in future work. Many studies on ship detection use the SAR detection dataset (SSDD), which consists of quad-pol SAR imagery with spatial resolution of 1–15 m. The ships appear in these images as small bright regions. The study [45] also looked at a wide-swath Sentinel-1 image, which is comparable to the data source used here.
For the task of ice floe detection, various researchers have used different data platforms. Studies conducted by Hall et al. [13], Lu et al. [14], Heyn et al. [15,16], and Wang et al. [17] used vessel mounted camera sensors to obtain photographic data to identify ice floes. However, due to the oblique sensor placement, accurate measurements of sensor height, tilt, and focal length are required to calculate the geometric distortion. Moreover, compensation for ship sway is required for the success of these methods.
Images obtained from SAR provide a continuous stream of high spatial resolution data irrespective of the weather conditions and natural illumination. Earlier studies [6,7,8,9,10,11] aimed to solve the problem of ice floe identification in two steps. The first step involved the ice–water segmentation while the second step involved delineating different floes. Studies by Steer et al. and Toyota et al. [6,7] involved different thresholding methods for sea ice segmentation followed by morphological dilation/erosion operations to split different floes. Holt et al. [8] used local dynamic thresholding [46] and shrinking/growing algorithm [47] for floe segmentation. Hwang et al. [9] proposed a segmentation technique using Kernel Graph Cuts (KGC) [48] for ice–water segmentation and a combination of distance transformation, watershed [49], and a rule-based boundary revalidation processing for floe splitting [50]. Graphical models such as Markov Random Field and Conditional Random Field have also been used for the task of sea ice segmentation [10,11]. Due to the presence of speckle noise in SAR images, it can be difficult to segment sea ice floes using traditional machine learning techniques.
Recently, Convolutional Neural Networks (CNNs) have been proven to be good at learning the low- and high-level abstract features from raw images. Hence, they have been extensively used in tasks such as image classification [19,51,52,53], semantic segmentation [18,54], and object detection [55,56]. Long et al. [54] introduced a fully convolutional network for the task of image segmentation, while Chen et al. [57] proposed a combination of CNNs and CRFs to tackle poor localization of CNNs. Ronnerberger et al. [18] proposed an encoder–decoder-based network for the task of medical image segmentation. Later, Chen et al. proposed the DeepLab family networks [57,58,59] with dilated convolutions to reduce the computational complexity while maintaining the same receptive field.
Recently, Singh [60] et al. compared various segmentation models (e.g., DeepLab [57], UNet [18], SegNet [61], DenseNet [9]) for the task of river ice floe segmentation. Zhang et al. [62] introduced a convolutional network with dual attention streams for ice segmentation in rivers. Both of these studies use optical image datasets. To improve on pixel representations of CNNs and take advantage of residual learning, we propose RUF, an encoder–decoder network with residual blocks integrated with a convolutional CRF and trained in an end-to-end manner with dual loss function for the task of ice floe segmentation.

3. Dataset

The geographical area of interest for the dataset used in this paper spans over the Hudson Strait located in Eastern Canada, and its outflow into the Labrador Sea in the North Atlantic. The dataset is composed of 9 RADARSAT-2 C-band ScanSAR wide-beam mode images acquired in HH (horizontal transmit and horizontal receive) polarization. The images were acquired at a center frequency of 5.405 GHz with a 500-Km swath width and provide a nominal pixel spacing of 50 m. The SAR images were captured from the area as shown in Figure 1, with the red polygon describing the extent of one SAR image. Information regarding the image acquisition dates, central latitude and longitude, and instances of annotated floes is given in Table 1.

3.1. Data Preprocessing

SAR images have a grainy ‘salt and pepper’ appearance, also called speckle noise, which is caused due to random interference between coherent returns. To reduce this intrinsic contamination of speckle noise, the SAR images were downsampled four-fold. The downsampling operation was carried out by averaging over 4 × 4 pixel nonoverlapping blocks. Downsampling operation changed the nominal pixel spacing to 200 m with a reduction in data volume to 1 / 16 of the original. The local average filtering operation helps to reduce the speckle noise [63,64]. Downsampling is the result of this average filtering. The local average filtering also helps with reducing the data volume and makes it more manageable for training the neural network models.
Note that the Hudson Strait and the neighboring geographical areas of interest have a long coastline, and the images have significant land cover. As ice floes are a water body phenomenon, pixels representing land in the images were masked as black, as shown in Figure 2. To generate the land masks for the images, we applied a threshold of 0 m-elevation to the elevation masks and later a Gaussian blur with ( 5 , 5 ) kernel size to remove the rough edges along the sea–land boundaries. The elevation masks were generated using A C E 2 _ 5 M i n digital elevation model with bilinear interpolation using the European Space Agency (ESA) toolbox, Sentinel Application Platform.

3.2. Data Annotation

The dataset contains 9 images with 1627 manual annotations of ice floes. For our experiments, a single class was defined for annotation, namely, ‘floe’, and all the remaining pixels were categorized as background. In the pixel-level annotations, we used the following criteria to annotate a closed contour as a floe:
  • The contour contains consolidated ice as determined via visual inspection.
  • At least 30% of the contour boundary is in contact with seawater.
  • The contour contains at least 60 pixels.
These criteria were chosen to eliminate the closed contours in ice covers (frozen in floes) and to reduce noise artifacts. The floes were annotated through visual inspection of the imagery using the Computer Vision Annotation Tool (CVAT).

3.3. Dataset Split

We split the dataset into the train, validation, and test sets with a rough 60:20:20 ratio split, such that the training set contains 60 % of annotated floes while validation and test sets each contain 20 % of floe annotations. Full SAR images were placed in each set in order to have truly independent samples across different sets, as shown in Table 1. Splitting the dataset as such allows us to ascertain better generalization of the model.

4. Methodology

This section introduces the proposed RUF network architecture. The proposed model leverages the advantages of residual CNNs and Convolutional Conditional Random Fields (Conv-CRFs) [20] alongside a dual loss function. Residual blocks allow a network to train easily while the UNET skip connections help with easy information propagation between different layers of the network. To facilitate learning of the network weights, we jointly train the RES-UNET and Conv-CRF parts with the dual loss function. The overall architecture of the proposed network is illustrated in Figure 3. Information about various components of the network is given in the subsequent subsections.

4.1. RES-UNET

4.1.1. Residual Block

Neural network architectures with multiple layers, commonly known as deep neural networks, can learn richer features than their shallower counterparts [65,66]. Even though these deep neural network architectures are effective, they do struggle with a degradation problem where, upon adding more layers to an already deep neural network, the training accuracy of the model decreases. This is counter-intuitive since a deeper model should be able to fit the training data as well as a shallower model. To overcome this degradation problem, He et al. [19] proposed residual neural networks with stacked residual blocks. Given an input x l and output of the l-th residual unit x l + 1 , a residual block can be illustrated as
y l = h ( x l ) + R ( x l , w l ) , x l + 1 = f ( y l ) ,
where R ( · ) is the residual function, f ( y l ) is an activation function, and h ( x l ) is an identity mapping function.

4.1.2. UNET

The UNET [18] is a fully convolutional image segmentation architecture with symmetric downsampling and upsampling paths. To help with projecting the discriminatory features learned at different levels of the downsampling path, UNET uses skip connections. Skip connections help in integrating the location information in the downsampling paths to the contextual information in the upsampling paths. Rather than adding the input to the output in the case of residual blocks, a skip connection concatenates the input from the downsampling path to the output of the upsampling path.

4.2. Convolutional Conditional Random Field: Conv-CRF

In a semantic segmentation task, the pixelwise predictions of the CNN models are prone to having inaccurate boundaries. To reduce inaccuracies in the boundary, global and contextual information models such as CRFs can be used in conjunction with the CNNs [57].
In the case of semantic segmentation, the label of each pixel is x i { 1 , , X } , where i is a pixel in image I with N pixels. In modern approaches, a fully connected CRF (FC-CRF) takes the CNN’s output to compute the unary potentials [57,67]: Ψ u ( x i ) R X . Further, the pairwise potentials that account for the joint distribution of pixel pairs i , j as Ψ p ( x i , x j ) = μ ( x i , x j ) K ( f i , f j ) , where μ ( x i , x j ) is a compatibility transformation such as in Potts model μ ( x i , x j ) = [ x i x j ] , and K is a kernel function such as the Gaussian kernel function.
Conv-CRFs [20] have an add-on assumption of conditional independence over the FC-CRFs. In the Conv-CRF model, two pixels i , j are considered conditionally independent when their L1 norm is greater than a threshold: d ( i , j ) > k , where d ( · ) is the L1 norm and k is the distance threshold or the filter size. This means that the pairwise potential is zero for all the pixels at a distance greater than the threshold k. The Gibbs energy for a label sequence x can then be written as
E ( x ) = i N Ψ u ( x i ) + i = 1 N j = 1 k × k Ψ p ( x i , x j ) .
This greatly reduces the computational complexity. Teichmann et al. [20] also introduced a new message passing kernel that is similar to 2d convolutions of CNNs and can be efficiently implemented using convolutional libraries.
Efficient computations and exact message-passing lead to better run-time and performance when compared to FC-CRFs, which makes Conv-CRFs a better candidate for our architecture.

4.3. End-To-End Training

For training the proposed architecture, we first feed the SAR images to the base RES-UNET network, where pixelwise segmentation maps from the base network are fed to the Conv-CRF network as the unary. The Conv-CRF network cleans up spurious predictions and enhances object boundary predictions. Training these two parts in an end-to-end manner allows the gradients to flow through the whole pipeline and enables both networks to learn simultaneously. Hence, with this approach, we optimize both models with respect to each other to provide optimum results.

4.4. Dual Loss Function

For the task of multiclass classification, network weights are generally trained using the categorical cross entropy loss. In case of segmentation, losses involving ground truth and prediction overlap are generally employed. To train our network, we optimize the proposed RUF architecture using a weighted dual loss function including Binary Cross Entropy (BCE) loss and Soft Dice (SD) loss:
L t o t a l = α · L B C E + ( 1 α ) · L D i c e s ,
where α [ 0 , 0.5 , 1 ] is the weight parameter. BCE loss measures the classification accuracy of the model prediction and it increases as the prediction diverges from the ground truth [68]. SD loss, which is derived from Dice Coefficient, measures the similarity between two sets [69]. BCE loss and SD loss can be defined as Equations (4) and (5), respectively:
L B C E = i = 1 N [ y i log ( y ^ i ) + ( 1 y i ) log ( 1 y ^ i ) ] ,
L D i c e s = 1 2 i = 1 N y i y ^ i i = 1 N y ^ i + i = 1 N y i ,
where y i is the label or ground truth and y ^ i is the prediction for the ith pixel.

4.5. RUF Architecture

RUF is a five-level deep convolutional network with symmetric downsampling and upsampling paths, as shown in Figure 3. The downsampling path encodes the image into a condensed representation while the upsampling path decodes this information into pixelwise categorization. The downsampling path has four residual blocks. Each residual block contains multiple residual units built with two 3 × 3 convolutional layers and a residual connection. The convolutional layers are accompanied by a BatchNorm2d layer with a ReLu activation function. Rather than employing a max-pooling operation to downsample the feature maps [18], we use down-convolutional blocks with strided convolutions. Max-pooling downsamples the feature maps by taking the maximum value in pooling window to represent the pixels in that window, whereas strided convolutions allow the network to summarize the pixels in that receptive field. Strided convolutions allow the network to learn the spatial relationships without losing localization accuracy, as in max-pooling when downsampling is performed multiple times. A bottleneck in the network forces the model to compress the information to learn useful features from the previous layers. The upsampling path has a similar structure to the downsampling path. Feature maps are upsampled at each level using up conv blocks employing transposed convolutions. Skip connections concatenate the output of each level in the downsampling path to the upsampling path and help in combining coarse information with finer information. At the final level, a 1 × 1 convolution projects the multichannel feature maps to our intermediate segmentation mask. This mask is then processed in conjunction with the input image to calculate the unary and pairwise potentials of the Conv-CRF for further refinement. A softmax operation is applied to the output of Conv-CRF, which is later thresholded at 50 percent confidence to obtain the prediction mask.

5. Metrics

The following metrics were used to evaluate the proposed approach and perform a comparison with other standard segmentation approaches.
  • Mean Intersection over Union (mIoU): mIOU is a popular metric that is often used for the task of semantic segmentation. mIoU is calculated by averaging the Jaccard Score (J) over all the given classes in a segmentation task. Jaccard Score represents the ratio of the area of intersection between the ground truth (G) and predicted segmentation (P) maps to the union between ground truth and predicted segmentation maps:
    J ( G , P ) = | G P | | G P | ,
    where G and P are the ground truth and predicted segmentation maps, respectively.
  • Mean Pixel Accuracy (mPA): mPA is the ratio of correctly classified pixels per class, then averaged over the total number of classes. For k + 1 classes (k foreground classes and one background):
    m P A = 1 k + 1 i = 0 k p i i j = 0 k p i j .
  • F1 Score: F1 Score is the harmonic mean of Precision and Recall. Precision and Recall:
    P r e c i s i o n = T P T P + F P , R e c a l l = T P T P + F N ,
    where T P is the number true positives, F P is the number of false positives, and F N is the number of false negatives.
  • Dice Score: Dice Coefficient is the ratio of twice the area of intersection between ground truth and predicted segmentation maps, with the total number of pixels in the maps:
    D i c e = 2 | G P | | G | + | P | .
    Both Dice Coefficient and IoU are positively correlated. In the case of binary segmentation, where foreground class is considered the positive class, F1 Score and Dice Coefficient are identical:
    F 1 = 2 | G P | | G | + | P | = D i c e .
    Hence, for our binary segmentation problem of ice floes, in this approach, we use Dice Coefficient as F 1 Score hereafter.

6. Experiments and Results

6.1. Training Procedure

There are a total of 1140 annotated floes in the training set, which comprise approximately 60% of the overall annotated floes. Due to limited training images, we employ a random patch draw policy to train the model. An image is first randomly selected from the training set. We then randomly draw a patch of the given patch size from this selected image. Figure 2 illustrates the training patch selection process. A patch is considered a valid training sample if it does not contain more than 50% black area (due to image boundaries or land masking) or contains at least one floe either fully or partially. For example, a patch with 70% land and containing one floe is a valid sample. The above process is repeated until we find enough training samples for the training batch. We randomly rotate the eligible training samples by 0, 90, 180, or 270 degrees for data augmentation to increase the overall data available to train the models. This data augmentation provides a regularization effect that helps the models to generalize better on the overall dataset by reducing overfitting. The model training pipeline is illustrated in Figure 4.

6.2. Validation and Testing Procedure

The validation and test sets contain 2 images each, which account for approximately 20% of the overall annotated floes. To validate and test the model, patches are extracted serially from the images with an overlap of 50%. Figure 2 provides details about selecting validation and testing patches. The model validation and testing pipeline is illustrated in Figure 5. The validation dataset is used to check if the model is overfitting while the test dataset is used to compare different models.

6.3. Setup

The code was implemented using Pytorch 1.3.1 and Torchvision 0.4.2 open source frameworks. For all experiments, model weights were initialized using kaiming uniform initialization [70]. To optimize the model parameters further, we used the ADAM optimizer at an initial 1 × 10 4 learning rate. The Conv-CRF was initialized using default parameters from [20,71] with an exception, We removed the Gaussian blur as the SAR images in our dataset are already downsampled with 4-fold averaging.

6.4. UNET Backbone-Selection

To select the primary segmentation network for our pipeline, we first compared the UNET architecture with different backboned UNETs. The key difference between a UNET and a backboned-UNET is that we replace the two convolutional layers, and the 2 × 2 max-pooling operations for each level in the downsampling path are replaced with the different convolutional blocks of the backbone architecture, while the skip connections and the convolutional layers in the upsampling path remain the same. Information regarding the number of parameters and comparison between different UNET architectures is given in Table 2. We use VGG19, Inception V3, and ResNet34 to construct different UNET architectures. We observe that UNET architecture with a ResNet34-based encoder achieves the best scores on the validation set and is selected for further improvement. Refer to Table 2 for more information.

6.5. Joint Training with Conv-CRF

With the UNET backbone selected, we examine different configurations of joint training. We experiment with decoupled learning, stepwise learning, and end-to-end learning approaches. The results of which are illustrated in Table 3. Decoupled learning is based on the assumption that CRF needs an accurate prediction of the unary [71] to learn efficiently. In this approach, initially, only the CNN model is training till loss convergence, i.e., when the loss begins to saturate. After this step, the CRF is trained as a standalone model with the CNN output as unary to the CRF. The gradients never flow through the whole model in an iteration. Stepwise learning is similar to decoupled learning but in the second step, both the CNN and CRF are trained jointly such that weights of the whole model are updated in the second step. End-to-end learning involves training the whole architecture jointly from the very beginning such that the gradients flow through the whole network from the first epoch.
We observed that for our task, the end-to-end learning approach was more stable and yielded preeminent results. The main reason for the success of this approach is that when the CNN and CRF are trained together, they are able to coadapt easily with respect to each other. Thus, the end-to-end learning approach was used for further experiments.

6.6. Dual Loss Function Selection

To select the optimum weights of our dual loss function, we trained different RUF models using various α values for L t o t a l in Equation (3). We evaluated three different configurations with α [ 0 , 0.5 , 1 ] , which yield only BCE loss, equally weighted BCE and Dice loss, and only Dice loss, respectively. The results of our experiments are given in Table 4. We observe that using a dual loss comprising of equally weighted BCE and Dice loss helps the network to generate better predictions.
We observed that training with either only BCE or Dice loss did help the model loss to converge quickly but the predictions from such models lack discrete boundaries, as illustrated in Figure 6. BCE is a local loss that accounts for the correctness of individual pixels and is biased towards the majority class (background). Dice loss deals with the problem of unbalanced datasets by design, such that the network is disincentivized to ignore the minority class. But Dice loss can have high variance over batches as all object instances are given the same weight irrespective of the object size. As Dice loss focuses on the similarity of two sets and BCE loss is the sum of the distance between the corresponding pixels in ground truth and prediction mask, combining them allows the network to optimize the loss in both image-level and pixel-level domains. We observe that models trained using dual loss yield better metrics and visual predictions than those without, irrespective of using the CRF component, as in Figure 6.

6.7. Patch Size Selection

As the overall segmentation network configuration is selected, we investigate further to choose the optimum patch size to train our model. We trained various RUF models beginning with a patch size of 96 × 96 pixels and a batch size of 64 patches. We observed that this model was able to identify small floes, while it struggled to detect floes that appear partially in the patch. To overcome this issue, we decided to increase the patch size until we obtained optimum results. To train the subsequent models with bigger patch sizes, we decreased the batch size to obtain optimum GPU memory utilization. Results of our experiment are given in Table 5.
We observed a direct correlation between a bigger patch size and the segmentation accuracy, as evident from increasing metrics in Table 5—with an exception to the patch size of 576 × 576 pixels, where a considerable dip in performance was observed. To investigate this issue, we trained models with patch sizes 432 × 432 and 528 × 528 with various batch sizes. The results of this experiment are presented in Table 6. We observed that when the training patch size increases from 384 at 12 batch size to 432 at 10 batch size, the model’s performance increases. However, when the 432-sized patch’s batch size is decreased to 8, a dip in performance is observed. When we changed the training patch size from 480 at 8 batch size to 528 at 7 batch size, we observed that the model’s performance decreases, although the patch size has increased. A similar trend has also been discussed in [72], such that increasing the training batch size helps the model to learn and converge quickly while yielding better results. Due to the GPU memory constraint, a trade-off between patch size and batch size had to be made in order to obtain an optimum patch and batch size.

6.8. Comparison with Other Models

Using the optimal input size and other network parameters, we obtained the best results using our proposed method. We then compared our method with several other standard segmentation models that are typically used in SAR image segmentation. Test set images are used in this comparison. To gauge the iterative improvement component of our method, we plot the predictions of UNET, RES-UNET, and RUF together in Figure 7. Upon initial observation, we notice that both UNET and RES-UNET generate a greater number of false-positive predictions compared to RUF. We observe that RUF is able to detect floes of various shapes and sizes, while UNET struggles to detect bigger floes. UNET is also unable to delineate the floes in close proximity to each other. The main reason for this issue would be the absence of residual connections in the UNET, as the model is unable to learn finer features compared to other more complex models. It can also be observed that both UNET and RES-UNET generate multiple broken boundary predictions where a single floe is segmented as multiple floes.
We also compare RUF with other standard segmentation models such as FCN-8 (FCN) and DeepLabV3 (DLV3), the predictions of which are illustrated in Figure 8. We can observe that the predictions from both FCN and DeepLab do not suffer from false-positives; however, they do struggle to detect bigger floes and delineate floes in close proximity. The absence of a dense component such as a conditional random field to enhance finer segmentation details would explain this issue. Apart from these qualitative observations, the superiority of RUF for the task of ice floe segmentation is also evident from the quantitative analysis, as given in Table 7. Zoomed-in segmentation results of a patch are given in Figure 9.

7. Conclusions

In this paper, we presented a sea ice floe segmentation network for SAR images. RUF is introduced as a Conv-CRF embedded residual UNET, end-to-end trained using a dual loss function. With our model, we demonstrate that we can fully exploit the usefulness of Conv-CRF for the task of segmentation by integrating it inside a deep learning algorithm and training it in an end-to-end fashion. We performed extensive analysis for selecting model parameters to ensure the reliability of the proposed model. The experimental results confirm that our proposed model can achieve higher scores on all metrics and visually superior results that show a reduced number of false alarms and better representation of floe shape than other approaches on the limited dataset used in the study. It is noteworthy that this has been achieved with fewer weights than other leading approaches. We observed that in some instances, RUF was unable to detect floes of very large sizes. These floes were comparable to the prediction window/patch size. Upon investigation, we noticed that in our images, there were a greater number of mid-sized and small-sized floes compared to large-sized floes. In the future, we aim to use a much larger dataset over a wider geographic region to include a greater variety of floes. For example, including marginal ice zones, such as those of Antarctica, Baffin Bay, and the Beaufort Sea, would allow a combination of ice floes and larger icebergs in addition to multiyear ice floes. We would like to remove the 4x4 downsample and use full-resolution images. This may require more lightweight architectures, such as EfficientNet [73]. Our end goal is to provide floes from SAR that can be used for various purposes, such as the validation of sea ice concentration from passive microwave data, to provide information for wildlife habitat studies, and for studies of wave–ice interactions.

Author Contributions

A.S.N., D.K. and K.A.S. conceived and designed the framework of the study. A.S.N. and D.S. completed the data collection and processing. A.S.N. and D.K. completed the algorithm design and the data analysis. A.S.N. completed the experiments. A.S.N. was the lead author of the manuscript with contributions by D.K., D.S. and K.A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Canadian ArcticNet program and MEOPAR the Marine Environmental Observation Prediction and Response Network.

Data Availability Statement

The data is not publicly available due to 3rd party data policy.

Acknowledgments

The RADARSAT-2 data used in this paper was provided by the Canadian Ice Service (CIS). All SAR images are copyrighted to MacDONALD, Dettwiler and Associates Ltd. (2010)-All Rights Reserved. RADARSAT is an official mark of the Canadian Space Agency.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Barnhart, K.R.; Miller, C.R.; Overeem, I.; Kay, J.E. Mapping the future expansion of Arctic open water. Nat. Clim. Chang. 2016, 6, 280–285. [Google Scholar] [CrossRef]
  2. Andrews, J.; Babb, D.; Barber, D. Climate change and sea ice: Shipping in Hudson Bay, Hudson Strait, and Foxe Basin (1980–2016). Elem. Sci. Anth. 2018, 6, 19. [Google Scholar] [CrossRef] [Green Version]
  3. Barber, D.; Babb, D.; Ehn, J.; Chan, W.; Matthes, L.; Dalman, L.; Campbell, Y.; Harasyn, M.; Firoozy, N.; Theriault, N.; et al. Increasing mobility of high Arctic sea ice increases marine hazards off the east coast of Newfoundland. Geophys. Res. Lett. 2018, 45, 2370–2379. [Google Scholar] [CrossRef]
  4. Scheuchl, B.; Flett, D.; Caves, R.; Cumming, I. Potential of RADARSAT-2 data for operational sea ice monitoring. Can. J. Remote Sens. 2004, 30, 448–461. [Google Scholar] [CrossRef]
  5. Smirnov, V.; Bychkova, I. Satellite monitoring of ice features to ensure safety of offshore operations in the Arctic seas. Izv. Atmos. Ocean. Phys. 2015, 51, 935–942. [Google Scholar] [CrossRef]
  6. Steer, A.; Worby, A.; Heil, P. Observed changes in sea-ice floe size distribution during early summer in the western Weddell Sea. Deep Sea Res. Part II Top. Stud. Oceanogr. 2008, 55, 933–942. [Google Scholar] [CrossRef]
  7. Toyota, T.; Haas, C.; Tamura, T. Size distribution and shape properties of relatively small sea-ice floes in the Antarctic marginal ice zone in late winter. Deep Sea Res. Part II Top. Stud. Oceanogr. 2011, 58, 1182–1193. [Google Scholar] [CrossRef] [Green Version]
  8. Holt, B.; Martin, S. The effect of a storm on the 1992 summer sea ice cover of the Beaufort, Chukchi, and East Siberian Seas. J. Geophys. Res. Ocean. 2001, 106, 1017–1032. [Google Scholar] [CrossRef] [Green Version]
  9. Hwang, B.; Ren, J.; McCormack, S.; Berry, C.; Ayed, I.B.; Graber, H.C.; Aptoula, E. A practical algorithm for the retrieval of floe size distribution of Arctic sea ice from high-resolution satellite Synthetic Aperture Radar imagery. Elem. Sci. Anth. 2017, 5, 38. [Google Scholar] [CrossRef] [Green Version]
  10. Clausi, D.A.; Yue, B. Comparing cooccurrence probabilities and Markov random fields for texture analysis of SAR sea ice imagery. IEEE Trans. Geosci. Remote Sens. 2004, 42, 215–228. [Google Scholar] [CrossRef]
  11. Zhu, T.; Li, F.; Heygster, G.; Zhang, S. Antarctic Sea-Ice Classification Based on Conditional Random Fields From RADARSAT-2 Dual-Polarization Satellite Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2451–2467. [Google Scholar] [CrossRef]
  12. Cooke, C.L.V.; Scott, K.A. Estimating Sea Ice Concentration From SAR: Training Convolutional Neural Networks With Passive Microwave Data. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4735–4747. [Google Scholar] [CrossRef]
  13. Hall, R.J.; Hughes, N.; Wadhams, P. A systematic method of obtaining ice concentration measurements from ship-based observations. Cold Reg. Sci. Technol. 2002, 34, 97–102. [Google Scholar] [CrossRef]
  14. Lu, P.; Li, Z. A method of obtaining ice concentration and floe size from shipboard oblique sea ice images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2771–2780. [Google Scholar]
  15. Heyn, H.M.; Blanke, M.; Skjetne, R. Ice condition assessment using onboard accelerometers and statistical change detection. IEEE J. Ocean. Eng. 2019, 45, 898–914. [Google Scholar] [CrossRef] [Green Version]
  16. Heyn, H.M.; Knoche, M.; Zhang, Q.; Skjetne, R. A system for automated vision-based sea-ice concentration detection and floe-size distribution indication from an icebreaker. In Proceedings of the International Conference on Offshore Mechanics and Arctic Engineering, Trondheim, Norway, 25–30 June 2017; American Society of Mechanical Engineers: New York, NY, USA, 2017; Volume 57762, p. V008T07A012. [Google Scholar]
  17. Wang, Q.; Li, Z.; Lu, P.; Lei, R.; Cheng, B. 2014 summer Arctic sea ice thickness and concentration from shipborne observations. Int. J. Digit. Earth 2019, 12, 931–947. [Google Scholar] [CrossRef]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical iMage Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. Teichmann, M.T.T.; Cipolla, R. Convolutional CRFs for Semantic Segmentation. arXiv 2018, arXiv:cs.CV/1805.04777. [Google Scholar]
  21. Chen, G.; Gao, Z.; Wang, Q.; Luo, Q. U-net like deep autoencoders for deblurring atmospheric turbulence. J. Electron. Imaging 2019, 28, 053024. [Google Scholar] [CrossRef]
  22. Hu, X.; Naiel, M.A.; Wong, A.; Lamm, M.; Fieguth, P. RUNet: A Robust UNet Architecture for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  23. Lei, Y.; Tian, S.; He, X.; Wang, T.; Wang, B.; Patel, P.; Jani, A.B.; Mao, H.; Curran, W.J.; Liu, T.; et al. Ultrasound prostate segmentation based on multidirectional deeply supervised V-Net. Med. Phys. 2019, 46, 3194–3206. [Google Scholar] [CrossRef]
  24. Haugen, J.; Imsland, L.; Løset, S.; Skjetne, R. Ice observer system for ice management operations. In Proceedings of the Twenty-First International Offshore and Polar Engineering Conference, Maui, HI, USA, 19–24 June 2011. [Google Scholar]
  25. Leigh, S.; Wang, Z.; Clausi, D.A. Automated ice–water classification using dual polarization SAR satellite imagery. IEEE Trans. Geosci. Remote Sens. 2013, 52, 5529–5539. [Google Scholar] [CrossRef]
  26. Hoekstra, M.; Jiang, M.; Clausi, D.A.; Duguay, C. Lake Ice-Water Classification of RADARSAT-2 Images by Integrating IRGS Segmentation with Pixel-Based Random Forest Labeling. Remote Sens. 2020, 12, 1425. [Google Scholar] [CrossRef]
  27. Xie, T.; Perrie, W.; Wei, C.; Zhao, L. Discrimination of open water from sea ice in the Labrador Sea using quad-polarized synthetic aperture radar. Remote Sens. Environ. 2020, 247, 111948. [Google Scholar] [CrossRef]
  28. Wang, L.; Scott, K.A.; Xu, L.; Clausi, D.A. Sea ice concentration estimation during melt from dual-pol SAR scenes using deep convolutional neural networks: A case study. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4524–4533. [Google Scholar] [CrossRef]
  29. Cheng, A.; Casati, B.; Tivy, A.; Zagon, T.; Lemieux, J.F.; Tremblay, L.B. Accuracy and inter-analyst agreement of visually estimated sea ice concentrations in Canadian Ice Service ice charts using single-polarization RADARSAT-2. Cryosphere 2020, 14, 1289–1310. [Google Scholar] [CrossRef] [Green Version]
  30. Karvonen, J. Baltic sea ice concentration estimation using SENTINEL-1 SAR and AMSR2 microwave radiometer data. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2871–2883. [Google Scholar] [CrossRef]
  31. Shi, L.; Karvonen, J.; Cheng, B.; Vihma, T.; Lin, M.; Liu, Y.; Wang, Q.; Jia, Y. Sea ice thickness retrieval from SAR imagery over Bohai sea. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; pp. 4864–4867. [Google Scholar]
  32. Karvonen, J.; Shi, L.; Cheng, B.; Similä, M.; Mäkynen, M.; Vihma, T. Bohai Sea ice parameter estimation based on thermodynamic ice model and Earth observation data. Remote Sens. 2017, 9, 234. [Google Scholar] [CrossRef] [Green Version]
  33. Zakhvatkina, N.Y.; Alexandrov, V.Y.; Johannessen, O.M.; Sandven, S.; Frolov, I.Y. Classification of sea ice types in ENVISAT synthetic aperture radar images. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2587–2600. [Google Scholar] [CrossRef]
  34. Park, J.W.; Korosov, A.A.; Babiker, M.; Won, J.S.; Hansen, M.W.; Kim, H.C. Classification of Sea Ice Types in Sentinel-1 SAR images. Cryosphere Discuss. 2019, 2019, 1–23. [Google Scholar]
  35. Nagi, A.S.; Minhas, M.S.; Xu, L.; Scott, K.A. A Multi-Scale Technique to Detect Marginal Ice Zones Using Convolutional Neural Networks. In Proceedings of the IGARSS 2020-2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 3035–3038. [Google Scholar]
  36. Sola, D.; Nagi, A.S.; Scott, K.A. Identifying Sea Ice Ridging in SAR Imagery Using Convolutional Neural Networks. In Proceedings of the IGARSS 2020-2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 6930–6933. [Google Scholar]
  37. Lang, F.; Yang, J.; Yan, S.; Qin, F. Superpixel segmentation of polarimetric synthetic aperture radar (sar) images based on generalized mean shift. Remote Sens. 2018, 10, 1592. [Google Scholar] [CrossRef] [Green Version]
  38. Zhang, W.; Xiang, D.; Su, Y. Fast Multiscale Superpixel Segmentation for SAR Imagery. IEEE Geosci. Remote. Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  39. Ciecholewski, M. River channel segmentation in polarimetric SAR images: Watershed transform combined with average contrast maximisation. Expert Syst. Appl. 2017, 82, 196–215. [Google Scholar] [CrossRef]
  40. Ijitona, T.B.; Ren, J.; Hwang, P.B. SAR sea ice image segmentation using watershed with intensity-based region merging. In Proceedings of the 2014 IEEE International Conference on Computer and Information Technology, Xi’an, China, 11–13 September 2014; pp. 168–172. [Google Scholar]
  41. Braga, A.M.; Marques, R.C.; Rodrigues, F.A.; Medeiros, F.N. A median regularized level set for hierarchical segmentation of SAR images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1171–1175. [Google Scholar] [CrossRef]
  42. Jin, R.; Yin, J.; Zhou, W.; Yang, J. Level set segmentation algorithm for high-resolution polarimetric SAR images based on a heterogeneous clutter model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 4565–4579. [Google Scholar] [CrossRef]
  43. Zhang, T.; Zhang, X.; Shi, J.; Wei, S. Depthwise separable convolution neural network for high-speed SAR ship detection. Remote Sens. 2019, 11, 2483. [Google Scholar] [CrossRef] [Green Version]
  44. Zhang, T.; Zhang, X. High-speed ship detection in SAR images based on a grid convolutional neural network. Remote Sens. 2019, 11, 1206. [Google Scholar] [CrossRef] [Green Version]
  45. Zhang, T.; Zhang, X.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.; Pan, D.; Li, J.; Su, H.; Zhou, Y. LS-SSDD-v1.0: A Deep Learning Dataset Dedicated to Small Ship Detection from Large-Scale Sentinel-1 SAR Images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  46. Haverkamp, D.; Soh, L.K.; Tsatsoulis, C. A comprehensive, automated approach to determining sea ice thickness from SAR data. IEEE Trans. Geosci. Remote Sens. 1995, 33, 46–57. [Google Scholar] [CrossRef] [Green Version]
  47. Tsatsoulis, C.; Kwok, R. Analysis of SAR Data of the Polar Oceans: Recent Advances; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  48. Salah, M.B.; Mitiche, A.; Ayed, I.B. Multiregion image segmentation by parametric kernel graph cuts. IEEE Trans. Image Process. 2010, 20, 545–557. [Google Scholar] [CrossRef] [PubMed]
  49. Kato, Z.; Berthod, M.; Zerubia, J. A hierarchical Markov random field model and multitemperature annealing for parallel image classification. Graph. Model. Image Process. 1996, 58, 18–37. [Google Scholar] [CrossRef]
  50. Ren, J.; Hwang, B.; Murray, P.; Sakhalkar, S.; McCormack, S. Effective SAR sea ice image segmentation and touch floe separation using a combined multi-stage approach. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 1040–1043. [Google Scholar]
  51. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  52. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  53. Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Darrell, T.; Keutzer, K. Densenet: Implementing efficient convnet descriptor pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
  54. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  55. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 142–158. [Google Scholar] [CrossRef]
  56. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  57. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  58. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  59. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  60. Singh, A.; Kalke, H.; Loewen, M.; Ray, N. River ice segmentation with deep learning. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7570–7579. [Google Scholar] [CrossRef] [Green Version]
  61. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  62. Zhang, X.; Jin, J.; Lan, Z.; Li, C.; Fan, M.; Wang, Y.; Yu, X.; Zhang, Y. ICENET: A Semantic Segmentation Deep Network for River Ice by Fusing Positional and Channel-Wise Attentive Features. Remote Sens. 2020, 12, 221. [Google Scholar] [CrossRef] [Green Version]
  63. Pogson, L.; Geldsetzer, T.; Buehner, M.; Carrieres, T.; Ross, M.; Scott, K.A. Collecting empirically derived SAR characteristic values over one year of sea ice environments for use in data assimilation. Mon. Weather Rev. 2017, 145, 323–334. [Google Scholar] [CrossRef]
  64. Clausi, D.; Qin, A.; Chowdhury, M.; Yu, P.; Maillard, P. MAGIC: Map-guided ice classification system for operational analysis. In Proceedings of the 2008 IAPR Workshop on Pattern Recognition in Remote Sensing (PRRS 2008), Tampa, FL, USA, 7 December 2008; pp. 1–4. [Google Scholar]
  65. Rolnick, D.; Tegmark, M. The power of deeper networks for expressing natural functions. arXiv 2017, arXiv:1705.05502. [Google Scholar]
  66. Mhaskar, H.; Liao, Q.; Poggio, T. Learning functions: When is deep better than shallow. arXiv 2016, arXiv:1603.00988. [Google Scholar]
  67. Zheng, S.; Jayasumana, S.; Romera-Paredes, B.; Vineet, V.; Su, Z.; Du, D.; Huang, C.; Torr, P.H. Conditional random fields as recurrent neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1529–1537. [Google Scholar]
  68. Ma, Y.-D.; Liu, Q.; Qian, A.B. Automated image segmentation using improved PCNN model based on cross-entropy. In Proceedings of the 2004 International Symposium on Intelligent Multimedia, Video and Speech Processing, Hong Kong, China, 20–22 October 2004; pp. 743–746. [Google Scholar]
  69. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2017; pp. 240–248. [Google Scholar]
  70. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1026–1034. [Google Scholar]
  71. Krähenbühl, P.; Koltun, V. Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials. arXiv 2012, arXiv:cs.CV/1210.5644. [Google Scholar]
  72. Smith, S.L.; Kindermans, P.J.; Ying, C.; Le, Q.V. Don’t decay the learning rate, increase the batch size. arXiv 2017, arXiv:1711.00489. [Google Scholar]
  73. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
Figure 1. The geographical area of interest of captured SAR scenes. Red polygon represents the geographic extent of a SAR image used in the study.
Figure 1. The geographical area of interest of captured SAR scenes. Red polygon represents the geographic extent of a SAR image used in the study.
Remotesensing 13 02460 g001
Figure 2. Patch generation for training, validation, and testing. Training: input SAR image ( h × w ) is land masked and patches of size 480 × 480 are selected randomly from the image. For a patch to be considered as a valid training sample, it should not contain more than 50% black pixels. Validation and testing: land masked SAR image is padded using black pixels. Patches are drawn serially from the land masked SAR image with a patch overlap of 50%. Total training/validation patches extracted from an image are ( h p × w p ) ÷ ( 480 × 480 ) . Image acquired on date 2012-06-06 at time 10:44:18 along the Canadian east coast with central coordinate 60°42′8″N, 65°1′W.
Figure 2. Patch generation for training, validation, and testing. Training: input SAR image ( h × w ) is land masked and patches of size 480 × 480 are selected randomly from the image. For a patch to be considered as a valid training sample, it should not contain more than 50% black pixels. Validation and testing: land masked SAR image is padded using black pixels. Patches are drawn serially from the land masked SAR image with a patch overlap of 50%. Total training/validation patches extracted from an image are ( h p × w p ) ÷ ( 480 × 480 ) . Image acquired on date 2012-06-06 at time 10:44:18 along the Canadian east coast with central coordinate 60°42′8″N, 65°1′W.
Remotesensing 13 02460 g002
Figure 3. Figure of the proposed network architecture. The lower branch of the architecture consists of the downsampling/encoding path. The upper branch of the architecture consists of the upsampling/decoding path. Skip connections (interconnections between the two parts) help with integrating the location information in the downsampling paths to the contextual information in the upsampling path. The output of the final layer (loosely labeled mask) in the upsampling path is fed to the Conv-CRF as unary. Conv-CRF processes the loosely labeled mask along with input image to generate the Prediction Mask.
Figure 3. Figure of the proposed network architecture. The lower branch of the architecture consists of the downsampling/encoding path. The upper branch of the architecture consists of the upsampling/decoding path. Skip connections (interconnections between the two parts) help with integrating the location information in the downsampling paths to the contextual information in the upsampling path. The output of the final layer (loosely labeled mask) in the upsampling path is fed to the Conv-CRF as unary. Conv-CRF processes the loosely labeled mask along with input image to generate the Prediction Mask.
Remotesensing 13 02460 g003
Figure 4. Model training pipeline. Input SAR image is processed as illustrated in Figure 2 to extract patches for the training batch. These patches are fed to the model and the model parameters are updated according to the loss.
Figure 4. Model training pipeline. Input SAR image is processed as illustrated in Figure 2 to extract patches for the training batch. These patches are fed to the model and the model parameters are updated according to the loss.
Remotesensing 13 02460 g004
Figure 5. Model testing pipeline. Input SAR image is processed as illustrated in Figure 2 to feed patches serially to the trained RUF model for inference. These patches are then reconstructed to yield the segmentation mask for the whole image.
Figure 5. Model testing pipeline. Input SAR image is processed as illustrated in Figure 2 to feed patches serially to the trained RUF model for inference. These patches are then reconstructed to yield the segmentation mask for the whole image.
Remotesensing 13 02460 g005
Figure 6. Comparison between the predictions of RUF model trained using BCE, Dice, and BCE + Dice loss. It can be observed that the predictions obtained using BCE + Dice loss have discrete continuous boundaries similar to the ground truth (GT).
Figure 6. Comparison between the predictions of RUF model trained using BCE, Dice, and BCE + Dice loss. It can be observed that the predictions obtained using BCE + Dice loss have discrete continuous boundaries similar to the ground truth (GT).
Remotesensing 13 02460 g006
Figure 7. Comparison between the predictions of UNET, RES-UNET, and RUF for the task of ice floe segmentation on patches from the test set. Red ovals highlight segmentation results where both UNET and RES-UNET are unable to properly segment a fully or partially visible ice floe. Blue ovals highlight segmentation results where both UNET and RES-UNET produced many false-positive predictions. It can be observed that the proposed RUF architecture is able to produce finer segmentation results compared to other backbone architectures on our dataset. Here, Patch and GT denote original image patch input and ground truth, respectively. The sea ice concentration (SIC) masks are overlaid on the respective patches for comparison between sea ice as continuous (passive microwave data) vs. discrete (processed ice floe masks from SAR) information.
Figure 7. Comparison between the predictions of UNET, RES-UNET, and RUF for the task of ice floe segmentation on patches from the test set. Red ovals highlight segmentation results where both UNET and RES-UNET are unable to properly segment a fully or partially visible ice floe. Blue ovals highlight segmentation results where both UNET and RES-UNET produced many false-positive predictions. It can be observed that the proposed RUF architecture is able to produce finer segmentation results compared to other backbone architectures on our dataset. Here, Patch and GT denote original image patch input and ground truth, respectively. The sea ice concentration (SIC) masks are overlaid on the respective patches for comparison between sea ice as continuous (passive microwave data) vs. discrete (processed ice floe masks from SAR) information.
Remotesensing 13 02460 g007
Figure 8. Comparison between the predictions of FCN-8, DeepLabV3, and RUF for the task of ice floe segmentation on patches from the test set. Red ovals highlight segmentation results where both FCN and DLV3 are unable to properly segment a fully or partially visible ice floe. Green ovals highlight segmentation results where both FCN and DLV3 are unable to delineate floes in close proximity. It can be observed that the proposed RUF architecture is able to produce finer segmentation results compared to the other frequently used segmentation methods in satellite imaging on our dataset. Here, Patch and GT denote original image patch input and ground truth, respectively. The sea ice concentration (SIC) masks are overlaid on the respective patches for comparison between sea ice as continuous (passive microwave data) vs. discrete (processed ice floe masks from SAR) information.
Figure 8. Comparison between the predictions of FCN-8, DeepLabV3, and RUF for the task of ice floe segmentation on patches from the test set. Red ovals highlight segmentation results where both FCN and DLV3 are unable to properly segment a fully or partially visible ice floe. Green ovals highlight segmentation results where both FCN and DLV3 are unable to delineate floes in close proximity. It can be observed that the proposed RUF architecture is able to produce finer segmentation results compared to the other frequently used segmentation methods in satellite imaging on our dataset. Here, Patch and GT denote original image patch input and ground truth, respectively. The sea ice concentration (SIC) masks are overlaid on the respective patches for comparison between sea ice as continuous (passive microwave data) vs. discrete (processed ice floe masks from SAR) information.
Remotesensing 13 02460 g008
Figure 9. Zoom-in on the segmentation result of a patch. Here, Patch and GT denote original image patch input and ground truth, respectively, where GT are the floes that were annotated manually according to our selection criteria. The sea ice concentration (SIC) mask is overlaid on the patch for comparison between sea ice as a continuous (passive microwave data) vs. discrete (processed ice floe masks from SAR) information.
Figure 9. Zoom-in on the segmentation result of a patch. Here, Patch and GT denote original image patch input and ground truth, respectively, where GT are the floes that were annotated manually according to our selection criteria. The sea ice concentration (SIC) mask is overlaid on the patch for comparison between sea ice as a continuous (passive microwave data) vs. discrete (processed ice floe masks from SAR) information.
Remotesensing 13 02460 g009
Table 1. Description of RADARSAT-2 scenes dataset that was used to train, validate, and test various model architectures.
Table 1. Description of RADARSAT-2 scenes dataset that was used to train, validate, and test various model architectures.
DatasetScene AcquisitionNumber of FloesCentral Coordinate
DateTime
Train2008-11-1311:29:0020463°32′40″N, 75°22′25″W
2011-06-1222:12:5320860°28′50″N, 68°47′22″W
2012-06-0522:42:5917762°38′24″N, 77°2′1″W
2013-06-0311:25:5434862°47′46″N, 72°38′1″W
2013-06-1422:34:3520362°38′14″N, 74°56′25″W
Total 1140
Validation2012-05-0811:29:4921263°32′40″N, 75°22′25″W
2012-06-0610:44:1812860°42′8″N, 65°1′W
Total 340
Test2011-06-1510:56:385962°20′26″N, 67°29′54″W
2014-06-1511:29:4028863°30′41″N, 75°22′W
Total 347
Table 2. Comparison of different UNET backbones on validation set. Best results are in bold.
Table 2. Comparison of different UNET backbones on validation set. Best results are in bold.
ModelParametersmIOUF1mPA
UNET7,788,54570.5058.7774.83
VGG-UNET20,267,28171.5260.7173.62
Inception-UNET27,161,26471.5960.8974.68
RES-UNET25,658,40272.5762.8880.40
Table 3. Comparison of different RUF joint training approaches on validation set. Best results are in bold.
Table 3. Comparison of different RUF joint training approaches on validation set. Best results are in bold.
MethodmIoUF1mPA
Decoupled68.2554.2372.90
Stepwise69.7357.2774.87
End-to-end75.0767.2881.83
Table 4. Comparison of different loss functions for various model architectures using end-to-end training approach on validation set. BCE—Binary cross entropy loss; Dice—Dice loss; BCE + Dice—Equally weighted BCE and Dice Loss. Best results are in bold.
Table 4. Comparison of different loss functions for various model architectures using end-to-end training approach on validation set. BCE—Binary cross entropy loss; Dice—Dice loss; BCE + Dice—Equally weighted BCE and Dice Loss. Best results are in bold.
ModelLoss Function
BCEDiceDice + BCE
mIoUF1mPAmIoUF1mPAmIoUF1mPA
UNET65.7648.6666.3965.6448.6171.9670.5158.7874.83
RES-UNET70.8359.4074.4170.3258.4475.4172.5762.8880.40
RUF (Ours)70.9159.5975.3671.3960.4874.6175.0767.2881.83
Table 5. Comparison of different models for various patch sizes on validation set. Best results are in bold.
Table 5. Comparison of different models for various patch sizes on validation set. Best results are in bold.
Patch SizeBatch SizemIoUF1mPA
966470.5358.8174.09
1923270.7759.3877.65
2881670.7559.2474.16
3841271.9959.7774.40
480875.0767.2881.83
576669.8557.4472.13
Table 6. Comparison of patch size and batch size on validation set. Best results are in bold.
Table 6. Comparison of patch size and batch size on validation set. Best results are in bold.
Patch SizeBatch SizemIoUF1mPA
3841271.9959.7774.40
4321073.7763.9878.38
432871.8760.1874.82
480875.0767.2881.83
528773.6563.7477.87
528669.1256.9771.65
576669.8557.4472.13
Table 7. Comparison between various segmentation architectures on test set. Best results are in bold.
Table 7. Comparison between various segmentation architectures on test set. Best results are in bold.
ModelParametersmIoUF1mPA
UNET7,788,54569.2056.5471.87
RES-UNET25,658,40271.7661.2474.66
FCN854,304,08669.8857.4872.24
DLV360,991,06267.4752.4368.94
RUF (ours)26,119,20874.5266.2677.88
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nagi, A.S.; Kumar, D.; Sola, D.; Scott, K.A. RUF: Effective Sea Ice Floe Segmentation Using End-to-End RES-UNET-CRF with Dual Loss. Remote Sens. 2021, 13, 2460. https://doi.org/10.3390/rs13132460

AMA Style

Nagi AS, Kumar D, Sola D, Scott KA. RUF: Effective Sea Ice Floe Segmentation Using End-to-End RES-UNET-CRF with Dual Loss. Remote Sensing. 2021; 13(13):2460. https://doi.org/10.3390/rs13132460

Chicago/Turabian Style

Nagi, Anmol Sharan, Devinder Kumar, Daniel Sola, and K. Andrea Scott. 2021. "RUF: Effective Sea Ice Floe Segmentation Using End-to-End RES-UNET-CRF with Dual Loss" Remote Sensing 13, no. 13: 2460. https://doi.org/10.3390/rs13132460

APA Style

Nagi, A. S., Kumar, D., Sola, D., & Scott, K. A. (2021). RUF: Effective Sea Ice Floe Segmentation Using End-to-End RES-UNET-CRF with Dual Loss. Remote Sensing, 13(13), 2460. https://doi.org/10.3390/rs13132460

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop