Next Article in Journal
A White-Box Sociolinguistic Model for Gender Detection
Previous Article in Journal
Modelling Approach for the Prediction of Machinability in Al6061 Composites by Electrical Discharge Machining
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RoiSeg: An Effective Moving Object Segmentation Approach Based on Region-of-Interest with Unsupervised Learning

1
School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China
2
School of Computer Science and Technology, Soochow University, Suzhou 215006, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(5), 2674; https://doi.org/10.3390/app12052674
Submission received: 7 February 2022 / Revised: 27 February 2022 / Accepted: 1 March 2022 / Published: 4 March 2022
(This article belongs to the Topic Applied Computer Vision and Pattern Recognition)

Abstract

:
Traditional video object segmentation often has low detection speed and inaccurate results due to the jitter caused by the pan-and-tilt or hand-held devices. Deep neural network (DNN) has been widely adopted to address these problems; however, it relies on a large number of annotated data and high-performance computing units. Therefore, DNN is not suitable for some special scenarios (e.g., no prior knowledge or powerful computing ability). In this paper, we propose RoiSeg, an effective moving object segmentation approach based on Region-of-Interest (ROI), which utilizes unsupervised learning method to achieve automatic segmentation of moving objects. Specifically, we first hypothesize that the central n × n pixels of images act as the ROI to represent the features of the segmented moving object. Second, we pool the ROI to a central point of the foreground to simplify the segmentation problem into a classification problem based on ROI. Third but not the least, we implement a trajectory-based classifier and an online updating mechanism to address the classification problem and the compensation of class imbalance, respectively. We conduct extensive experiments to evaluate the performance of RoiSeg and the experimental results demonstrate that RoiSeg is more accurate and faster compared with other segmentation algorithms. Moreover, RoiSeg not only effectively handles ambient lighting changes, fog, salt and pepper noise, but also has a good ability to deal with camera jitter and windy scenes.

1. Introduction

Many researchers have proposed efficient solutions to solve foreground detection in video object segmentation problems. Among these solutions, deep neural network (DNN) methods have impressive performance with high accuracy. However, DNN architectures need enough datasets and time to train the network for improving the accuracy, which makes it not suitable for some special scenarios without enough training samples (e.g., detection of air-dropped objects in military operations) or with strict time requirements (e.g., the interception of a cannon against a shell). Moreover, these DNN-based methods also require high-performance computing units to complete all the tasks, which is too expensive for ordinary people. Background subtraction and frame difference are commonly adopted in solving the video object segmentation problems [1,2]. There are several challenges existing in background subtraction and frame difference, such as including various illumination changes, camera jitter, dynamic background, camouflage, shadows, bootstrapping and video noise [3,4]. Although many useful algorithms for background modeling have been designed, their performance is limited due to the complexity of algorithms, for example, background subtraction and the modeling of a scene based on each pixel of each frame [5]. Moreover, the accuracy of these algorithms is to some extent effected by wind noise or camera jitter [6].
To deal with these challenges, we propose RoiSeg, an effective object segmentation approach based on Region-of-interest (ROI), which utilizes unsupervised learning method to achieve automatic segmentation of moving objects. RoiSeg hypothesizes the central n∗n pixels of images as the ROI to reflect the features of moving object, then the classification of all pixels is turned into that of ROI central points. In the field of classification, the supervised learning methods usually provide a better accuracy compared with the unsupervised learning methods, however, they inevitably need more annotated datasets, hence increasing the workload of computing units [7]. To address this problem, RoiSeg adopts an automatic generation method based on ROI to produce the training samples with the unsupervised learning method. Moreover, RoiSeg also implements an online sample classifier to compensate the imbalance of different classes.
We highlight our main contributions as follows:
  • We propose RoiSeg, an effective object segmentation approach based on ROI, which utilizes unsupervised learning method to achieve automatic segmentation of moving objects. RoiSeg not only effectively handles ambient lighting changes, fog, salt and pepper noise, but also has a good ability to deal with camera jitter and windy scenes.
  • We hypothesize the central n*n pixels as the ROI and simplify the foreground segmentation into a classification problem based on ROI. In addition, we propose an automatic generation method to produce the training samples and implement an online sample classifier to compensate the imbalance of different classes, respectively.
  • We also conduct extensive experiments to evaluate the performance of RoiSeg and the experimental results demonstrate that RoiSeg is more accurate and faster compared with other segmentation algorithms.
The rest of this paper is organized as follows. Section 2 presents a review of related works. The description of RoiSeg is demonstrated in Section 3. The comparison experiments are given in Section 4. Finally, the conclusion is drawn in Section 5.

2. Related Work

Video segmentation has attracted great attention and many researchers have proposed to use DNN methods to solve this problem due to its impressive performance in this field. However, DNN is obviously not suitable for scenarios with a small/no training samples. Background subtraction, a crucial step in video object segmentation has attracted great attention in the last two decades. The main idea of background subtraction is to build a background model with a fixed number of frames. This model can be designed by different methods, such as statistical, fuzzy, neuro-inspired, and so forth. Among these methods, statistical methods have been intensively studied and widely used in various applications [8,9,10,11]. For example, Xue et al. developed a message passing algorithm termed offline denoising-based turbo message passing subtracting the background successfully with a lower mean squared error and better visual quality for both offline and online compressed video background subtraction [12]. Stauffer and Grimson implemented a parametric probabilistic background model [13]. In this model, distributions of each pixel color updated through an online expectation-minimization algorithm, were represented by a sum of weighted Gaussian distributions defined in a given color space: the Gaussian Mixture Model (GMM). Culibrk et al. adopted a neural network to determine whether each pixel of the image belongs to the foreground or the background [14]. Yu et al. established a spatio-color model based on both foreground and background, which used Expectation Maximization (EM) to track the parameters of GMM [15]. Gallego et al. used EM in the same way but modeled foreground and background at the region level and pixel level, respectively [16]. Cuevas and Garcia proposed an algorithm for foreground extraction and background updating using fuzzy functions and modeled both foreground and background in a non-parametric way [17].
These algorithms mainly implemented foreground detection on each pixel of a frame and may not be able to segment some parts of the background into foreground, resulting in a lower accuracy than DNN methods. However, these algorithms can provide some real-time results to meet some time-sensitive tasks. In addition, it is useful to first segment an ROI by frame difference before clustering and classification. In this paper, we propose RoiSeg, an effective object segmentation method based on ROI, which utilizes unsupervised learning to improve the accuracy of foreground segmentation and ensure the real-time performance. RoiSeg includes two crucial designing methods, namely clustering and classification methods.
Cluster analysis is a statistical multivariate analysis technique which is a common method of unsupervised machine learning [18]. It divides a set of data points into several classes, with the data points in each cluster being very similar but the data points in different clusters being very different [19]. K-means is an excellent clustering method based on segmentation. It iteratively calculates the distance from each point to the K-cluster center, so that K clusters can be found in a given data set [20]. Seiffert et al. presented an efficient initial seed selection method, RDBI, to improve the performance of the K-means filtering method by locating the seed points at dense, well-separated areas of the dataset [21]. Nidheesh et al. presented an improved, density-based version of K-Means, the key idea of which is to select as the initial centroids data points which belong to dense regions and which are adequately separated in feature space [22]. The Gaussian mixed model (GMM) is a classic statistical model, in which samples are generated by a Gaussian mixture distribution and the expectation maximization (EM) algorithm is used to update the parameters of the model [23]. Unlike traditional methods of cluster analysis based on heuristic or distance-based procedures, finite mixture modeling provides a formal statistical framework on which to base the clustering procedure [24]. Theoretically, all the data points can fit as long as the GMM has enough components, but the relationship between the number of modes and the number of components in the mixture is very complex so it is particularly important to determine the number of components. In this paper, we only use two unsupervised clustering algorithms: the GMM and K-Means.
The popular Naive Bayesian classifier performs well in dealing with discrete data [25]. Naive Bayes can perform surprisingly well in classification tasks where the probability, itself calculated by the Naive Bayes classifier, is not important. In recent years, many scholars have studied Naive Bayesian classifiers and suggested several algorithms to improve their predictive accuracy [26,27,28]. However, classifiers trained with imbalanced data tend to generate results with a high true negative rate and low true positive rate. In data mining and machine learning, it is difficult to establish an effective classifier for imbalanced data [29]. Therefore, many scholars have proposed various methods to compensate for it [30]. The common methods are as follows: algorithmic-level methods, data-level methods, cost-sensitive methods, and ensembles of classifiers [31].
The threshold method and one-class learning method are the most efficient algorithm-level solutions; the former sets different thresholds at different learning stages for different types of data, whereas the latter uses specific data to train the classifier. Data-level solutions are based on preprocessing the collected imbalanced training data set by either downsampling or oversampling strategies. Gustavo showed that resampling solutions can effectively solve the class imbalance problem and optimize classifier performance [32]. In particular, preprocessing the imbalanced data before constructing the classier is simple and efficient because the advantage of the data-level solution is to make the sampling and classifier training processes independent [33]. The data preprocessing method is based on the resampling of imbalanced data. Oversampling approaches are used to increase the number of data samples in the minority class and downsampling approaches are used to reduce the number of data samples in the majority class, respectively [34]. Common resampling methods include the synthetic minority oversampling technique (SMOTE) and random undersampling (RUS). RUS [21] performs similarly to SMOTE, but is based on a downsampling process where some examples are removed from the majority class. Lin et al. presented a clustering-based undersampling, which uses the K-nearest to cluster the minority class into the class subset which consists of the difference between the majority class and the minority class, resulting in a balanced training set [33].

3. Design of RoiSeg

In this section, we will describe the designing process of RoiSeg. As shown in Figure 1, RoiSeg consists of three modules, namely, ROI-central-point generation and feature extraction, automatic training-sample generation, and an online sample classifier. The purpose of RoiSeg is to classify the foreground through the ROI central points. In the first module, the frame difference and canny edge detection are used to transform the background modeling of each pixel into an ROI-central-point-based classification problem, which greatly reduces the amount of data operation. We then extract the features of the ROI central points and provide them to the automatic training-sample generation. In the second module, the characteristics extracted from the ROI central points are made into training samples using ROI-central-point-based sample clustering and a proposed trajectory-based class classifier. In the third module, we explore the training samples and find that they are class imbalanced. A K-means oversampling method is proposed to solve the class imbalance problem and a means to update training samples online is employed to compensate for the weaknesses of the Bayesian classifier.

3.1. ROI-Central-Point Generation

We use an imbalance degree η proposed in [35] to demonstrate the imbalance between the foreground and background, as shown in Table 1.
η = sum ( F ) sum ( B ) ,
where sum(F) and sum(B) are the sums of foreground and background pixels. We compute η on several subsets of the BMC database plus two self-captured sequences (“My_video1”, “My_video2”). The result reveals that the foreground and background are relatively imbalanced. ∞ means the number of foreground is far greater than the number of background.
As shown in Figure 2, the moving targets often include the foreground and the background [13]. The frame difference is intended to compute the difference between the current frame and the previous frame in the video sequence and then segment the moving targets. Suppose we have obtained the foreground frame shown in Figure 2c. There is a significant change in the pixel value between the current and previous frames at the position of the moving target. The moving target in the current frame is copied to the corresponding position in the previous frame, and a new previous frame is obtained. The moving target will not appear again if the new previous frame is subtracted from the current frame, as shown in Figure 3.
Following this principle, suppose we have classified the foreground and background. The background of the current frame is copied to the corresponding position in the previous frame, and a new previous frame is obtained. Foreground detection is then done if the new previous frame is subtracted from the current frame, as shown in Figure 3. When the moving target is detected by the frame difference (shown in Figure 3b), the Canny algorithm is used to detect the contours of the moving target and then the bounding boxes of the contours are obtained (as shown in Figure 3c). We call the region of the bounding boxes the Region of Interest (ROI). Therefore, the classification of foreground and background can be regarded as the classification of the bounding boxes. Furthermore, we use the center of the bounding box (the ROI central point) to represent the bounding box so that foreground detection is transformed into an ROI-central-point-based classification problem.

3.2. ROI-Based Noise Filter

Figure 3c shows that the areas of the bounding boxes for noise are much smaller than those of the foreground, because the bounding boxes for the vehicles and pedestrians that we pay attention to are often larger than those of other moving targets [36,37]. Based on this assumption, a bounding-box-area-based noise filter is proposed to remove the bounding boxes whose area is below a preset threshold. Here, the threshold is set to 0.1%. In this paper, we use 12 frame sequences as the experimental test set. Ten of them are from the BMC dataset [38], and two are hand-captured high-resolution crowd walking videos taken with a top-view camera, in which jitter was generated by shaking the camera. The description of the experimental test set is shown in Table 2. The filtering thresholds of the 12 videos are shown in Table 3. For the video test sets “112”, “122”, “212”, “222”, “312”, “322”, “412”, and “422”, we find that after filtering with the preset threshold, all the foregrounds are recognized, as shown in Figure 4.
There are two reasons for this phenomenon: first, the frame difference has good suppression on ambient lighting changes, so the moving cast shadows and fog cannot be detected. Second, the areas of the bounding boxes caused by salt and pepper noise, and so forth, are usually far smaller than those of the foreground, such as cars and pedestrians. However, for the video test sets “512”, “522”, “My_video1”, and “My_video2”, noise produced by the wind or the camera jitter and by other dynamic background factors dominates. The areas of the bounding boxes of such noise are random, usually varying with the strength of the wind and the magnitude of the jitter. When the area of a bounding box is larger than the threshold, noise is not be removed by area-based filtering. The filtering result is shown in Figure 4. It is important to achieve high classification accuracy in the Performance Comparison for the experiment (shown in Table 4) because a lot of noise is removed. For the test sets “112”, “122”, “212”, “222”, “312”, “322”, “412”, and “422”, we successfully obtain the foreground after filtering, so we do not need to use the second and third modules to classify the foreground and the background. This is why we do not use test sets “112”, “122”, “212”, “222”, “312”, “322”, “412”, and “422” in the experiment to classify the foreground- and background-based ROI central points and instead only use the test sets “512”, “522”, “My_video1”, and “My_video2”.

3.3. Automatic Training-Sample Generation

In Section 3.2, we get input samples, but these samples are unlabeled original samples and cannot be used as training samples for the classifier. In order to obtain labeled training samples, we propose an ROI-central-point-clustering method and a class detector.

3.3.1. ROI Pooling and Feature Extraction

For those videos not classified with the aforementioned noise filter, we are required to develop further feature extraction methods to enable noise reduction.
Let Z t = z i t be a frame at time t, z i t represent each pixel in Z t , and i generally refers to the i-th element in the set. We choose the r i , g i , b i color feature and the coordinate x i , y i of z i t as the spatio-color feature space z i t = r i t , g i t , b i t , x i t , y i t . For the ROI central point, a 5-tuple vector in the spatio-color feature space is selected as the z i t = r i t , g i t , b t t , x i t , y t t classification-learning feature. In Section 3.1, we used the center of the bounding box to represent the ROI central point. We also process the ROI central point with mean-pooling. Experiments show that using the pooled ROI central point z i t = r i t , g i t , b t t , x i t , y t t as a learning feature can provide good foreground detection results, as shown in Table 4 in the columns for the proposed method. The advantage of this method is that it reduces the computation load and guarantees good classification accuracy.

3.3.2. ROI Central Point Based Sample Clustering

We noticed that the background noise is often bound to a specific area. In adjacent frames, however, a moving target’s coordinates are also similar. Regardless of the foreground or background, the same type of target has more similar color features R, G, and B. In this section, we cluster ROI central points with similar characteristics. A GMM is a function to estimate the probability density of the exact polymorphism. It has excellent performance in clustering and its general form is as follows:
P x i = j = 1 k α j G x i ; μ j ; j .
The weighted coefficient α j is satisfied as follows:
j = 1 k α j = 1 , 0 α j 1 .
The j-th component ( j = 1 , , k ) is shown below:
G x i ; μ j ; j = j 1 2 ( 2 π ) d 2 e 1 2 x j μ j r j 1 x j μ j ,
where μ j and j represent the i t 1 mean vector and covariance matrices, respectively. We choose z i t = r i t , g i t , b t t , x i t , y t t as x i , which is the input of the GMM. All ROI central points in the first l frames (here we set l = 5) are used as input samples for the GMM. For example, we use the 3-Component GMM, so the ROI central points are clustered into 3 similar clusters, red, blue and yellow sets, as shown in Figure 5. However, the number of components determines the clustering accuracy of GMM. We choose precision ( P ), recall (R), and F-Measure (F) for performance evaluation:
precision = T P T P + F N ,   recall = T P T P + F P ,
F = 2 × precision × recall precision   + recall .
T P , F N and F P are the number of true positive, false negative, and false positive pixels, respectively. Figure 6 shows that F increases with the number of components increases. However, when this number reaches 6, the growth of F tends to be slow. Figure 7 shows that FPS drops with the increase of the number of components. The reasons are as follows: with the number of GMM components increasing, the data-fitting ability of the GMM is gradually enhanced, so the F value increases; meanwhile, the computation load is also increasing, which leads to the reduction of FPS; finally, the number of components increasing makes it easy for the GMM to over-fit the data. As such, we set the number of components to 6. Figure 6 demonstrated that F is between 0.87 and 0.91, which means that we can use the pooled ROI center point as the input of the GMM.

3.3.3. Trajectory Based Class Classifier

Figure 5 shows that in the 1st, 2nd, and 4th clips, the red set represents the foreground, and the blue and yellow sets represent the background. In the 3rd clip, pedestrians were incorrectly detected as another cluster, unlike in the 1st, 2nd, 4th clip. The reason for this is that it is difficult to classify the foreground and background using a GMM because it is a clustering method.
Therefore, we propose a trajectory-based-classifier method to foreground and background. Suppose G t s + 1 G t are the clustering results for the first s frames:
G t = g 1 t , g 2 t , , g m t , m = 1 , 2 , , k ,
where G t means the tth frame clustering result and m = k means there are k cluster in G t . g m t = x m 1 t , x m 2 t , x m n t , n = 1 , 2 , , num, where g m t represents the mth cluster, x n n t = r i t , g i t , b i t , x i t , y i t represents the pooled ROI central point, and num is the total number of ROI central points in g m t . Clusters in G t the current frame are matched one-by-one with clusters in of G t 1 the previous frame and if the matching is successful, they are considered the same type of target. Our proposed trajectory-based-classifier method is as follows:
(1) Calculate the mean-feature g ¯ m t of each cluster
g ¯ m t = 1 n u m n = 1 m o n x m n t = 1 n u m n = 1 m m r n t , n = 1 m m g n t , n = 1 m u n b n t , n = 1 m o n x n t n = 1 m o n y n t .
The mean-feature of all clusters in each frame is:
G ¯ t = g ¯ 1 t , g ¯ 2 t , , g ¯ m t , m = 1 , 2 , , k .
(2) Find the same class in adjacent frames
In adjacent frames, if a cluster g t 1 in the previous frame is the same type of target as a cluster g t in the current frame, they have similar g ¯ t . If they are not the same target type, they often have significant differences. The g ¯ t of each cluster in the previous and current frames is ordered accordingly, from small to large. The g ¯ t of the same sequence location is the same type of target because g ¯ t of the same type of target is very similar in adjacent frames. In this way, we can compare two ascending-order clusters in adjacent frames to get the same classes of foreground objects, thus reducing the cost of computation. Then we use this method to match the same type of target in the first s frames. For example, the first 4 frames matched the result shown in Figure 8. As a result, we changed the classification in Figure 5 and classified the pedestrians as red, the chairs in the middle as blue, and the chairs on the right as yellow, as shown in Figure 8.
(3) Label Positive and Negative Samples with Trajectories
We need to identify each class as foreground or background. When we obtain a class of targets in the first s frames, its displacement can be calculated. The proposed principle is as follows: The position coordinates of the mean-feature of the same cluster in the first s frames are denoted as [ ( x ¯ t s + 1 , y ¯ t s + 1 ) , ( x ¯ t s + 2 , y ¯ t s + 2 ) , , ( x ¯ t , y ¯ t ) ] . Then, we compute the moving distance of the clusters in t 1 , , t s + 2 , t s + 1 and t. We assume that the foreground displacements are increasing in a certain direction in adjacent frames, while the background often remains still or exhibits jitter with uncertain directions. If the moving distance of one cluster is increasing in the first s frames, it is classified as foreground; otherwise it is classified as background.
X = 0 , x ¯ t 1 x ¯ t 2 + y ¯ t 1 y ¯ t 2 , < , < x ¯ t s + 2 x ¯ t 2 + y ¯ t s + 2 y ¯ t 2 , < x ¯ t s + 1 x ¯ t 2 + y ¯ t s + 1 y ¯ t 2 1 , else
where X = 0 means foreground, marked as X F , and X = 1 means background, marked as X B . After the foreground and background classes are categorized, the original samples can be used as a labeled training sample, represented as X = X F , X B , v = 0 , 1 . In Figure 9, the red and yellow sets represent foreground and background, respectively. In this method, the correct selection of the value s has a great impact on the accuracy of the foreground and background detection. If s is too small, it is easy to misjudge a background element as foreground. If s is too large, the method may misjudge part of the foreground as background. Therefore, we experiment on video test sets to decide s. We define precision for foreground and background detection as c:
c = s u m ( T r u e ) s u m ( F r a m e ) ,
where s u m ( F r a m e ) is the sum of the test frame sequences, and s u m ( T r u e ) is the sum of the correctly judged test frame sequences. In Table 5, we find that when s = 6 , c has the highest value in most cases, and can almost reach 0.9.

3.4. Online Sample Classifier

3.4.1. Imbalance Compensation

The training samples are class imbalanced, as shown in Table 6. To better train the online classifier, we need to resample the training set to obtain balanced data. A K-means oversampling method is adopted to compensate for this imbalance. The methods are as follows:
(1) Calculate the total number of samples in the foreground, NF, and in the background, NB, respectively. Then, calculate the difference between them: K = N B N F .
(2) Use the K-means method to preprocess minority classes to get K clusters, calculate the mean of each cluster, then use the means as new minority-class samples. The new samples are added to each training sample so we get a set of balanced training samples.
The imbalance degrees are different in Table 1 and Table 6, because we use ROI-based area filtering, mentioned in Section 3.2.

3.4.2. Online Sample Updating

We use the new balanced training samples as training sets for the Naive Bayesian classifier. We use the ROI center points of the latest frame as the test set. After the Naive Bayesian classifies the test set, we get foreground and background represented as follows:
X = { r 1 , g 1 , b 1 , x 1 , y 1 , r N F , g N F , b N F , x N F , y N F , r 1 , g 1 , b 1 , x 1 , y 1 , r N B , g N B , b N B , x N B , y N B } .
Thus, we can get background bounding boxes and copy the pixels in the bounding boxes to the previous frame. Using the frame difference algorithm to process the new previous frame and the current frame, the foreground can then be segmented. To update the trained samples online, the newly classified foreground and background are added to the training samples. Then, we train the Naive Bayesian classifier with the updated training samples so that we can detect the foreground online.
The experimental results of the training classifier with imbalanced training sets are shown in Figure 10, and the precision results are shown in Figure 11a. The experimental results of training the classifier with the balanced and updated training sets are shown in Figure 12, and the precision is shown in Figure 11b. Red represents the background and yellow represents the foreground, respectively. Figure 11a,b demonstrated that P increased by 22.7 % , R increased by 2.1 % , and F increased by 23.4 % . This means using the proposed K-means oversampling method and online updating training samples to compensate for the weaknesses of the classifier is effective, especially for P and F.

4. Evaluation

In this section, we compared RoiSeg with the traditional foreground segmentation algorithms. Because the CNN foreground segmentation algorithms are not suitable for these scenarios due to their strict real-time requirements. Sobral tested and compared 29 background subtraction algorithms and recommended five of the best, namely DPWrenGABGS, MixtureOfGaussianV1BGS, MultiLayerBGS, PixelBasedAdaptiveSegmenter and LBAdaptiveSOM [38]. In this paper, we used these five algorithms to compare with our proposed method and the hardware of our experiment is a Lenovo desktop with Intel(R) Core(TM) i5-4590 CPU @ 3.3 GHz, 8 GB RAM, Win 10 64bit system. Because the foreground detected by frame difference has an aperture, we manually filled some aperture in the foreground in order to evaluate our algorithm using P, R and F. As the size of the self-captured sequences “My_video1” and “My_video2” were 1280 × 720 and that of the sequences provided by the BMC database was 640 × 480, the FPS we give was the average for all test sequences.
In Section 3.2, for the video sequences “112”, “122”, “212”, “222”, “312”, “322”, “412”, and “422”, we found that after filtering with the preset threshold, all foregrounds were recognized. Thus, we did not need a classifier to distinguish foreground and background. The results of running these five algorithms and the proposed RoiSeg on these eight sequences are shown in Table 4 and Figure 13. The average P, R and F of the five algorithms and proposed RoiSeg were computed, as shown in Table 4. We could see that the proposed RoiSeg had the best R and the best F on some sequences. For video sequences “112” and “122”, which were without noise, and “212” and “222”, which had salt and pepper noise, the proposed noise filter method did not function best; however, the R and F still reached over 0.9. For video sequences “312” and “322”, which having moving cast shadows, and “412” and “422”, which were foggy, the proposed noise filter method functions best. A non-optimal P indicated that our method produced more false positives than other algorithms. The reason for this was that the foreground detected by frame difference has ghosts, and the increased false positives were mainly located at the boundaries of moving objects, which was not harmful for visual observation. From Table 4, we found the proposed noise filter method had the highest FPS. This is due to the effectiveness of the proposed ROI-based noise filter, which mainly focuses on the ROI instead of the full frame.
For the video sequences “My_video1”, “My_video2”, “512”, and “522”, with wind and camera jitter, the noise could not be removed by filtering. Thus, we needed a classifier to distinguish foreground and background. The results of applying these five algorithms and the proposed RoiSeg algorithm are shown in Table 4 and Figure 14. From Figure 14, we could observe that the proposed RoiSeg algorithm produced better visual results than the five algorithms. Furthermore, Table 4 shows that the RoiSeg algorithm had the best F, which proved that our method had the best overall performance. After the proposed clustering and online classification work in Section 3.3 and Section 3.4, the FPS of our method decreases. However, Table 4 shows that the FPS of our method still outperformed those of the other approaches.
We also evaluated the performance of RoiSeg on different datasets with the metrics of the average pixel error rate (APFPER) and the joint intersection overlap (IoU) [39]. APFPER measured the number of misclassified pixels and IoU was to calculate the combined intersection of the estimated and true split plots for evaluating the split performance. We compared RoiSeg with state-of-the-art unsupervised learning methods on FBMS dataset, as shown in Figure 15 and Table 7. It is observed that the image saliency methods rendering of information within frames can produce unsatisfactory results, and even some images miss foreground objects, mainly because the time correlation in the image sequence to convey the target information was not taken into consideration [40]. However, these foreground segmentation methods based on motion perform better than the image saliency methods [41,42,43]. RoiSeg estimated the target object in a more cluttered background with higher real-time boundary and splits video objects in a completely unsupervised manner. We also conducted experiments to compare the performance of RoiSeg and other segmentation methods on a different dateset (SegTrack), as shown in Table 8 and Table 9.
In Table 8, References [41,42,43,45,47,48,49,50] are unsupervised learning methods, while [39,51] are the supervised learning methods. The results demonstrated that RoiSeg could meet the requirements of most tasks, although its performance was not as good as state-of-the-art segmentation methods. In Table 9, References [41,44,45,48,52] are unsupervised learning methods, while [46,50,51] are supervised learning methods. Among them, ref [46] utilized the CNN methods and VOC 2011 [53] for pre-training and testing, respectively. Table 9 shows that the result of IoU evaluation on RoiSeg was similar to that of APFPER and the CNN-based approaches had an absolute advantage in supervised segmentation tasks, but it relied on too much data. We also conducted extensive experiments to evaluate the real-time performance as shown in Table 10 and RoiSeg achieved a better performance in terms of real-time operations. In summary, RoiSeg outputted the expected results on some video sequences compared to the best performing unsupervised learning methods. There is a gap compared to the supervised method and the CNN method, and RoiSeg is better in real-time operations with the average processing time of 45 ms.

5. Conclusions and Future Work

In this paper, we propose RoiSeg, an effective object segmentation method, which consists of three modules, ROI-central-point generation and feature extraction, automatic training-sample generation, and an online sample classifier. RoiSeg can be applied to a number of scenarios where datasets are difficult to obtain and require high real-time performance. We also conduct extensive experiments and the results demonstrate that the frames per second of RoiSeg is 95.84, which is better than other algorithms, and the classification accuracy is 92.4%. Future work may fall into two categories. First, to find better algorithms to detect stopped objects, we plan to introduce Kalman filtering to predict the state of the stopping target the next time. Second, we will try to design a deep neural network algorithm to study the segmentation of the foreground in long-term scenarios.

Author Contributions

Conceptualization, Z.Z. and Z.P.; data collection, Z.T.; analysis and interpretation of results, Z.Z. and F.G.; validation, Z.Z. and F.G.; writing—original draft preparation, Z.Z. and F.G.; writing—review and editing, F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by China Postdoctoral Science Foundation (2020M671597), Jiangsu Postdoctoral Research Foundation (2020Z100), the National Science Foundation of the Jiangsu Higher Education Institutions of China (20KJB520002), Suzhou Planning Project of Science and Technology (No. SYG202024), and the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

BMC (Background Models Challenge) provides videos for testing our background subtraction algorithm. For more description of the BMC dateset and how to use the BMC dateset, please refer to this website: http://backgroundmodelschallenge.eu/#evaluation; (accessed on 2 January 2022). The Freiburg-Berkeley Motion Segmentation Dataset (FBMS) has a total of 720 frames annotated. FBMS-59 comes with a split into a training set and a test set. For more description of motion segmentation dataset and how to use evaluation code, please refer to this website: https://lmb.informatik.uni-freiburg.de/resources/datasets/moseg.en.html; (accessed on 2 January 2022). SegTrack is a video segmentation dataset. For more description of SegTrack dataset, please refer to this website: https://web.engr.oregonstate.edu/~lif/SegTrack2/dataset.html; (accessed on 2 January 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Agrawal, S.; Natu, P. Segmentation of Moving Objects using Numerous Background Subtraction Methods for Surveillance Applications. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2020, 9, 2553–2563. [Google Scholar]
  2. Qu, Z.; Chen, Z. An intelligent vehicle image segmentation and quality assessment model. Future Gener. Comput. Syst. (FGCS) 2021, 117, 426–432. [Google Scholar] [CrossRef]
  3. Li, L.; Wang, Z.; Hu, Q.; Dong, Y. Adaptive Nonconvex Sparsity Based Background Subtraction for Intelligent Video Surveillance. IEEE Trans. Ind. Inform. (TII) 2020, 17, 4168–4178. [Google Scholar] [CrossRef]
  4. Garcia-Garcia, B.; Bouwmans, T.; Silva, A.J.R. Background subtraction in real applications: Challenges, current models and future directions. Comput. Sci. Rev. 2020, 35, 100204. [Google Scholar] [CrossRef]
  5. Kalsotra, R.; Arora, S. Background subtraction for moving object detection: Explorations of recent developments and challenges. Vis. Comput. 2021. [Google Scholar] [CrossRef]
  6. Sultana, M.; Mahmood, A.; Bouwmans, T.; Jung, S.K. Dynamic background subtraction using least square adversarial learning. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 3204–3208. [Google Scholar]
  7. Liu, X.; Van De Weijer, J.; Bagdanov, A.D. Exploiting Unlabeled Data in CNNS by Self-supervised Learning to Rank. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2019, 41, 1862–1878. [Google Scholar] [CrossRef] [Green Version]
  8. Prasad, D.K.; Prasath, C.K.; Rajan, D.; Rachmawati, L.; Rajabally, E.; Quek, C. Object Detection in a Maritime Environment: Performance Evaluation of Background Subtraction Methods. IEEE Trans. Intell. Transp. Syst. (TITS) 2018, 20, 1787–1802. [Google Scholar] [CrossRef]
  9. Zhao, C.; Sain, A.; Qu, Y.; Ge, Y.; Hu, H. Background Subtraction Based on Integration of Alternative Cues in Freely Moving Camera. IEEE Trans. Circuits Syst. Video Technol. (TCSVT) 2018, 29, 1933–1945. [Google Scholar] [CrossRef]
  10. Chen, Y.Q.; Sun, Z.L.; Lam, K.M. An Effective Subsuperpixel-Based Approach for Background Subtraction. IEEE Trans. Ind. Electron. (TIE) 2019, 67, 601–609. [Google Scholar] [CrossRef]
  11. Huang, L.; Fu, Y.; Chen, R.; Yang, S.; Qiu, H.; Wu, X.; Zhao, S.; Gu, Y.; Li, P. SNR-adaptive OCT Angiography Enabled by Statistical Characterization of Intensity and Decorrelation with Multi-variate Time Series Model. IEEE Trans. Med. Imaging (TIP) 2019, 38, 2695–2704. [Google Scholar] [CrossRef] [Green Version]
  12. Xue, Z.; Yuan, X.; Yang, Y. Denoising-Based Turbo Message Passing for Compressed Video Background Subtraction. IEEE Trans. Image Process. (TIP) 2021, 30, 2682–2696. [Google Scholar] [CrossRef] [PubMed]
  13. Stauffer, C.; Grimson, W.E.L. Adaptive Background Mixture Models for Real-time Tracking. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Fort Collins, CO, USA, 23–25 June 1999; Volume 2, pp. 246–252. [Google Scholar]
  14. Culibrk, D.; Marques, O.; Socek, D.; Kalva, H.; Furht, B. Neural Network Approach to Background Modeling for Video Object Segmentation. IEEE Trans. Neural Netw. (TNN) 2007, 18, 1614–1627. [Google Scholar] [CrossRef] [PubMed]
  15. Yu, T.; Zhang, C.; Cohen, M.; Rui, Y.; Wu, Y. Monocular Video Foreground/Background Segmentation by Tracking Spatial-color Gaussian Mixture Models. In Proceedings of the IEEE Workshop on Motion and Video Computing (WMVC), Austin, TX, USA, 23–24 February 2007; pp. 5–13. [Google Scholar]
  16. Gallego, J.; Pardas, M.; Haro, G. Bayesian Foreground Segmentation and Tracking using Pixel-wise Background Model and Region Based Foreground Model. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 3205–3208. [Google Scholar]
  17. Cuevas, C.; Garcia, N. Efficient Moving Object Detection for Lightweight Applications on Smart Cameras. IEEE Trans. Circuits Syst. Video Technol. (TCSVT) 2013, 23, 1–14. [Google Scholar] [CrossRef]
  18. Wu, J.; Dong, M.; Ota, K.; Li, J.; Guan, Z. Big Data Analysis-based Secure Cluster Management for Optimized Control Plane in Software-defined Networks. IEEE Trans. Netw. Serv. Manag. (TNSM) 2018, 15, 27–38. [Google Scholar] [CrossRef] [Green Version]
  19. Afshang, M.; Dhillon, H.S. Poisson Cluster Process Based Analysis of HetNets with Correlated User and Base Station Locations. IEEE Trans. Wirel. Commun. (TWC) 2018, 17, 2417–2431. [Google Scholar] [CrossRef]
  20. Bu, Z.; Li, H.J.; Zhang, C.; Cao, J.; Li, A.; Shi, Y. Graph K-means Based on Leader Identification, Dynamic Game, and Opinion Dynamics. IEEE Trans. Knowl. Data Eng. (TKDE) 2019, 32, 1348–1361. [Google Scholar] [CrossRef]
  21. Seiffert, C.; Khoshgoftaar, T.M.; Van Hulse, J.; Napolitano, A. RUSBoost: A Hybrid Approach to Alleviating Class Imbalance. IEEE Trans. Syst. Man Cybern.-Part A Syst. Hum. 2009, 40, 185–197. [Google Scholar] [CrossRef]
  22. Nidheesh, N.; Nazeer, K.A.; Ameer, P. An Enhanced Deterministic K-Means Clustering Algorithm for Cancer Subtype Prediction from Gene Expression Data. Comput. Biol. Med. 2017, 91, 213–221. [Google Scholar] [CrossRef]
  23. Zhou, Y.; Rangarajan, A.; Gader, P.D. A Gaussian Mixture Model Representation of Endmember Variability in Hyperspectral Unmixing. IEEE Trans. Image Process. (TIP) 2018, 27, 2242–2256. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, C.; Yan, Z.; Pedrycz, W.; Zhou, M.; Li, Z. A Weighted Fidelity and Regularization-based Method for Mixed or Unknown Noise Removal from Images on Graphs. IEEE Trans. Image Process. (TIP) 2020, 29, 5229–5243. [Google Scholar] [CrossRef]
  25. Jiang, L.; Zhang, L.; Li, C.; Wu, J. A Correlation-based Feature Weighting Filter for Naive Bayes. IEEE Trans. Knowl. Data Eng. (TKDE) 2018, 31, 201–213. [Google Scholar] [CrossRef]
  26. Kim, S.B.; Han, K.S.; Rim, H.C.; Myaeng, S.H. Some Effective Techniques for Naive Bayes Text Classification. IEEE Trans. Knowl. Data Eng. (TKDE) 2006, 18, 1457–1466. [Google Scholar]
  27. Wu, D.; Jiang, Z.; Xie, X.; Wei, X.; Yu, W.; Li, R. LSTM Learning with Bayesian and Gaussian Processing for Anomaly Detection in Industrial IoT. IEEE Trans. Ind. Inform. (TII) 2019, 16, 5244–5253. [Google Scholar] [CrossRef] [Green Version]
  28. Yerima, S.Y.; Sezer, S. DroidFusion: A Novel Multilevel Classifier Fusion Approach for Android Malware Detection. IEEE Trans. Cybern. 2018, 49, 453–466. [Google Scholar] [CrossRef] [PubMed]
  29. Yu, H.; Yang, X.; Zheng, S.; Sun, C. Active Learning from Imbalanced Data: A Solution of Online Weighted Extreme Learning Machine. IEEE Trans. Neural Netw. Learn. Syst. (TNNLS) 2018, 30, 1088–1103. [Google Scholar] [CrossRef]
  30. Galar, M. A Review on Ensembles for the Class Imbalance Problem: Bagging-, Boosting-, and Hybrid-Based Approaches. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2012, 42, 463–484. [Google Scholar] [CrossRef]
  31. Beyan, C.; Fisher, R. Classifying Imbalanced Data Sets using Similarity Based Hierarchical Decomposition. Pattern Recognit. 2015, 48, 1653–1672. [Google Scholar] [CrossRef] [Green Version]
  32. Batista, G.E.; Prati, R.C.; Monard, M.C. A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data. ACM Sigkdd Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  33. Lin, W.C.; Tsai, C.F.; Hu, Y.H.; Jhang, J.S. Clustering-based Undersampling in Class-imbalanced Data. Inf. Sci. 2017, 409, 17–26. [Google Scholar] [CrossRef]
  34. Hu, S.; Liang, Y.; Ma, L.; He, Y. MSMOTE: Improving Classification Performance When Training Data is Imbalanced. In Proceedings of the International Workshop on Computer Science and Engineering, Qingdao, China, 28–30 October 2009; Volume 2, pp. 13–17. [Google Scholar]
  35. Zhang, X.; Zhu, C.; Wu, H.; Liu, Z.; Xu, Y. An Imbalance Compensation Framework for Background Subtraction. IEEE Trans. Multimed. (TMM) 2017, 19, 2425–2438. [Google Scholar] [CrossRef]
  36. Cui, Y.; Wang, Q.; Yuan, H.; Song, X.; Hu, X.; Zhao, L. Relative Localization in Wireless Sensor Networks for Measurement of Electric Fields under HVDC Transmission Lines. Sensors 2015, 15, 3540–3564. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. He, M.; Luo, H.; Chang, Z.; Hui, B. Pedestrian Detection with Semantic Regions of Interest. Sensors 2017, 17, 2699. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Sobral, A.; Vacavant, A. A Comprehensive Review of Background Subtraction Algorithms Evaluated with Synthetic and Real Videos. Comput. Vis. Image Underst. (CVIU) 2014, 122, 4–21. [Google Scholar] [CrossRef]
  39. Tsai, D.; Flagg, M.; Nakazawa, A.; Rehg, J.M. Motion coherent tracking using multi-label MRF optimization. Int. J. Comput. Vis. (IJCV) 2012, 100, 190–202. [Google Scholar] [CrossRef]
  40. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M.H. Saliency detection via graph-based manifold ranking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 3166–3173. [Google Scholar]
  41. Wang, W.; Shen, J.; Yang, R.; Porikli, F. Saliency-aware video object segmentation. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2017, 40, 20–33. [Google Scholar] [CrossRef] [PubMed]
  42. Fu, H.; Cao, X.; Tu, Z. Cluster-based co-saliency detection. IEEE Trans. Image Process. (TIP) 2013, 22, 3766–3778. [Google Scholar] [CrossRef] [Green Version]
  43. Zhou, F.; Bing Kang, S.; Cohen, M.F. Time-mapping using space-time saliency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 3358–3365. [Google Scholar]
  44. Zhang, D.; Javed, O.; Shah, M. Video object segmentation through spatially accurate and temporally dense extraction of primary object regions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 628–635. [Google Scholar]
  45. Papazoglou, A.; Ferrari, V. Fast object segmentation in unconstrained video. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, NSW, Australia, 1–8 December 2013; pp. 1777–1784. [Google Scholar]
  46. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  47. Brox, T.; Malik, J. Object segmentation by long term analysis of point trajectories. In Proceedings of the European Conference on Computer Vision (ECCV), Heraklion, Greece, 5–11 September 2010; pp. 282–295. [Google Scholar]
  48. Jang, W.D.; Lee, C.; Kim, C.S. Primary object segmentation in videos via alternate convex optimization of foreground and background distributions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 696–704. [Google Scholar]
  49. Li, F.; Kim, T.; Humayun, A.; Tsai, D.; Rehg, J.M. Video segmentation by tracking many figure-ground segments. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, NSW, Australia, 1–8 December 2013; pp. 2192–2199. [Google Scholar]
  50. Wen, L.; Du, D.; Lei, Z.; Li, S.Z.; Yang, M.H. Jots: Joint online tracking and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 2226–2234. [Google Scholar]
  51. Tsai, Y.H.; Yang, M.H.; Black, M.J. Video segmentation via object flow. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3899–3908. [Google Scholar]
  52. Xiao, F.; Jae Lee, Y. Track and segment: An iterative unsupervised approach for video object proposals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 933–942. [Google Scholar]
  53. Everingham, M.; Van Gool, L.; Williams, C.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes Challenge 2011 (VOC 2011) Results. Available online: http://www.pascal-network.org/challenges/VOC/voc2011/workshop/index.html (accessed on 2 January 2022).
Figure 1. The framework of RoiSeg.
Figure 1. The framework of RoiSeg.
Applsci 12 02674 g001
Figure 2. Frame difference. (a) Current frame. (b) Previous frame. (c) Binary frame.
Figure 2. Frame difference. (a) Current frame. (b) Previous frame. (c) Binary frame.
Applsci 12 02674 g002
Figure 3. Copy all ROI in current frame to previous frame. (a,d,g) are current frame; (b,e) are previous frame; (c,f,h,i) are binary frame. (ac) show the Process that copy all ROI in current frame to previous frame. (df) show the Process that selectively copy ROI of current frame to previous frame. (gi) The moving target detected by the frame difference.
Figure 3. Copy all ROI in current frame to previous frame. (a,d,g) are current frame; (b,e) are previous frame; (c,f,h,i) are binary frame. (ac) show the Process that copy all ROI in current frame to previous frame. (df) show the Process that selectively copy ROI of current frame to previous frame. (gi) The moving target detected by the frame difference.
Applsci 12 02674 g003
Figure 4. Experimental results with filtering. Left to right: “112”, “122”, “212”, “222”, “312”, “322”, “412”, “422”,“512”, “522”, “My_video1”, “My_video2”. Top to bottom: Original frame, Binary frame without filtering, Binary frame with filtering.
Figure 4. Experimental results with filtering. Left to right: “112”, “122”, “212”, “222”, “312”, “322”, “412”, “422”,“512”, “522”, “My_video1”, “My_video2”. Top to bottom: Original frame, Binary frame without filtering, Binary frame with filtering.
Applsci 12 02674 g004
Figure 5. The clustering result with the 3-Component GMM of the 7-th, 8-th, 9-th, 10-th frames in My_video1.
Figure 5. The clustering result with the 3-Component GMM of the 7-th, 8-th, 9-th, 10-th frames in My_video1.
Applsci 12 02674 g005
Figure 6. F, P, R of varying the number of GMM components.
Figure 6. F, P, R of varying the number of GMM components.
Applsci 12 02674 g006
Figure 7. FPS of varying the number of GMM components.
Figure 7. FPS of varying the number of GMM components.
Applsci 12 02674 g007
Figure 8. The first 4 frames matched this result.
Figure 8. The first 4 frames matched this result.
Applsci 12 02674 g008
Figure 9. Result of detecting foreground and background: the red sets represent foreground and the yellow sets represent background.
Figure 9. Result of detecting foreground and background: the red sets represent foreground and the yellow sets represent background.
Applsci 12 02674 g009
Figure 10. Experimental results from the Naive Bayes with imbalanced training sets. Top to bottom: results for “My_video1”, “My_video2”, “512”, “522”.
Figure 10. Experimental results from the Naive Bayes with imbalanced training sets. Top to bottom: results for “My_video1”, “My_video2”, “512”, “522”.
Applsci 12 02674 g010
Figure 11. (Left) Performance evaluation for classifier with imbalanced training sets. (Right) Performance evaluation for classifier with the balanced and updated training sets.
Figure 11. (Left) Performance evaluation for classifier with imbalanced training sets. (Right) Performance evaluation for classifier with the balanced and updated training sets.
Applsci 12 02674 g011
Figure 12. Experimental results from the Naive Bayes with the balanced and updated training sets. Top to bottom: results for “My_video1”, “My_video2”, “512”, “522”.
Figure 12. Experimental results from the Naive Bayes with the balanced and updated training sets. Top to bottom: results for “My_video1”, “My_video2”, “512”, “522”.
Applsci 12 02674 g012
Figure 13. Experimental results on eight sequences. Left to right: original images, DPWrenGABGS, MixtureOfGaussianV1BGS, MultiLayerBGS, PixelBasedAdaptiveSegmenter, LBAdaptiveSOM, proposed method. Top to bottom: “112”, “122”, “212”, “222”, “312”, “322”, “412”, “422”.
Figure 13. Experimental results on eight sequences. Left to right: original images, DPWrenGABGS, MixtureOfGaussianV1BGS, MultiLayerBGS, PixelBasedAdaptiveSegmenter, LBAdaptiveSOM, proposed method. Top to bottom: “112”, “122”, “212”, “222”, “312”, “322”, “412”, “422”.
Applsci 12 02674 g013
Figure 14. Experimental results on four sequences. Left to right: original images, DPWrenGABGS, MixtureOfGaussianV1BGS, MultiLayerBGS, PixelBasedAdaptiveSegmenter, LBAdaptiveSOM, proposed method. Top to bottom: ”512”, “522”, “My_video1”, “My_video2”.
Figure 14. Experimental results on four sequences. Left to right: original images, DPWrenGABGS, MixtureOfGaussianV1BGS, MultiLayerBGS, PixelBasedAdaptiveSegmenter, LBAdaptiveSOM, proposed method. Top to bottom: ”512”, “522”, “My_video1”, “My_video2”.
Applsci 12 02674 g014
Figure 15. Comparison between RoiSeg and state-of-the-art unsupervised learning methods on FBMS.
Figure 15. Comparison between RoiSeg and state-of-the-art unsupervised learning methods on FBMS.
Applsci 12 02674 g015
Table 1. Imbalance Degree of BMC Database And Self-captured Sequences.
Table 1. Imbalance Degree of BMC Database And Self-captured Sequences.
Dataset112122212222312322412422512522My_video1My_video2
η 0.04890.1440.3420.314
Table 2. The description of the experimental test set.
Table 2. The description of the experimental test set.
SequencesDescriptionSize
112Cloudy, without acquisition noise, as normal mode640 × 480
122640 × 480
212Cloudy, with salt and pepper noise during the whole sequence640 × 480
222640 × 480
312Sunny, with noise, generating moving cast shadows640 × 480
322640 × 480
412Foggy, with noise, making both background and foreground hard to analyze640 × 480
422640 × 480
512Wind, with noise, producing a moving background640 × 480
522640 × 480
My_video1Camera jitter1280 × 720
My_video21280 × 720
Table 3. Threshold of the bounding box areas to classify foreground.
Table 3. Threshold of the bounding box areas to classify foreground.
Video Sequences112122212222312322412422512522My_video1My_video2
Number of Video clips1502150314991499149915011499149914991499390390
number of pixels in a bounding box30421830421830421830421830421810001000
Total area of frame covered by bounding boxes for noise (%)0.10.070.10.070.10.070.10.070.10.070.10.1
Table 4. Performance evaluation of the five algorithm and proposed RoiSeg.
Table 4. Performance evaluation of the five algorithm and proposed RoiSeg.
BMC
Sequences
DPWren GABGSMixture Of Gaussian V1BGSMultiLayer BGSPixel Based Adaptive SegmenterLBAdaptive SOMProposed RoiSeg
PRFFPSPRFFPSPRFFPSPRFFPSPRFFPSPRFFPS
1120.870.870.8770.20.960.740.8489.30.920.950.9350.880.90.8915.60.860.920.8920.60.890.930.91115
1220.910.870.8977.60.960.70.870.50.910.940.932.20.90.880.8913.20.880.930.922.30.910.940.92120.6
2120.920.860.8958.30.970.740.8470.30.940.940.942.50.890.890.898.20.790.770.7815.50.890.930.9170.6
2220.930.860.959.20.960.70.8170.60.940.930.933.50.90.870.897.60.890.920.9114.20.910.940.9285.1
3120.650.780.7170.40.980.680.873.80.960.870.912.40.880.870.8711.20.520.840.6419.20.890.930.91103.2
3220.890.780.8363.20.950.650.7765.90.940.850.894.30.90.80.8512.30.540.850.6615.10.910.940.9288.3
4120.530.760.6262.10.980.690.8187.70.710.840.773.10.850.820.8411.50.510.780.6113.30.890.930.9198.1
4220.530.750.6269.30.970.640.7775.80.770.790.783.90.850.770.8110.40.510.780.6215.10.910.940.9285.8
5120.630.860.7373.40.820.740.7876.80.650.930.764.10.820.890.8614.10.520.880.6618.30.810.910.86102.3
5220.80.860.8370.30.910.690.7972.20.880.930.93.30.890.870.8812.40.670.920.7821.60.890.930.9199.1
My_video10.380.840.5412.10.680.520.5916.30.760.890.830.50.820.890.855.10.420.80.556.50.90.910.8642.23
My_video10.30.830.4411.30.80.480.613.50.750.840.790.450.750.850.84.50.230.870.363.20.860.860.8639.62
Table 5. The precision for classify the foreground and background.
Table 5. The precision for classify the foreground and background.
DatasetVideo Clips2345678910
51214990.5280.6830.8560.9140.9240.9190.9030.8390.789
52214990.5790.6540.8870.9120.9080.8780.8540.8030.776
My_video13900.4710.5890.7180.8420.9030.8790.8030.7530.684
My_video23900.5210.5710.7330.8830.8990.8570.7940.7090.649
Table 6. Imbalance degrees of training sets.
Table 6. Imbalance degrees of training sets.
Imbalance Degree512522Video1Video2
η 0.1840.5380.4210.523
Table 7. Comparison based on IoU between RoiSeg and other methods on FBMS dataset.
Table 7. Comparison based on IoU between RoiSeg and other methods on FBMS dataset.
VideoRoiSeg[42][44][45][41][46]
Bear263.5187.5221.1486.8170.1188.92
Cars515.6210.7138.7317.3838.5260.11
Cars930.1719.5528.9252.4460.0877.82
Cats178.8319.7581.4983.1185.7270.13
People158.6356.0664.8253.3368.1277.07
People555.8210.7184.4351.8156.4173.31
Rabbits256.0120.4147.8128.3271.0679.12
Avg.51.2332.1052.4853.3164.2975.21
Table 8. Comparison based on APFPER between RoiSeg and other methods on SegTrack dataset.
Table 8. Comparison based on APFPER between RoiSeg and other methods on SegTrack dataset.
VideoFramesUnsupervisedSupervised
RoiSeg[47][44][45][48][49][50][41][39][51]
Birdfall30352217155189144199468140252454
Cheetah29776890633806617599196862211421217
Girl21125338591488169811951164759599113041755
Monkeydog715572843654723543221434350563683
Parachute514128552202212002421113195235502
Avg. 67012215726775025052516459699922
Table 9. Comparison based on IoU between RoiSeg and other methods on SegTrack dataset.
Table 9. Comparison based on IoU between RoiSeg and other methods on SegTrack dataset.
VideoFramesUnsupervisedSupervised
RoiSeg[44][45][52][48][41][50][51][46]
Birdfall3060.9171.4337.3972.5273.2174.5178.7157.4178.83
Cheetah2950.1258.7540.9161.2164.2264.3466.1233.8275.31
Girl2170.9481.9171.2186.3786.6788.7284.6487.8588.84
Monkeydog7165.2174.2473.5874.0776.1278.0482.1554.3585.65
Parachute5190.1293.9388.0895.9294.6294.894.4294.5295.61
Avg. 67.4676.0562.2378.0278.9780.0881.2165.5984.85
Table 10. Realtime comparison between RoiSeg and other methods.
Table 10. Realtime comparison between RoiSeg and other methods.
MethodRoiSeg[44][45][50][51][41][46]
Value0.0535.120.531.820.843.260.38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Pei, Z.; Tang, Z.; Gu, F. RoiSeg: An Effective Moving Object Segmentation Approach Based on Region-of-Interest with Unsupervised Learning. Appl. Sci. 2022, 12, 2674. https://doi.org/10.3390/app12052674

AMA Style

Zhang Z, Pei Z, Tang Z, Gu F. RoiSeg: An Effective Moving Object Segmentation Approach Based on Region-of-Interest with Unsupervised Learning. Applied Sciences. 2022; 12(5):2674. https://doi.org/10.3390/app12052674

Chicago/Turabian Style

Zhang, Zeyang, Zhongcai Pei, Zhiyong Tang, and Fei Gu. 2022. "RoiSeg: An Effective Moving Object Segmentation Approach Based on Region-of-Interest with Unsupervised Learning" Applied Sciences 12, no. 5: 2674. https://doi.org/10.3390/app12052674

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop