An Image Segmentation Method Using an Active Contour Model Based on Improved SPF and LIF

: Inhomogeneous images cannot be segmented quickly or accurately using local or global image information. To solve this problem, an image segmentation method using a novel active contour model that is based on an improved signed pressure force (SPF) function and a local image ﬁtting (LIF) model is proposed in this paper, which is based on local and global image information. First, a weight function of the global grayscale means of the inside and outside of a contour curve is presented by combining the internal gray mean value with the external gray mean value, based on which a new SPF function is deﬁned. The SPF function can segment blurred images and weak gradient images. Then, the LIF model is introduced by using local image information to segment intensity-inhomogeneous images. Subsequently, a weight function is established based on the local and global image information, and then the weight function is used to adjust the weights between the local information term and the global information term. Thus, a novel active contour model is presented, and an improved SPF-and LIF-based image segmentation (SPFLIF-IS) algorithm is developed based on that model. Experimental results show that the proposed method not only exhibits high robustness to the initial contour and noise but also effectively segments multiobjective images and images with intensity inhomogeneity and can analyze real images well.


Introduction
Image segmentation is an important task in the field of image analysis and object detection and aims to segment an image into distinctive subregions that are meaningful to analyze [1].Segmentation is the intermediate step between image processing and image analysis as well as the bridge from lowto high-level research in computer vision.Inhomogeneity, noise, and low contrast in real images have increased the difficulty of image segmentation [2].
Over the past few decades, many segmentation methods have been proposed.The active contour model (ACM), which was proposed by Kass et al. [3], has been proven to be an efficient framework for image segmentation.The fundamental idea of the ACM framework is to control a curve to move toward its interior normal and then stop on the true boundary of an object based on an energy minimization model [4].The two main shortcomings of ACM algorithms are (1) sensitivity to the initial position and (2) difficulties related to topological changes [5].Generally, existing ACM methods can be roughly divided into the following types, edge-based models [6][7][8][9] and region-based models [10][11][12][13][14].
The geodesic active contour (GAC) model [15] is the most typical of edge-based methods.Owing to the edge-indicator function, the model can stop at high-contrast image gradients [16].Edge-based models have distinct disadvantages.For example, these methods can effectively segment to segment intensity-inhomogeneous images.Moreover, a weight function is established to adjust the weights between the SPF model and the LIF model.Thus, a novel ACM model is presented, and an image segmentation algorithm is investigated.Experimental results demonstrate that our model involves simpler computation, exhibits faster convergence, and can effectively segment multiobjective images and intensity-inhomogeneous images.Furthermore, the proposed method is highly robust to the initial contour and noise.
The remainder of this paper is structured as follows.Section 2 briefly reviews the GAC, C-V, SBGFRLS, and LIF models.In Section 3, by combining the improved SPF function with the LIF model, a novel ACM is presented, and using this model, an image segmentation algorithm is designed.Then, the experimental results and analysis are discussed in Section 4. Section 5 presents the conclusions.

The GAC Model
The GAC model uses image gradient information from the boundary of an object [33].Suppose that I: Ω⊂R 2 is an image domain, I: Ω → R 2 is an input image, and C(q) is a closed curve.Then, the GAC model is formalized by minimizing the following energy functional as where g is a strictly decreasing function.
Usually, a satisfactory edge stopping function (ESF) should be defined, which is regular and positive at object boundaries [21], e.g., where G σ denotes the Gaussian kernel function and G σ *I describes the convolution operation of I with G σ .Using the steepest descent method and the calculus of variations, we obtain the Euler-Lagrange form of Equation (1), which is written as where k is the curvature of the contour and → N is the inward normal to the curve.A constant velocity term α is typically added to increase the propagation speed [21].Thus, Equation (3) can be rewritten as The corresponding level set formulation is described as where φ represents the level set function and α is the balloon force that controls the shrinkage or expansion of the contour.
The GAC model utilizes the image gradient to construct an ESF, which can stop the contour evolution on object boundaries.When images have weak boundaries or the initial contour is far from the desired object boundary, the GAC model will fail to find the target [18,22].

The C-V Model
The C-V model is proposed based on the assumption that the original image intensity is homogeneous.The energy functional of the C-V model [34] is expressed as where λ 1 and λ 2 are positive constants that regulate image driving force inside and outside the contour, c 1 represents the mean gray value of the target area and the background area in the evolution curve C, and c 2 represents the mean gray value of the target area and the background area outside the evolution curve C. By minimizing Equation ( 6), one has c 1 and c 2 , which are described, respectively, as where H(φ) is the Heaviside function.
In practice, the Heaviside function H(φ) and the Dirac delta function δ(φ) must be approximated by smooth functions H ε (φ) and δ ε (φ) when ε→0, which are typically expressed as follows, respectively By incorporating the length and area energy terms into Equation ( 6) and further minimizing the length and area of the level set curve, the corresponding partial differential equation is described as where µ, ν, λ 1 , and λ 2 denote the corresponding coefficients, all of which are positive constants; ∇ is the gradient operator; µ controls the smoothness of the zero level set; ν increases the propagation speed; and λ 1 and λ 2 control the image data driving force inside and outside the contour, respectively.Because c 1 and c 2 are related to the global information inside and outside the curve, this model can segment blurred images and images with weak gradients more effectively than the edge-based model can, and it is insensitive to the initialization location [22,35].However, when the internal and external intensities of the curve are inhomogeneous, c 1 and c 2 cannot express the local information precisely, which leads to the failure of image segmentation [2].

The SBGFRLS Model
The SBGFRLS model is proposed based on the traditional C-V model and the GAC model, thereby seizing the advantages of both models [21].In the SBGFRLS model, an SPF function is used to substitute ESF in the GAC model, and thus the level set formulation of the SBGFRLS can be expressed as where spf (I(x)) in Equation ( 12) is an SPF function, which can be given as where c 1 and c 2 represent the gray mean values of regions outside and inside the contour, computed using Equations ( 7) and (8), respectively.The SBGFRLS model can reduce the cost of the expensive reinitialization of the traditional level set method and is more efficient than traditional models.The model stops the contour evolution, even with blurred edges, without any a priori training.However, the model assumes that the region to be segmented is homogeneous.This assumption occasionally holds in general clinical cases.When facing heterogeneous intensity distributions, the detection accuracy can fall significantly because the fundamental assumption is violated [36,37].Moreover, the SBGFRLS model can become trapped in a local minimum without proper initialization, which leads to poor segmentation performance [38][39][40].

The LIF Model
The local fitted image (LFI) formulation [31] is defined based on local image information, based on which the LIF model is investigated.This model can segment intensity-inhomogeneous images [41].The LIF model is expressed as follows where I LFI is a local fitted image, and any x ∈ Ω.
It follows that I LFI can be calculated as where m 1 and m 2 are expressed, respectively, as φ is the zero level set of a Lipschitz function that represents the contour C; H ε (φ) is the regularized Heaviside function, as defined in Equation ( 9); and W k (x) is a rectangular window function.
In our experiment, W k (x) is a truncated Gaussian window with a standard deviation of σ and size (4k + 1) × (4k + 1), where k is the greatest integer that is smaller than σ.Similarly, the segmentation results can be achieved if a constant window is chosen [31].
According to the calculus of variations and the gradient descent method, the following partial differential equation can be obtained by minimizing E LIF : where δ ε (φ) is the regularized Dirac delta function [32], which is calculated as indicated in Equation (10).
According to the complexity analysis and experimental results in [31,32,41,42], the LIF model is more efficient than the LBF model.However, neither model can handle noisy and intensity-inhomogeneous images well [41,42].

Improved SPF Function
The main strategy of the ACM based on region information is to construct a driving force, which is based on the information of the image region [43].The region function modulates the sign of the pressure forces using region information such that the contour shrinks when it is outside the object of interest and expands when it is inside the object.For this reason, these external forces are sometimes called SPF [43].Zhang et al. [22] proposed the SBGFRLS model, which utilizes the statistical information inside and outside the contour to construct a region-based SPF function [37].However, an SPF function is simply based on image information.Thus, the corresponding model cannot segment intensity-inhomogeneous images or images with weak boundaries [36,41].
In this study, the global information of image I is used to divide the image into two parts, inC and outC, and the level set function is then introduced into the new SPF function.
Using global region information and combining c 1 and c 2 , a global fitted image formulation is defined as where H ε (φ) defined in Equation ( 9) is the regularized Heaviside function and c 1 and c 2 are calculated by Equations ( 7) and ( 8), respectively, and .* describes matrix multiplication.By employing the above-defined global fitted image, a new SPF function is defined as According to the construction approach of the SPF function, a new partial differential equation is defined as where α is the balloon force that controls the shrinkage or the expansion of the contour.In this paper, according to the concept of a balloon force established previously [44], a balloon force is reconstructed to change the evolution rate of the level set function adaptively, which is defined as The new SPF is more efficient than the traditional ACM models because this function avoids the expensive cost of the reinitialization step.Moreover, the SPF is less sensitive to the initialization location.However, the SPF function is constructed with only global image information.Therefore, it appears difficult to handle images with intensity inhomogeneity using this approach.

Active Contour Model Based on Improved SPF and LIF
Zhang et al. [31] constructed the LIF model, which can effectively process nonhomogeneous images through local image information.Unfortunately, the model is sensitive to the initial curve and noise [2].To construct a model that can process nonhomogeneous images and reduce the dependence on the location of the initial contour, this subsection combines the new SPF function with the existing LIF model to form a new ACM based on local and global image information.
By combining the new SPF function with the LIF model, the new level set evolution equation is defined as where δ ε (φ), defined in Equation (10), is the regularized Dirac delta function and λ is a new weight coefficient.
Here, λ is a weight function that can be employed to dynamically adjust the ratio between the local and the global term in image segmentation.Namely, the image information term playing a crucial role in segmenting an image can be selected.
Based on the local and global image information, the weight coefficient λ is defined as where A is defined in Equation ( 21) and B is defined as where m 1 and m 2 are defined in Equation ( 16).
It is noted that the selection of the weight parameter λ is important in controlling the influence of the local and the global terms.Li et al. [45] declared that the local term is critical to the initialization to some extent; a global term is incorporated into the local framework, thereby forming a hybrid ACM.Therefore, with the mutual assistance of the local force and the global force, the robustness to the initialization can be improved, and the global force is dominant if the evolution curve is away from the object.When the contour is placed near the object boundaries, the LIF model plays a dominant role, and fine details can be detected accurately.In contrast, the new SPF model plays a key role when the contour is located far from the object boundaries, and owing to the assistance of the SPF, a flexible initialization is allowed.It follows that the automatic adjustment between the LIF and SPF models in our ACM is very distinct.Furthermore, the objective of the dynamic adjustment is to determine an optimal result for image segmentation.
In general, the new proposed SPFLIF-IS model not only solves the problem that the intensity-inhomogeneous images cannot be accurately segmented by using the global image information but also overcomes the primary shortcoming that the model based on the local image information is sensitive to noise and the initial contour.

Algorithm Steps
The procedures of image segmentation are illustrated in Figure 1.After the abovementioned image segmentation algorithm has been applied, an improved SPF and LIF-based image segmentation (SPFLIF-IS) algorithm using ACM can be implemented and described as Algorithm 1, which is summarized as follows.Step 1: Initialize the level set function φ, and set the coefficients ∆t, n, and ε.
Step 2: Calculate the Heaviside function and the Dirac delta function using Equations ( 9) and ( 10), respectively.Step 3: For n = 1: iterNum // iterNum is the total number of iterations.
Step 6: Introduce the LIF model.
Step 8: Calculate the level set evolution equation using Equation (22).
Step 9: If the evolution of the curve is stable, then output the segmentation result.Else, return to Step 4.
Step 10: End for It is well known that convolution operations are the most time-consuming with respect to the time complexity of an algorithm.Therefore, it is necessary to explain the complexity of the convolution operation.When an algorithm requires a convolution operation, the time cost is approximately O(n 2 × N) [46], where N is the image size and n is the Gaussian kernel.The values of N are greater than n 2 .
Because the C-V model [34] must be reinitialized in every iteration, its time cost is very high, and the computational complexity is O(N 2 ) [31].The LBF model [24] usually needs to perform four convolution operations in each iteration, which greatly increases the computational time complexity.This situation indicates that the time complexity is O(itr × 4 × n 2 × N), where the parameter itr is the number of iterations.In contrast, the SBGFRLS model [23] must perform three convolution operations, two of which are derived by gradient calculation (horizontal and vertical), and the other involves mask image and filter mask.Thus, the total computational complexity of the SBGFRLS model is O(itr × 3 × n 2 × N).The LIF model [31] performs two convolution operations in each iteration.Thus, the total computational time required for the LIF model is O(itr × 2 × n 2 × N).For the SPFLIF-IS algorithm, the computational complexity is mainly concentrated in Step 6.In Algorithm 1, Step 6 is the most time-consuming to calculate in the LIF model.The computational complexity of our proposed method is O(itr × 2 × n 2 × N), where n is the size of the Gaussian kernel function and N is the image size.Since in most cases, N >> n 2 , the complexity of SPFLIF-IS is O(N) approximately, which is close to that of the LIF model in [31].It follows that our proposed method is much more computationally efficient than the C-V model [31], the LBF model [24], and the SBGFRLS model [23].Because the SPFLIF-IS algorithm decreases the number of Gaussian convolution operations required, its time costs and number of iteration operations are drastically reduced.Therefore, the computational complexity of our SPFLIF-IS method is lower than that of the other related ACMs [6,8,11,12,15,17,20,23,24,31,34,43].

Experiment Preparation
In this section, comprehensive segmentation results for all algorithms compared are presented to validate the performance of our proposed method on various representative synthetic and real images with respect to different characteristics.Following the experimental techniques for image segmentation designed by Ji et al. [42], these selected images are mostly corrupted with one or more degenerative characteristics, including additive noise, low contrast, weak edges, and intensity inhomogeneity.Unless otherwise specified, the same parameters are employed as follows, ∆t = 1, n = 5, ε = 1.5, and ϕ 0 (x, y) = 1 : (x, y) ∈ in(c) or φ 0 (x, y) = −1 : (x, y) ∈ out(c).The Gaussian kernel plays an important role in practical applications; the kernel is a scale parameter controlling region scalability from small neighborhoods to the entire image domain [31].In general, the value of the scale parameter should be appropriately selected from practical images.It is well known that an excessively small value may cause undesirable results, whereas an excessively large value can lead to high computational complexity [31,36].Thus, the Gaussian kernel size controlling the regularization of the level set function should be chosen according to practical cases [36].Following the experimental techniques designed in [31,36], the σ selected in our experiments is typically less than 10.All of models compared in this paper are tested in MATLAB R2014a in a Windows 7 environment using a 3.20 GHz Intel (R) Core i5-3470M processor with 4 GB RAM.

Segmentation Results of Images with Intensity Inhomogeneity
To demonstrate the satisfactory performance and effectiveness of the SPFLIF-IS model, a series of experimental results are presented.We compare our model with the following five existing models: (1) the C-V (The code is available at [47]) model [34], (2) the LBF (The code is available at [48]) model [24], (3) the LIF (The code is available at [47]) model [31], (4) the SBGFRLS (The code is available at [47]) model [23], and (5) the LSACM model [28].The five representative ACM algorithms are the state-of-the-art level set methods published recently for image segmentation.The algorithms show improvements over the classical ACM and are specially selected based on the level set method for comparison experiments.The chosen parameters for these models can be found in [23,24,28,31,34].The segmentation results obtained for images with intensity inhomogeneity using the six models are illustrated in Figure 2, where the original images shown in Figure 2a can be found in [2].Note that because the visual evaluations in Figure 2 are partial to subjective measures, to strengthen the objective results of our experiments, the corresponding tables should be added to defend the arguments for all the tested images in the following visual evaluations, in which each failure is clearly labeled to avoid ambiguity.To more clearly illustrate this state, the following symbols are adopted in the tables: F1: fail to detect boundaries, F2: nonideal boundaries detected, F3: fail to detect internal boundaries, and T: true boundaries detected.Table 1 objectively describes the segmentation results of Figure 2 in detail.It can be clearly concluded from Table 1 that the LBF performs as well as the SPLIF-IS, the C-V exhibits slightly bad results, and the LSACM produces the worst results.Therefore, the experimental results shown in Figure 2 2.

Segmentation Results of Multiobjective Images
This portion of our experiment concerns the segmentation results obtained for multiobjective images.The SPFLIF-IS method is consistently compared with the five abovementioned methods (C-V, LBF, SBGFRLS, LIF, and LSACM).The original multiobjective images and the segmentation results of the six models are shown in Figure 3, where the original images shown in Figure 3a are derived from [42,49].Although our model identifies most of the boundaries of the first image, the boundaries are subtle different when compared with those detected by the LBF model.As shown in Figure 3d,f, the SBGFRLS model and the LSACM model obviously fail to segment the first, second, and fourth multiobjective images.The true boundaries of the third image cannot be extracted by the C-V model, the LBF model, the SBGFRLS model, or the LIF model; the results are shown in Row 3 of Figure 3. Table 2 describes the segmentation results shown in Figure 3.As shown in Table 2, the SPFLIF-IS yields the best results, the C-V performs as well as the LBF, and the LIF exhibits the worst results.Figure 3 and Table 2 clearly show that our proposed SPFLIF-IS method can segment the fourth image, but the other comparison methods cannot.The experimental results indicate that the SPFLIF-IS model can efficiently segment the multiobjective images.
Table 2. Description of the segmentation results of Figure 3.

Segmentation Results of Noisy Images
The following subsection describes the experimental segmentation results obtained for noisy images.The SPFLIF-IS model is still compared with the C-V, LBF, SBGFRLS, LIF, and LSACM models.Figure 4 illustrates the original images with different noise intensities and compares the results of the six state-of-the-art segmentation methods, where the original images without noise in Figure 4a are derived from [46].In Figure 4, Row 1 shows the original images and the segmentation results.Row 2 to Row 5 show the added Gaussian noise with zero means and different variances (σ = 0.01, 0.02, 0.03, 0.05).Figure 4c,f shows that the LBF model and the LSACM model cannot analyze the five images.Although the C-V model and the SBGFRLS model can segment the first and the second image, neither model performs well when the noise intensity increases; the results are shown in Figure 4b,d.Figure 4e shows that the LIF model could analyze the images without Gaussian noise well.With respect to the segmentation of the images containing Gaussian noise, the LIF model exhibits poor performance.As shown in Figure 4g, the object boundaries are accurately extracted by our proposed SPFLIF-IS model.Table 3 describes the segmentation results of Figure 4. Table 3 shows that the SPFLIF-IS model yields the best results, the C-V model performs as well as the SBGFRLS model, and the LBF model performs as poorly as the LSACM model.The experimental results demonstrate that the SPFLIF-IS model can effectively eliminate the interference of the noise and complete the segmentation of the noisy images.

Segmentation Results of Texture Image
This part of our experiment tests the segmentation performance of texture images.Figure 5a shows the original texture image, which is derived from [4].Moreover, the compared models are still the C-V, LBF, SBGFRLS, LIF, and LSACM models.According to Figure 5c,e,f, the object boundaries of the first image are not identified by the LBF, LIF, or LSACM model, respectively.Most of the boundaries are obtained by the SBGFRLS model.However, the internal details are not recognized; the detailed results are illustrated in Figure 5d. Figure 5e shows that the LIF model fails to segment the second image.Although the C-V, LBF, SBGFRLS, and LSACM models recognize the true boundaries of the second image, some boundaries lie in the middle of the image; the results are illustrated in Row 2 of Figure 5. Table 4 describes the segmentation results of Figure 5. Table 4 shows that the SPFLIF-IS model performs the best, the C-V model exhibits the second best performance, and the LIF model shows as poor a performance as the LSACM model.The SPFLIF-IS model can eliminate the interference of the image texture and analyze the texture image well.

Segmentation Results of Real Images
In this subsection, we continue testing our algorithms, this time using real images.The SPFLIF-IS method is compared with the same five methods (C-V, LBF, SBGFRLS, LIF, and LSACM).The original real images and the segmentation results of the six models are shown in Figure 6, where the first and second images in Figure 6a can be found in the literature [27,31], and the third and fourth images shown in Figure 6a are selected from Berkeley segmentation data set 500 (BSDS500) (The code is available at [50]).The first image in the third and fourth columns shows that the LBF and SBGFRLS models fail to segment the image; the results are shown in Figure 6c,d.The first image in the fifth and sixth columns shows that most of the boundaries are obtained by the LIF and LSACM models.However, the internal details are not recognized; the results are illustrated in Figure 6e,f.The LBF, LIF, and LSACM models fail to segment the second and third images, as shown in Figure 6c,e,f.The object boundaries of the fourth images are not accurately extracted by the C-V, LBF, SBGFRLS, LIF, and LSACM models, as shown in the fourth rows of Figure 6.As shown in Figure 6g, the object boundaries are accurately extracted by our proposed model.Table 5 objectively offers the segmentation results of Figure 6.Table 5 indicates that the SPFLIF-IS achieves the best results; the C-V exhibits slightly better results than those obtained by the LBF, SBGFRLS, LIF, and LSACM; and the LBF performs as poorly as the LSACM.The segmentation results demonstrate that our SPFLIF-IS model can efficiently analyze real images and yield great segmentation results.
Table 5. Description of the segmentation results of Figure 6.

Comparative Evaluation Results
In addition to using visual evaluation, the accuracy of the target region segmentation can be assessed quantitatively and objectively using the DICE coefficient (DICE) [51,52] and the Jaccard similarity index (JSI) [53].Following the experimental techniques designed in [42,54], test images are selected randomly from the BSDS500 database.Note that BSDS500 contains hundreds of natural images whose ground-truth segmentation maps have been generated by multiple individuals [40,55].To enhance the coherency of our work with the abovementioned algorithms, three comparative experiments are performed on many real-world color images, which are selected from the Berkeley segmentation data set 500 (BSDS500) and consist of a set of natural images.
The first part of this experiment involves evaluating the value of the DICE for twenty representative real-world color images, which are chosen from the Berkeley segmentation data set 500 (BSDS500).The algorithms compared are the C-V model [34], the LBF model [24], and the LIF model [31].
The DICE, also called the overlap index, is the most frequently used metric for validating image segmentations.The DICE measures how well the segmentation results S match the ground truth G.When the value of the DICE is close to 1, the segmentation results have high accuracy.The formula for the DICE is given as where Ω S describes the segmented volume and Ω G denotes the ground truth [56,57].The DICE values of the segmentation results obtained by applying the four models to segment Berkeley color images are listed in Table 6, where the Mean describes the average values of the DICE for all test image data.Table 6 shows that the SPFLIF-IS method yields the best values for the DICE on the twenty image data, and the corresponding Mean is also the largest.The results indicate that our SPFLIF-IS model outperforms the C-V, LBF, and LIF models.In summary, these results demonstrate that our SPFLIF-IS method is indeed efficient and outperforms these currently available approaches.The next section of the experiment involves testing the value of the JSI coefficient for the twenty representative real-world color images in Table 6.The algorithms compared are still the C-V model [34], the LBF model [24], and the LIF model [31].
The JSI is the second statistical measure used for quantitative evaluation in this paper.The JSI is calculated by The accuracy of the segmentation results for the Berkeley color images is measured by the JSI value, as shown in Figure 7.A JSI value close to 1 indicates favorable segmentation results.Figure 7 shows that the SPFLIF-IS method exhibits the best JSI values for the twenty image data.For image IDs 3063, 14092, 41006, and 147091, the JSI values of the C-V model are very close to those of the SPFLIF-IS model.For image ID 227092, the JSI values yielded by the C-V and LBF models are close to those of the SPFLIF-IS model.However, Figure 7 clearly illustrates that the SPFLIF-IS method yields greater JSI values than those generated by the C-V, LBF, and LIF models.
In the final part of this experiment, to fully validate the advantages of our SPFLIF-IS method in terms of the DICE and JSI, the five state-of-the-art methods ((1) the C-V model [34], (2) the LBF model [24], (3) the LIF model [31], (4) the SBGFRLS model [23], and (5) the LSACM model [28]) are applied to eight real color image data selected from the Berkeley segmentation data set 500 (BSDS500).The experimental results are shown in Table 7, where the Mean describes the average values of the DICE and JSI for all test image data.
The foregoing experimental analysis demonstrates that our proposed method is designed based on an improved SPF function and the LIF method.The model combines the merits of global image information and local image information and can segment noisy images and multiobjective images well.By contrast, the C-V model and the SBGFRLS model are constructed with global image information alone, based on the assumption that the region to be segmented is homogeneous.Unfortunately, this assumption is not suitable for intensity-inhomogeneous images [2,31,35].The LBF model and LIF model use local information to segment intensity-inhomogeneous images and obtain desirable segmentation results; thus, the models are sensitive to the initial position and image noise [2,36].
The LSACM model is proposed based on the local statistical information of an image; therefore, this model is robust to noise while suppressing intensity overlap to some extent.Nevertheless, this model is assumed that the image gray is separable in a relatively small area, and the offset is smooth in the entire image area.The model is easily trapped in a local minimum and involves high computational complexity [58,59].It follows from Table 7 that the values of the DICE and JSI of the SPFLIF-IS method are the highest for the eight real image data, and the Mean is also the largest.Thus, the experimental results obtained for synthetic and real images further demonstrate the superior performance of our method.Therefore, our model is able to obtain better DICE and JSI values than those yielded by the methods compared.

Discussion
According to the experimental results and evaluations presented above, the validity and stability of our proposed model are fully verified, and the contributions of the proposed model can be summarized as follows.
(1) The new model is regularized by a Gaussian kernel, which avoids the expensive computation associated with reinitialization.It follows that the model has low computational complexity.
(2) Our proposed model makes the best use of global and local image information.The model solves the problem of not accurately segmenting intensity-inhomogeneous images faced by the traditional image segmentation model and overcomes the shortcoming that local image information is sensitive to the initial contour and noise.
(3) Compared with the existing C-V, LBF, SBGFRLS, LIF, and LSACM models, the SPFLIF-IS model exhibits high robustness to the initial contour and noise and quickly and accurately segments inhomogeneous and multiobjective images.

Conclusions
In this paper, to segment intensity-inhomogeneous images quickly and accurately, an image segmentation method using a novel ACM based on an improved SPF function and an LIF model is proposed.The model combines the advantages of global and local information terms in segmenting intensity-inhomogeneous images.Moreover, a weight function is established to adjust the weights between the local information term and the global information term.Thus, a novel ACM model is presented, and an image segmentation algorithm is thereby established.To demonstrate the effectiveness of our proposed model, several experiments are designed in our study.The results indicate that our model not only segments inhomogeneous and multiobjective images effectively but also exhibits high robustness to the initial contour and noise.However, at present, it is difficult to determine a suitable Gaussian kernel size for all the images, and considering the uncertainty of real-world complex images, the proposed method will not be suitable in all the cases.As future work, we plan to accommodate the Gaussian kernel size automatically, which can be used to control region scalability from a small neighborhood to the entire image domain.This approach is considered to be more accurate and efficient in segmenting complex images and reducing computational complexity.

Figure 1 .
Figure 1.The graphical process of image segmentation.

Algorithm 1 .
SPFLIF-IS Input: An original image Output: The result of image segmentation

Figure
Figure 2b,d,f shows that the C-V model, the SBGFRLS model, and the LSACM model fail to analyze the first image with intensity inhomogeneity.As shown in Figure 2d,f, the SBGFRLS model and the LSACM cannot yield the ideal segmentation results for the second image.The object boundaries of the third image are not identified by the LIF model, and the results are shown in Figure 2e. Figure 2e,f shows that the true boundaries of the fourth image are not accurately extracted by the LIF model or the LSACM model.The SPFLIF-IS model detects the true boundary, and the results are illustrated in Figure 2g.Meanwhile, Figure 2c,g shows that the LBF model perform as well as the SPLIF-IS model.Note that because the visual evaluations in Figure2are partial to subjective measures, to strengthen the objective results of our experiments, the corresponding tables should be added to defend the arguments for all the tested images in the following visual evaluations, in which each failure is clearly labeled to avoid ambiguity.To more clearly illustrate this state, the following symbols are adopted in the tables: F1: fail to detect boundaries, F2: nonideal boundaries detected, F3: fail to detect internal boundaries, and T: true boundaries detected.Table1objectively describes the segmentation results of Figure2in detail.It can be clearly concluded from Table1that the LBF performs as well as the SPLIF-IS, the C-V exhibits slightly bad results, and the LSACM produces the worst results.Therefore, the experimental results shown in Figure2and Table 1 indicate that the SPFLIF-IS model can analyze the images with intensity inhomogeneity well.

Figure 4 .
Figure 4.The segmentation results obtained for images with strong noise using the six models.(a) Original image, (b) C-V model, (c) LBF model, (d) SBGFRLS model, (e) LIF model, (f) LSACM model, and (g) SPFLIF-IS model.

Figure 5 .
Figure 5.The segmentation results obtained for a texture image using the six models.(a) Original image, (b) C-V model, (c) LBF model, (d) SBGFRLS model, (e) LIF model, (f) LSACM model, and (g) SPFLIF-IS model.Table 4. Description of the segmentation results of Figure 5. Methods C-V LBF SBGFRLS LIF LSACM SPFLIF-IS Segmentation performance

Figure 7 .
Figure 7. JSI values of the image segmentation results using the four models for Berkeley color images.Table 7. DICE and JSI values of the results of image segmentation on fifteen Berkeley color images.Image ID

Table 1 .
and Table 1 indicate that the SPFLIF-IS model can analyze the images with intensity inhomogeneity well.Description of the segmentation results in Figure

Table 3 .
Description of the segmentation results of Figure 4.

Table 6 .
DICE values of the image segmentation results using the four models for Berkeley color images.