Next Article in Journal
Distributed Data Classification with Coalition-Based Decision Trees and Decision Template Fusion
Previous Article in Journal
Transformer and Pre-Transformer Model-Based Sentiment Prediction with Various Embeddings: A Case Study on Amazon Reviews
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Masi-Entropy Image Thresholding Based on Long-Range Correlation

by
Perfilino Eugênio Ferreira Júnior
1,
Vinícius Moreira Mello
1,
Enzo P. Silva Ribeiro
1 and
Gilson Antonio Giraldi
2,*
1
Department of Mathematics, Federal University of Bahia, Salvador 40170-110, BA, Brazil
2
Coordination of Mathematical and Computational Methods, National Laboratory for Scientific Computing, Petrópolis 25651-075, RJ, Brazil
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(12), 1203; https://doi.org/10.3390/e27121203
Submission received: 16 September 2025 / Revised: 1 November 2025 / Accepted: 18 November 2025 / Published: 27 November 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

Entropy-based image thresholding is one of the most widely used segmentation techniques in image processing. The Tsallis and Masi entropies are information measures that can capture long-range interactions in various physical systems. On the other hand, Shannon entropy is more appropriate for short-range correlations. In this paper, we have improved a thresholding technique based on Tsallis and Shannon formulas by using Masi entropy. Specifically, we replace the Tsallis information measure with Masi’s one, obtaining better results than the original methodology. As the proposed method depends on an entropic parameter, we designed a thresholding algorithm that incorporates a simulated annealing procedure for parameter optimization. Then, we compared our results with thresholding methods that use just Masi (or Tsallis), or a combination of them, Shannon, Sine, and Hill entropies. The comparison is enriched with a kernel version of a support vector machine, as well as a discussion of our proposal in relation to deep learning approaches. Quantitative measures of segmentation accuracy demonstrated the superior performance of our method in infrared, nondestructive testing (NDT), as well as RGB images from the BSDS500 dataset.

1. Introduction

Image segmentation is a mid-level processing technique used for image analysis. Such a method consists of splitting an image into several disjointed parts by grouping the pixels to form homogeneous regions regarding pixel features like intensities, textures, heat signatures, among other characteristics. The regions formed should be visually distinct, and their homogeneity should be based on correlations among image pixels. Errors in the segmentation could be caused by effects of illumination, shadowing, noise, partial occlusion, and subtle object-to-background changes.
One of the most basic and well-known ways to segment grayscale images is thresholding, which works by selecting one or more gray levels of the image, called thresholds, to separate objects from their background. When the objective is to separate the object region (foreground) from the background through only one gray level, the method is called bi-level thresholding [1,2,3,4,5,6,7,8]. This form of region segmentation results in a binary image, in which each region is either white or black. Otherwise, if the separation of the regions of interest from the background depends on different gray tones, the process is called multilevel thresholding [9,10,11,12,13,14,15,16].
The focus of this work is bi-level image thresholding. Entropy-based approaches constitute the most common techniques in this area. Basically, thresholding is implemented through an optimization technique that consists of selecting a threshold that maximizes the entropy of the segmented image. For instance, Kapur et al. maximize the Shannon entropy of segmented classes to obtain the optimal threshold values [17]. Suboptimal results of image thresholding can also be considered to deal with the time-consuming computation involved. In this line, the work [8] proposed to segment infrared (IR) images based on the maximum entropy of 2 D histograms and the PSO algorithm. Other works on the maximum entropy method proposed over the years have been based on Rényi’s [18], Tsallis [3], and Kaniadakis entropies [19].
IR images have a lot of low-frequency information, and the experimental tests have shown that some classical thresholding methods [5,6,7] are not efficient in this case. IR imaging is very useful in the military field for detecting objects with strong heat signatures, such as equipment and troop motions [20,21,22]. In addition, we can mention medical image applications [23], power equipment fault detection [24], and pedestrian detection on a scene [25].
Pixel intensities maintain a local long-range correlation in the neighborhood of the pixels in regions of an image. This fact is explored in [5], which is based on the long-range correlation between gray levels of an IR image and Tsallis entropy. In this case, the long-range interactions between the pixels are captured by the entropic parameter q inherent to the Tsallis entropy formula.
Another entropy from the context of thermodynamic science tested on IR images is the Masi one [6] that arises from the analysis between Rényi and Tsallis entropies [26]. These entropies cannot simultaneously manipulate long-range correlation, long-term memory, and fractal behavior [6]. Like Tsallis, Masi entropy is nonextensive, with a degree of nonextensivity measured by a parameter r.
The image histogram is a fundamental source of information for entropy-based methods [27,28]. In general, the appearance of a histogram is not that of two symmetrical portions with respect to an axis (threshold), which imposes difficulties on the thresholding techniques. To address this issue, ref. [29] presented a model that incorporates the Gumbel distribution to improve thresholding via cross-entropy in skewed histograms. In the same line, ref. [30] presents a method for automatic thresholding that allows selecting reasonable thresholds for images with unimodal, bimodal, multimodal, and peakless histograms.
The basic hypothesis for entropy-based bi-level thresholding methods is that the object and background regions are independent of each other. Recently, a method based on breaking this paradigm was proposed in [31], through the Tsallis entropy correlation.
In this work, we propose a new thresholding method based on long-range correlation and Masi entropy. Specifically, inspired by the methodology presented in [5], the novel technique computes the threshold through a max min optimization problem that involves the Masi and Shannon entropies.
The parameter selection of nonextensive entropies requires some effort on the part of the user. Optimization techniques and empirical methods have been extensively researched on this subject [4,32,33]. However, nothing good enough to obtain an optimal parameter that provides better thresholding has yet emerged in the literature. In our case, besides the selection by the user himself, we have provided a thresholding algorithm that includes a procedure based on the simulated annealing technique to automatically seek the near-optimum value of the entropic parameter r.
The experiments are concentrated on IR, nondestructive testing (NDT) images, as well as a subset of the colored images taken from the BSDS5000 dataset [34]. The main goals are to demonstrate the generalization capabilities of our approach and to allow the comparison with competing works. The results proved to be very competitive in relation to those of [3,5,6], with very low error measures. The former is the baseline for our proposal, while [6] applies only Masi, and [3] uses only Tsallis entropy. These facts have motivated the comparison between these state-of-the-art approaches and our proposal. In addition, we have used more recent techniques published in [35,36,37,38], as well as a support vector machine (SVM) implementation, from the Scikit-Learn library, to show how the proposed method fares versus a traditional learning-based approach. The efficiency of our proposal is assessed through measures of segmentation accuracy, demonstrating the superior performance of our method in infrared, nondestructive testing (NDT), and RGB images from the BSDS500 dataset. We also discuss aspects of our proposal in comparison to deep learning.
The organization of this paper is as follows: Section 2 discusses works related to our proposal. Section 3 presents the entropy functions investigated here. In Section 4, our method is presented. Then, Section 5 presents a theoretical analysis based on graph elements and the Principle of Maximum Entropy (PME). Section 6 shows the obtained computational results, and Section 7 discusses the main findings and some issues. Additionally, in Appendix A, a variation of our thresholding algorithm is presented to support ablation analysis.

2. Related Work

Nowadays, image segmentation approaches can be divided into two classes: traditional and deep learning methods. The former is composed of techniques based on edge extraction, fuzzy and morphological concepts, region representation, partial differential equations (PDEs), graph formulations, as well as stochastic and thresholding approaches [39]. On the other hand, deep learning methods rely on the universality of neural network computing, large annotated databases, theoretical and hardware developments that allow the training of deep architectures in feasible times [40]. Hence, despite the outstanding results obtained by deep neural networks in segmentation tasks, such performance is only possible if there is sufficient data for training, with proper diversity and representativeness, or if we can design appropriate architectures and/or methodologies for transfer learning, data augmentation, and data imbalance, which are tasks with their own issues (see [41,42] and ([43], Section 2.2)).
Thresholding methodologies achieve competitive results even if a small dataset is accessible for testing [1]. These methods do not depend on time-consuming training stages, like deep learning and some stochastic approaches. The parameter setting is simpler than edge extraction (like active contour models) and graph formulations (like graph cuts). Its formulation is simple and intuitive, allowing the development of algorithms with complexity lower than PDE-based methods, especially in the bi-level case. Those observations support the research in thresholding methods nowadays, as we have noticed in recent works for segmentation of thermograms [44], medical imaging [45], satellite images [46], color images [47], among others [48,49].
In [1], thresholding methods are separated into six groups depending on whether they are based on features related to the histogram shape, object attributes, stochastic correlation between pixels, local image characteristics, clustering algorithms, or entropy concepts. The latter class is derived inside information theory methodologies that are grounded on statistical mechanics elements [50]. This viewpoint started in [2] by applying the Shannon entropy and the concept of anisotropy coefficient. On the other hand, if we consider the entropy as the measure of the information contained in the image, then the best threshold is the one that maximizes the entropy of the result. This line is followed by [17] that seeks a threshold t that partitions the histogram into two regions that maximize the Shannon entropy of the corresponding joint probability distribution. However, such a proposal may fail if applied to distinct images with identical histograms, as pointed out in [18]. One direction to address this issue is the application of more general entropic concepts that include new parameters that can be customized to fit the application requirements.
That is the case of Renyi’s entropy used in [18] for bi-level thresholding, which uses an approach analogous to the one presented in [17] but replacing Shannon entropy with Renyi’s one. Shannon and Renyi entropies characterize systems known as extensive ones [51]. On the other hand, some nonextensive systems follow another entropy formula proposed by Tsallis in [52].
Tsallis entropy includes the entropic parameter q R and we can show that, in the limit q 1 , it equals the Shannon information measure. Tsallis entropy is the foundation of a new formalism in statistical mechanics where the level of nonextensivity of a physical system is quantified by the parameter q [53]. The work [3] was the first one to use the Tsallis entropy formalism for bi-level thresholding. Then, multilevel segmentation [9] approaches have been developed using Tsallis entropy with applications in medical imaging [45], multispectral image analysis [54], among others.
In this scenario, an important point is the distinction between long-range and short-range correlations. Specifically, image fields characterized by long-range correlations are more efficiently represented using Tsallis entropy, while Shannon entropy models short-range correlations [3,5].
Such a distinction raised the issue about the utilization of another entropy measure that is more flexible with respect to long-range correlation events but preserves the Shannon entropy characteristics. The Masi entropy, used in [6], is a choice in this direction that is dependent on the entropic parameter, r such that the limit r 1 recovers the Shannon measure. These facts have motivated our proposal that puts together Masi and Shannon entropies in a new thresholding approach. Our work is inspired by [5], which applies a max min optimization problem that involves Tsallis and Shannon information measures.

3. Entropy Functions

Entropy is the measure of disorder in physical systems or the measure of the amount of information that may be needed to specify the full microstates of the system. In 1948, Shannon added entropy to information theory, and his approach measures the uncertainty associated with a random variable or the amount of information produced by a process [55] as
S ( X ) = i = 1 n p i log p i ,
where X is a random variable that can take values { x 1 , , x n } , and p i = p ( x i ) is the corresponding probability of x i . Equation (1) is the Shannon entropy that describes systems that obey the following additive property: Let A and B be two random variables associated with independent statistical subsystems of a physical system. Then,
S ( A + B ) = S ( A ) + S ( B ) .
Systems of this kind are called extensive systems. A certain class of physical systems, which entail long-range interactions, long-time memory, and fractal-type structures, indicated the need for an extension. Tsallis entropy [52] extends its applications to so-called nonextensive systems using an adjustable parameter q. Tsallis entropy can explain a complex system class, such as long-range and long-memory interactions. It can be expressed as
S q ( X ) = 1 1 q i = 1 n p i q 1 ,
where X is a random variable, q R and S q ( X ) converge to S ( X ) in the limit q 1 . The function S q ( X ) has the following pseudo-additivity property for q 1 :
S q ( A + B ) = S q ( A ) + S q ( B ) + ( 1 q ) S q ( A ) S q ( B ) ,
where A and B are independent subsystems of a physical system. In 2005, Masi proposed a new entropy that combines the nonextensivity of Tsallis entropy and the additivity of Rényi’s entropy [26], namely
S r ( X ) = 1 1 r log 1 ( 1 r ) i = 1 n p i log p i ,
where r R ( r > 0 and r 1 ) . Moreover, S r satisfies
S r ( A + B ) = S r ( A ) + S r ( B ) ,
with A and B as before. The parameters q and r can be viewed as measures for the degree of nonextensivity that exists in the system for the Tsallis and Masi entropies, respectively.

Entropy Functions and Image Thresholding

For the image thresholding context, we will consider an image with n gray levels, and let p 1 , p 2 , , p n be the probability distribution of the levels. Here, two probability distributions can be derived from the original distribution, one for the background (class A) and the other for the object (class B). Their probabilities can be given by
A : p 1 P A , p 2 P A , , p t P A , B : p t + 1 P B , p t + 2 P B , , p n P B ,
where P A = i = 1 t p i and P B = i = t + 1 n p i .
In this case, Albuquerque’s method obtains an optimal threshold by maximizing the information measure between the two classes, where the objective function S q A + B = S q A + S q B + ( 1 q ) S q A S q B is parametrically dependent upon the threshold value t that separates the foreground and background. Hence, Albuquerque’s solution is given by
t o p t = a r g m a x S q A + B ( t ) .
The nonextensive parameter q represents the strength of the long-range correlation. Equation (4) indicates that there is a global correlation not only in the areas of foreground and background but also between them. Moreover, the strength of the global correlation is described by the same value q.
Regarding Masi entropy applied in Nie et al.’s method [6], although this entropy is nonextensive, it is also additive. The parameter r establishes the degree of nonextensivity and the strength of the long-range correlation property. The long-range correlation could be indicated by the Equation (6) whose optimum,
t o p t = a r g m a x S r A + B ( t ) ,
gives the threshold value in [6].
In practice, the global long-range correlation indicated by Equation (4) does not apply for IR images since the thresholding results only show local long-range correlation [5] in this case. Thus, the optimal value found by the maximization process is not sensitive to small variations of the nonextensive parameter q, as reported in [5]. In that reference, the authors argue that it would be inappropriate for these types of images to say that there is a global long-range correlation. The long-range correlation would be weaker in the background of some images in the context of Tsallis entropy. Thus, instead of simply maximizing the sum of the entropies of object and background, they decided to maximize both the Shannon entropy on the background of the image and Tsallis on the object. However, it is hard to obtain the absolute maxima of them with a single threshold unless their thresholds are equal. For this reason, the solution involved a trade-off that is addressed in [5] as the optimization problem:
t o p t = a r g m a x m i n S A ( t ) , S q B ( t ) ,
where S A is the Shannon entropy calculated on the background of the image and S q B is the Tsallis entropy calculated on the object. Some tested images apparently had a stronger long-range correlation in the background of the image. To mitigate this effect, Lin & Ou [5] also proposed an alternative way for the trade-off, interleaving the entropies:
t o p t = a r g m a x m i n S q A ( t ) , S B ( t )
where S q A is the Tsallis entropy calculated on the background of the image and S B is the Shannon entropy calculated on the object.

4. Proposed Method

Our proposal is to combine the Masi and Shannon entropies in a trade-off similar to that used by Lin & Ou in Equations (10) and (11). Two facts motivated the current proposal involving Masi entropy. First, its image thresholding results do not present a weak correlation in the background of images for the appropriate parameter r. Therefore, Masi entropy could be thought of as a method based on global long-range correlation. Second, the optimal value found in the maximization process of Masi entropy is very sensitive to the variation of the nonextensive parameter r. Thus, we propose a trade-off as follows:
Proposal 1 : t o p t = a r g m a x m i n S A ( t ) , S r B ( t )
where S A is the Shannon entropy calculated on the background of the image and S r B is the Masi entropy calculated on the object. When there are distributions of different gray levels in the image, we can propose an alternative way as in Equation (11).
Proposal 2 : t o p t = a r g m a x m i n S r A ( t ) , S B ( t )
where S r A is the Masi entropy calculated on the background of the image and S B is the Shannon entropy calculated on the object. This would be the case of having a weaker long-range correlation in the background of the image, for example.
The selection of r value can be executed by trial-and-error tests like performed in [3,5,6] to set q and r values. Although some works have been developed to automatically set entropic parameters [4,56,57], this subject remains an open issue in image processing. In our case, we uniformly sample a set of values over a predefined interval and select the one that best approximates the target segmentation, according to problem (9), to refine the search. The obtained value r = r is used to initialize the search for the best r for problem (12). Additionally, we offer a thresholding algorithm that includes a SA technique for entropic parameter optimization, to improve reproducibility and avoid ad hoc procedures for seeking the near-optimum r for our proposal.
The alternative form is used when a thresholding result close to the optimal value is not achieved by Equation (12). Then, we seek another r value to execute the optimization process given by Equation (13), instead of Equation (12). Experimentally, in all cases, it was possible to obtain a near-optimum solution through the ability to capture the long-range correlation properties among the pixels due to Masi entropy. In the following section, we report the results and offer quantitative comparisons with the other methods.

5. Theoretical Analysis

This section offers a justification for the proposed criterion in expressions (12)–(13) and uses the Principle of Maximum Entropy (PME) to discuss the long-range and short-range correlations in Tsallis, Shannon, and Masi entropies.

5.1. Justification by Isoperimetric Theory

It this section we follow [7,38] and apply graph elements to give a mathematical justification for Proposals 1 and 2 in Section 4. Thus, we start by modeling a grayscale image as a weighted undirected graph G = V , E , with V being the set of pixels and E V × V being a set of edges e i j = v i , v j V × V linking adjacent pixels v i and v j in V. Furthermore, we define a function w : E R + that calculates normalized weights:
w i j = exp β I v i I v j 2 E exp β I v m I v n 2 ,
where I : V 0 , 1 , 2 , , L represents gray level function. Hence, we define the following:
  • Degree of v i V : d i = w i j , e i j E .
  • Complement of S: S ¯ = V S .
  • Volume of S:
    d S = v i S d i ,
Given S V , the boundary of S is represented as S = e i j | v i S a n d v j S ¯ . Then, we can define
cut S , S ¯ = e i j S w i j .
In the thresholding segmentation theory, each set S is specified by a threshold t and, consequently, S = S t . With those elements, to each set S t V , we can define the isoperimetric ratio as follows [58]:
h S t = cut S t , S ¯ t min d S t , d S ¯ t .
Now, a global graph feature named isoperimetric constant is computed as follows [58]:
h G = min 0 t L cut S t , S ¯ t min d S t , d S ¯ t .
Expression (18) is applied as a segmentation criterion in [58]. The justification behind such a criterion is that cut S , S ¯ decreases when considering regions S and S ¯ with high contrast between each other. Moreover, some degree of homogeneity is expected in the interior of S (or S ¯ ), which increases the value of d S (or d S ¯ ). By considering min d S , d S ¯ in the denominator, there is a preference for uniform-intensity regions instead of larger areas. Homogeneous-intensity regions decrease the Masi and Shannon entropies since their maximum occurs for regions with a uniform-intensity histogram. Consequently, we can replace d S and d S ¯ to S A ( t ) and S r B ( t ) in expression (18) to obtain
h G = min 0 t L cut S t , S ¯ t min S A ( t ) , S r B ( t ) .
However, the weights w i j computed through Equation (14) are normalized, resulting in
h G = min 0 t L cut S , S ¯ min S A ( t ) , S r B ( t ) min 0 t L 1 min S A ( t ) , S r B ( t ) ,
which renders
t o p t = arg max 0 t L min S A ( t ) , S r B ( t ) ,
which is the proposed criterion in Equation (12). Analogously, we obtain Equation (13) by replacing d S and d S ¯ to S r A ( t ) and S B ( t ) in expression (18).

5.2. Long-Range Versus Short-Range Correlations

In this section, we follow [59] to formalize the stationary short-range correlation by an exponential function:
S x , y ; θ , σ = 1 2 θ exp x y σ , θ , σ > 0 ,
while the stationary long-range correlation will be described through the power-law function:
L x , y ; τ , μ , γ = τ 1 1 + x y / γ μ , τ , μ > 0 .
Firstly, if we perform the substitution x y d in expressions (22)–(23) and set θ = 0.5 , σ = 1.0 , and τ = μ = 1.0 , we obtain the plot in Figure 1:
We notice that for both models, the correlation decreases with respect to the distance d but function S x , y ; 0.5 , 1.0 decays faster than its counterpart. Expressions (22)–(23) are inspired in Shannon (S) and Tsallis ( S q ) entropies, respectively, as defined by expressions (1) and (3).
The second-order statistics associated with the information measures (1)–(3) are expressed by
i = 1 W p i = 1 ,
i = 1 W e i p i q i = 1 W p i q = U q ,
i = 1 W e i U q 2 p i q i = 1 W p i q = V q ,
where equation (24) is necessary for p i to be the probability associated to the state e i , for q = 1 expressions (25)–(26) recover the regular mean and variance quantities.
The Tsallis entropy offers a new formalism in which the real parameter q quantifies the level of nonextensivity of a physical system [53]. In particular, a general Principle of Maximum Entropy (PME) has been considered to find out distributions to describe such systems. In this PME, the goal is to seek the maximum of S q subjected to constraints (24) and (26), which gives the solution:
p j = 1 + ( q 1 ) k β e j U q 2 V q i = 1 W p i q 1 1 q j = 1 W 1 + ( q 1 ) k β e j U q 2 V q i = 1 W p i q 1 1 q ,
where β is a Lagrange multiplier and the denominator is just a normalization factor in order to fit constraint (24). Although expression (27) is not practical since the right-hand side also depends on the unknowns p j , it is usefull from theoretical viewpoint. Specifically, by comparing this expression with function (23) with the substitution x y d , we notice that if we set
d = ( q 1 ) e j U q 2 V q i = 1 W p i q ,
γ = k β ,
τ = j = 1 W 1 + ( q 1 ) k β e j U q 2 V q i = 1 W p i q 1 1 q ,
μ = 1 1 q ,
then we recognize the power-law kernel in both expressions (23)–(27), indicating that Tsallis entropy is more appropriate for model long-range correlation. On the other hand, an analogous PME can be formalized using the Shannon entropy (1). In this case, we must set q = 1 in the expression (25) and the obtained result is
p j = exp β e j U 1 2 j = 1 W exp β e j U 1 2 ,
where β is the Lagrange multiplier again. In this case, the comparison with the stationary short-range model (Equation (22)) shows that the exponential function dominates both expressions.
On the other hand, when considering the PME for Masi entropy, we obtain
p j = exp β r 1 + ( 1 r ) S p j r 1 e j U r 2 V r i = 1 W p i r S j = 1 W exp β r 1 + ( 1 r ) S p j r 1 e j U r 2 V r i = 1 W p i r S ,
which shows a more complex dependency that includes the Shannon entropy (S) besides the power-law p j r 1 . Hence, expression (29) shows that Masi entropy is more flexible regarding the trade-off long-range versus short-range correlation events being more appropriate for a wide range of images, as demonstrated in the computational experiments.

6. Experimental Results

In this section, the computational experiments are presented to demonstrate the capabilities of our proposal for segmentation tasks. Firstly, the experiments focus on IR and nondestructive testing (NDT) images to allow the comparison with the baseline [5], as well as [6,17] for ablation analysis, since [6] applies only Masi and [17] applies only Shannon entropy. Additionally, comparisons with [3] that uses only Tsallis is provided since we replace the Tsallis entropy with Masi’s one in our approach.
The parameterization of our proposal is also a point focused in the computational experiments since the results of Section 6.1 and Section 6.2 are obtained with ad hoc strategies. In this way, in Section 6.3, we present an algorithm that summarizes our proposal and includes a simulated annealing strategy to select the r value. With this algorithm, parameter selection and sensitivity analysis are discussed.
Section 6.4 delves into the ablation study by isolating the effect of Masi entropy in our proposal. Then, to show that our segmentation approach can be successfully applied to other types of images besides IR and NDT ones, experiments involving BSDS500 images are presented in Section 6.5. Additionally, a comparison with more recent entropy-based approaches is presented, and a support vector machine (SVM) is employed to demonstrate how the proposed method performs in comparison to learning-based approaches.
The set of NDT images used is composed of grayscale images with resolution 256 × 256 pixels. With respect to the IR images, they are also grayscale ones, but the resolution varies from 160 × 182 to 320 × 240 pixels. Both datasets can be acquired from the Google Drive (https://drive.google.com/file/d/1LzQNCjK4YeTzqGf-hYQ7Q_osi2Ery9m8/view?usp=sharing, (accessed on 15 September 2025)) and were assembled from freely available images on the Web, which were also used elsewhere [5,6,7,38].
The BSDS500 database [34] is composed of RGB images with resolution 481 × 321 pixels selected from the Corel image database using the following criterion: natural scenes that contain at least one discernible object. Besides the RGB images, the BSDS500 database gives the corresponding segmentation, given by the delineation of the target boundary of the objects by humans. Such ground truth is not suitable for our purposes, which require pixel-wise classification between foreground and background. Hence, we took 50 images and pre-processed their segmentations to generate the ground truth with the desired characteristics for our application. The obtained dataset is named BSDS500-50 throughout this text. Figure 2 presents some samples of BSDS500-50 together with their ground truth.
To infer the performance of our method, we adopted four quantitative measures. One of them is the misclassification error (ME) measure [1]. It can be written as
M E = 1 | B G T B T | + | F G T F T | | B G T | + | F G T |
where B G T and B T are the pixels of background in the ground-truth image and thresholded image, F G T and F T are the pixels of foreground in the ground-truth image and thresholded image, and | | is the cardinality of a set. The ME varies from 0 for a perfectly segmented image to 1 for a totally wrong binarized image. Another measure is the Jaccard similarity (JS), used in [20,60], which can be written as
J S ( F G T , F T ) = | F G T F T | | F G T F T | .
Differently from Equation (30), the correct segmentation obtains J S = 100 % . The third measure is the relative foreground area error (RAE) [61] that represents the accuracy and completeness of the segmented foreground. It can be expressed as
R A E = | | F G T | | F T | | m a x ( | F G T | , | F T | ) .
The lower the RAE value, the better the segmentation result. The fourth measure is F-measure [62,63] that is defined as the harmonic mean of precision and recall rate with weight α , which represents a compromise between under-segmentation and over-segmentation. Its value is provided by the expression
F α = ( 1 + α ) · P · R ( α · P ) + R ,
where P and R are the precision and recall rates given by
P = | F G T F T | | F T | a n d R = | F G T F T | | F G T | .
Generally, larger values for P and R indicate better segmentation results. F ff could be a balanced measure for segmentation results through a good choice of harmonic coefficient α . Usually, α is set to 0.5 for segmentation algorithms [62,63]. However, a segmentation that matches the ground truth obtains F α = 1 regardless of the value of α .
In addition, we apply the uniformity measure ( U M ) to evaluate the segmentation when the ground truth is not available. Specifically, given regions R 1 and R 2 of the grayscale image I, defined by the segmented image, the U M is computed as follows [64]:
U M R 1 , R 2 = 1 2 · σ 1 2 + σ 2 2 N · M · I m a x I m i n 2 ,
where σ i 2 is the variance with respect to the mean of region R i , N , M are the image dimensions, and I m a x , I m i n denote the maximum and minimum pixel intensities in the image. Consequently, a higher U M value indicates lower intra-region variances.
The most detailed part of applying parameterized entropies is choosing the entropic parameters. For the experiments performed with Tsallis entropy, the values of q were obtained as follows:
  • We automatically searched, in the range [ 0.1 , 0.9 ] with increments of 0.1 units, for the best value of q that minimized the ME measure between the segmented image using just the Tsallis entropy (Equation (4)) and the ground-truth image.
  • From this, the parameter q found is used in the combination of the Tsallis and Shannon entropies (Equation (10)). This can be corroborated in all the Figures generated in the text.
  • If ME value is larger than 0.1 then the same procedure is repeated using Equation (11).
On the other hand, the entropic parameter r is somewhat more complicated to handle. The following steps were next performed in Section 6.1 and Section 6.2:
  • The variation range of r, including all types of images tested, was in the range [ 0.7 , 1.7 ] , with increments of 0.1 units.
  • As we vary r in Masi expression (9), we notice that the ME values could be improved by oversampling within the current range. For example, if the ME value decreases for r within the subrange [ 1.3 , 1.4 ] , closer to the value 1.3 , we then test the value r = 1.32 . If the ME value decreases, we establish a new subinterval around this point, say [ 1.312 , 1.329 ] . In such a range, we should find a value of r * that minimizes M E .
  • Unlike the Tsallis entropy parameter, in general, the best value of r * for Masi entropy alone does not coincide with the best r value for the proposed max–min strategy in Equation (12). In the latter case, the same procedure is applied (items (a)–(b)), but now using Equation (12).
Items (a)–(c) above are ad hoc. A systematic algorithm will be presented in Section 6.3, based on simulated annealing (SA), to improve the generalization and reproducibility of our results.

6.1. Performance on IR Images

Table 1 shows thresholding values, as well as the number of misclassified pixels, ME, JS, RAE, and F α values, for the images. Additionally, the results with lower ME and RAE values and higher JS and F α values are marked in bold. We can observe in the last column of Table 1 that our proposal proved to be the most efficient, or tied for the best, among the experiments carried out.
Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 show more details about the results generated with the reported image thresholding methods:
  • Figure 3—000280 image: We notice that Lin & Ou’s method had an RAE equal to zero, which implies that the number of pixels in the foreground of the ground truth matched the number of pixels in the foreground of the thresholded image. However, visually, we can notice differences between the ground truth and the segmented image by Lin & Ou’s method. The other measures corroborate this observation, showing that the behavior of our method was slightly better than that of Lin & Ou’s as well as Nie et al.’s methods, whose results were similar. The proposed method exhibited the highest JS value for that image.
  • Figure 4—Airplane image: Our proposal surpasses all the others obtaining JS (Equation (31)) greater than 99 % and a F α value close to 1. Albuquerque’s and Lin & Ou’s methods match up and show inferior results if compared to the second-best approach, that is, Nie et al.’s method in this case. The image obtained with Lin & Ou’s method was generated with the alternative form given by Equation (11). This should be an indication that the correlation in the region of background is most strongly captured by Tsallis entropy (see the sharp peak in the histogram in Figure 4b).
  • Figure 5—Tank image: Our method overcomes all other techniques and provides more than 98 % of Jaccard similarity (Equation (31)). This example shows a significant difference with respect to the misclassified pixels and JS value in relation to the second-best result, that of Nie et al. Albuquerque and Lin & Ou’s methods perform far from our technique for all the considered measures.
  • Figure 6—Panzer image: Our method and Nie et al.’s approach achieve perfect segmentation. The number of misclassified pixels, ME, and RAE values are zero. The image obtained with Lin & Ou’s method was generated with the alternative form (11). However, both Albuquerque and Lin & Ou’s methods obtain segmentations far from the ground truth, as indicated by the values of the considered measures.
  • Figure 7—Car image: Our result outperforms their counterparts. Again, the alternative form (11) for Lin & Ou’s method was used to generate the corresponding image.
  • Figure 8—Sailboat image: Our results are much superior to other methods for all the measures considered. The image obtained with Lin & Ou’s method was generated with the alternative form (11).
The error measures reported in Table 1 show that our proposal is more efficient than competing techniques. The second-best technique, Nie et al.’s method, performs similarly as our approach only in the Panzer image, being inferior in the other cases. In addition, our proposal for image thresholding based on local long-range correlation and Masi entropy provides much better results than those based on Tsallis entropy. A range of q values used to generate images of Albuquerque’s and Lin & Ou’s methods was 0.1 q 0.9 .
The ranges of r values used for Nie et al.’s method and ours were 0.9 r < 1.4 and 0.7 < r < 1.3 ( r 1 ), respectively. As observed in [5], the optimal value found by the Tsallis entropy maximization process is not sensitive to the small variations in q. Differently, Masi entropy maximization process is quite sensitive to variations in r, even small ones. This fact is relevant for establishing the step for sampling the above interval to set r in Equations (12) and (13).
Figure 9 shows a test performed with an image sequence. Such images contain an object (a person) moving on the same background. As in Lin & Ou’s work [5], we assume that the strength of the long-range correlation of the images should be similar for Tsallis entropy. The authors judged that the correlation strength would be the same for that object on the images, which yields the same optimal q value. Thus, we kept q fixed at 0.8 , the same value for image 000280 from Figure 3. For the Nie et al method and ours, the ranges of r values found were 1.3 r < 1.4 and 1.1 r < 1.3 , respectively.
It can be seen from Table 1 that the Shannon entropy data for IR images are the same as those for Tsallis entropy, except for the Panzer image. Even so, the measurement values were relatively close in that case. A possible cause for this effect is that the Shannon entropy is obtained from the Tsallis entropy as q tends to 1, since in our experiments, we use q 0.8 . In fact, in general, the Tsallis entropy does not change much with variations in q. Therefore, given a value q 1 , the Tsallis entropy would remain almost constant within a certain range around the q 1 value. Thus, for q close to 1, the values of the Tsallis and Shannon entropies can be nearly equal, leading to similar segmentation results. For this reason, the experiment with the image sequence in Figure 9 was not performed with Shannon entropy.
Table 2 shows the JS obtained for this experiment. Our approach achieves the best results or performs equally well as the best methods. Lin & Ou’s method ties with our approach in only one of the listed cases (000340 image). Nie et al.’s method ties with our method in three of the images in the sequence (000360, 000400, and 000440 images).
Table 3 shows the threshold values, misclassified pixels, and misclassification error (ME) for the segmentation of the infrared images (IR) in Figure 9. Data in Table 3 corroborate our approach as the best among those analyzed since it outperforms the competing methods or ties with the best-performing one. Specifically, Table 3 shows a tie between the proposed method and the results of Lin & Ou’s technique (000340 image) and four ties with the Nie et al. approach (000320, 000360, 000400, and 000440 images). This highlights the competitive potential of the proposed approach against Lin & Ou and Nie et al. methods, which are based on nonextensive entropies. In the 000320 image, the ME and misclassified pixel values for our method and those of Nie et al. are the same, given by 0.000143229 and 11, respectively. Despite this, Table 2 reports that our approach obtained a JS value of 94.93 % , while Nie et al. obtained 94.91 % . Hence, we notice that the difference between their JS values is 2 hundredths. This shows a slight advantage of our approach over Nie et al.’s method in this case.

6.2. Performance on NDT Images

Figure 10, Figure 11 and Figure 12 form a set of three nondestructive testing (NDT) images, which are further tested. Table 4 presents quantitative performance metrics for the methods on these images. The last column of Table 4 shows that our proposal is the most efficient among the approaches considered. Below, we offer more details about the quantitative results reported in Table 4.
  • Figure 10—Gear image: Although the Lin & Ou’s and Nie et al.’s methods have obtained a good visual approximation of the ground truth, the result of our technique matches the ground-truth image with J S = 100 % , according to Table 4. For this image, the computational experiments have shown that Albuquerque’s method is very sensitive to the variation of parameter q. Since this image was not applied in Lin & Ou’s paper, the q values were determined by trial-and-error to optimize performance. This allowed Albuquerque’s result to be more than 94 % accurate according to JS. The image histogram suggests how easy it would be for the method to separate the regions of the image. As shown in Figure 10b below, there is a sharp peak at the beginning of the histogram. It can be considered as the strong correlations of the pixels in the foreground. Thus, as performed in Lin & Ou’s experiments, the alternative form, (11), was necessary for this image. The same occurred for our method, which used an alternative form to Equation (12) given by Equation (13). This happens due to a weaker long-range correlation in the background composed of the lightest area of the image.
  • Figure 11—Pcb image: In this example, Nie et al.’s technique surpasses Lin & Ou’s, having more than 90 % of similarity with the ground truth image. The image obtained with Lin & Ou’s method was generated with the alternative form (11). Even so, our method is more efficient than the others, making the accuracy of 99.33 % , as shown in Table 4.
  • Figure 12—Cell image: Lin & Ou’s method is more effective than Nie et al.’s method, in this case. However, it still does not overcome our result, which is 100 % accurate (see Table 4).
The presented set of results proves the superiority of our method against its counterparts in the application to IR and NDT images.

6.3. Selection of Entropic Parameter

The procedure used to set the entropic parameter values in Section 6.1 and Section 6.2, as explained in the introduction of Section 6, is ad hoc. In this section, we show a systematic procedure summarized on the following iterative Algorithm 1 that incorporates the proposed thresholding method (Section 4) and applies simulated annealing (SA) to update the r value in each iteration, besides the uniformity measure U M , defined in expression (34), and an error measure to compare the binarized image with the ground truth (if available).
Algorithm 1 Pseudo-code for the thresholding approach proposed based on SA.
1:
Input: grayscale image I with N pixels; image intensity range: 0 , L ; T i n i t i a l T m i n ; α ; m a r k o v l e n g t h ; M a x i t e r ; ground truth segmentation I g t ; Reference error E r e f ; Reference uniformity measure U M r e f ; d; r ^ ; f i x _ p a r ;
2:
Calculate image histogram h = h t ;
3:
Proposal 1
4:
Initial guess of entropic parameter: r ^ ;
5:
T 0 T i n i t i a l ;
6:
Objective function:
F 1 t , r ^ = m i n S A ( t ) , S r ^ B ( t ) ; 0 t L ;
7:
Solve the optimization problem:
t o p t r ^ = arg max 0 t L F 1 t , r ^ ;
8:
Binarization of input image: B i n a r y r ^ = T h r e s h o l d i n g I , t o p t r ^ ;
9:
if  f i x _ p a r = = 0  then
10:
    go to line 44;
11:
end if
12:
if ground truth I g t is available then
13:
    Calculate Misclassification Error: E 0 = E r r o r I g t , B i n a r y r ^ ;
14:
    If E 0 E r e f then go to Line 44;
15:
else
16:
    Calculate U M ( j ) = U M ( t o p t r ^ ) ;
17:
    If U M j U M r e f then go to Line 44;
18:
end if
19:
i 0 ;
20:
while  T i > T m i n and i < M a x i t e r  do
21:
    for  j = 1 , j + + , while j < m a r k o v l e n g t h ; do
22:
        Update r j r ^ + r a n d d , d ;
23:
        Solve the optimization problem:
t o p t r j = arg max 0 t L F 1 t , r j ;
24:
        Binarization of input image: B i n a r y r j = T h r e s h o l d i n g I , t o p t r j ;
25:
        if ground truth I g t is available then
26:
           Calculate Misclassification Error: E j = E r r o r I g t , B i n a r y r j ;
27:
           If E j E r e f then go to Line 44;
28:
           Calculate difference: Δ E = E j E j 1 ;
29:
           Acceptance criterion: If Δ E 0 then r ^ r j else p r o b = exp Δ E / T ;
30:
           If Δ E > 0 and r a n d 0 , 1 < p r o b then r ^ r j ;
31:
        else
32:
           Calculate U M ( j ) = U M ( t o p t r j ) ;
33:
           If U M j U M r e f then go to Line 44;
34:
           Calculate difference: Δ U M = U M j U M j 1 ;
35:
           Acceptance criterion: If Δ U M 0 then r ^ r j else p r o b = exp Δ U M / T ;
36:
           If Δ U M < 0 and r a n d 0 , 1 < p r o b then r ^ r j ;
37:
        end if
38:
    end for
39:
     i i + 1 ;
40:
     T i T i 1 · α ;
41:
end while
42:
Proposal 2
43:
Repeat steps 2–34 with the Objective Function:
F 2 t , r = m i n S r A ( t ) , S B ( t ) ; 0 t L ;
44:
Output: Threshold t o p t and parameter r ^ .
In Algorithm 1, we include the flag f i x _ p a r to add the possibility of using the algorithm just to obtain the segmentation but without optimizing the entropic parameter. Expressions (35) and (38) implement the objective functions of our proposal. Lines 16–35 incorporate the SA procedure to seek the best r value. The computational complexity of Algorithm 1 is defined by the following steps:
  • Computational complexity of Proposal 1:
    (a)
    Line 2: O N ;
    (b)
    Line 7: O L 2 ;
    (c)
    Line 8: O N ;
    (d)
    Line 11: O N ;
    (e)
    Line 14: O N ;
    (f)
    Line 20: O L 2 ;
    (g)
    Line 21: O N ;
    (h)
    Line 23: O N ;
    (i)
    Line 29: O N ;
    (j)
    Line 37: Analogous to Proposal 1.
Consequently, based on items (a)–(j) above, the computational complexity of Algorithm 1 is
C o m p Algorithm   1
= O M a x i t e r · m a r k o v l e n g t h · max N , L 2 .
which is not a drawback of our proposal, as will be shown in Section 6.5. Appendix A shows a variation of Algorithm 1 obtained by using just the Masi entropy. This algorithm is used to perform ablation analysis in the remaining text.
Besides computational complexity analysis, other important points regarding Algorithm 1 must be considered with respect to the behavior of our approach when varying the entropic parameter. Hence, we propose the following strategy. We selected the image that provided the best value for ME in our initial experiments, the panzer image in Figure 6. We set the following parameterization of Table 5 for Algorithm 1 unless stated otherwise. Then, we establish the steps listed below.
  • Sensitivity test: Another important point is the sensitivity to entropic parameter perturbation for the proposed method. We perturb the r ^ returned by Algorithm 1 for the panzer image to generate the sequence r k = r ^ + k Δ r , where k = 3 , 2 , 1 , 0 , 1 , 2 , 3 and Δ r = 0.01 . Then, we compute the sequence Δ M E ( k ) = | M E ( k ) M E r e f | , where M E ( k ) is the misclassified error of the segmentation obtained with Algorithm 1 with r ^ = r k and f i x _ p a r = 0 in the input of Algorithm 1, since in this case the r value is fixed. Additionally, M E r e f = 0.000518798828125 , which is the reference ME value for the panzer image. The other input parameters are maintained as specified in Table 5. Table 6 shows the data obtained in this procedure. Although the sensitivity S increases when r k deviates from the reference value r 0 , we notice that 0 Δ M E ( k ) 0.00823975 , which indicates the stability of Algorithm 1 with respect to parameter variation.
  • Convergence test: Algorithm 1 is executed n times with the parameterization of Table 5. The first execution uses r ^ as specified in that table. Next, each time it is executed, a parameter r ^ is returned, and this parameter is then used again as the initial guess of r for the next execution. This process repeats for a number n of times or until the algorithm converges, that is, the value of ME is equal to 0 or is fixed at a value close to it. The goal here is to check the efficiency of Algorithm 1 to achieve a near-optimum r value; that means it is expected that Algorithm 1 converges within a small n.
    Figure 13 shows the behavior of Algorithm 1 for the case n = 11 . From the seventh execution of that Algorithm, the ME value converges to 0.0003204345703125 , indicating the capability of the SA algorithm to output a near-optimum value for the entropic parameter r.
  • Noise sensitivity: Corrupted versions of the panzer reference image were created using a Gaussian distribution to verify the behavior of Algorithm 1 against noise. In this case, 10 noise levels were generated with the standard deviation σ ranging from [ 0.02 , 0.04 , , 2.0 ] , with increments of 0.02 . Furthermore, because the noise pattern must be different each time it is generated, for each noise level we generate 20 noisy images, leading to a total of 200 images. An average of the ME values was calculated for each σ . The r and t o p t values used for thresholding each image were provided by Algorithm 1 with the parameterization of Table 5.
    The ME results are shown in Figure 14, which shows an unexpected behavior since the ME for σ = 6 · 10 2 is higher than the ME for σ = 0.18 . Despite this, the ME values fall below 0.17 in most cases, which suggests low sensitivity against noise.

6.4. Ablation Analysis

We shall notice that Algorithm 1 has a stochastic process inside the WHILE loop that starts in line 17 (analogous to Algorithm A1, in line 16) due to the utilization of function rand. Consequently, different executions of each algorithm with the same input may produce different results for r ^ . Steered by this fact, in this section, we kept the panzer image as input and generate twenty r ^ values using twenty executions of Algorithm 1 and twenty more r ^ values were generated using twenty executions of Algorithm A1, yielding the sequence r ^ 1 j , r ^ 2 j , , r ^ 20 j ), j = 1 , 2 of values for r ^ . The parameterization is given in Table 5 for Algorithms 1 and A1. We then collected the ME and UM values, obtained using r ^ i j , i = 1 , , 20 , j = 1 , 2 and plotted the corresponding curves against r ^ . This experiment aims to verify the influence of the stochastic process on the output.
Figure 15 and Figure 16 show the plots of ME versus r ^ for Algorithms 1 and A1, respectively. Firstly, the range for r ^ for Algorithm 1 is [ 0.9 , 1.26 ] while the range for the other algorithm is [ 1.13 , 1.53 ] . Although both algorithms start with r ^ = 1.5 , their objective functions are different, which justifies the differences in the ranges.
In both cases, the ME values are very close to zero, if not reaching it. Algorithm 1, in this case, sets the ME value to 0 in 4 instances, within the range [ 0.9968 , 0.9975 ] of r ^ . On the other hand, Algorithm A1 sets the ME value to 0 in only 2 instances, in the range [ 1.230 , 1.232 ] of r ^ . In other words, the simulated annealing method leads to a higher convergence rate for Algorithm 1 with a slight advantage over Algorithm A1.
Note that the values obtained for r ^ in the original experiment (see Figure 6) are in the ranges provided by simulated annealing, as observed in Figure 15. It is important to note that more than one ME value was set to 0 in both cases. This demonstrates the existence of several local minima for ME, i.e., several local maxima for functions in expressions (35) and (A1). In this way, the experiments of Section 6.1 are validated by the latter (Algorithms 1 and A1).
Regarding the UM values, these are concentrated in values very close to 1 for both Algorithms 1 and A1. This corroborates all the observations made for the ME measure made above.

6.5. Performance on BSDS500 Images

This section has a two-fold proposal. Firstly, demonstrates the applicability of our proposal out of the context of IR and NDT images. Secondly, compare the performance of Algorithm 1 against counterpart methods based on entropy concepts, as well as the support vector machine (SVM) algorithm, which is a known machine learning approach.
The comparison with entropy-based methods is executed by considering the competing methods reported in the first column of Table 7: Albuquerque [3], based on the Tsallis entropy; Nie et al. [6], based on Masi entropy; Lin & Ou [5], based on Shannon and Tsallis entropies; Deng [37], based on Tsalis entropy; Manda & Hyun [36], based on Sine entropy; Elaraby & Moratal [35], based on Hill entropy; Proposals 1 and 2 of [38], based on Tsallis and Masi entropies. The performance of each technique is reported in Table 7. It is achieved by computing the mean and standard deviation (Std) of the ME over the segmentations of the set BSDS500-50, described in the introduction of Section 6. The BSDS500-50 images were converted to grayscale format since the considered methods do not work with RGB data. We observe that our approach is the most outstanding one, as its mean and standard deviation are the smallest. In all these experiments the parmeterization of Algorithm 1 is specified in Table 5.
Now, we compare Algorithm 1’s efficiency with the segmentation obtained by a SVM approach. The SVM is a machine learning technique that, in our application, seeks the classification hypersurface that maximizes the separation margin between foreground and background pixels. Hence, our SVM approach works in the pixel space. Consequently, we must separate two disjoint subsets: the training one ( D t r ), which in this case is composed of a pool of pixels used to compute the SVM hypersurface, and the test set ( D t e ) to evaluate the segmentation (inference) obtained.
As usual in the machine learning area, the selection of the D t r is critical to obtaining a good performance over the test set D t e . Since the BSDS500-50 images present different contents and characteristics, as observed in Figure 2, selecting images to compose the pixels in the training set is far from trivial. Hence, to offer a meaningful comparison between our approach and SVM, we randomly select the image 12074.jpg, shown in Figure 17, to generate the D t r . Hence, we generate a list of training points D t r = R i , G i , B i ; l i , 1 i N t r where the first three components are the color channels of each pixel, l i = 1 for the foreground and l i = 1 for the background, and N t r = 481 · 321 = 154401 . The SVM used is the non-separable one, with constant C = 1 and the RBF kernel.
Then, we perform inference over the D t e set, but processing one image at a time. Next, we selected the test images for which the SVM achieved an ME larger than 0.3 % and performed its segmentation with Algorithm 1 for comparison. Figure 17 pictures the training image and the test ones. The ME values for both SVM and Algorithm 1 over selected images can be analyzed in Figure 18.
Figure 18 shows that in most cases, our method outperforms that of SVM, except in four instances: 80099, 163062, 271008, and 299091. Obviously, if we change the D t r or the SVM parametrization, or use a more powerful machine learning method (see Section 7), we can obtain a different result. Despite this, the results presented in Figure 18 indicate that Algorithm 1 is competitive with respect to a traditional machine learning approach.

7. Discussion

The long-range correlations of an image can be captured by the Tsallis and Masi entropies. In particular, IR and NDT images show local long-range correlation instead of global long-range correlation. In this work, the image segmentation technique based on Tsallis entropy and long-range correlation proposed by Lin & Ou was improved. The combination of Masi and Shannon entropies not only outperforms that method but has also shown itself to be competitive against works that apply only Masi or only Tsallis models. Misclassification error, Jaccard similarity, misclassified pixels, relative foreground area error, and F—measure have proven the effectiveness of our method.
Regarding the comparison between Algorithm 1 and deep neural networks for image segmentation, like U-Net and others [65], some points must be contemplated:
  • The proposed approach does not rely on training stages. It is advantageous to avoid the computational complexity of such stages, which is a significant challenge for deep learning models.
  • The training process of machine learning models allows the incorporation of prior information into the segmentation model. It is an advantage of that stage despite its computational cost. In the case of Algorithm 1, given a dataset D composed of images with similar contents, we can take an image I from D, generate its ground truth I g t , and apply Algorithm 1 with I and I g t as input. The returned Masi parameter r ^ can be used as an initial guess of r for the remaining images from D. This strategy simulates a training process and could reduce the complexity of Algorithm 1, considering the low sensitivity of our proposal with respect to r verified in Section 6.3.
  • The comparison between SVM and Algorithm 1 shows that the latter is competitive regarding the former. The machine learning performance could be improved by using a deep neural network with sufficient degrees of freedom. However, considering the limited number of annotated images in the BSDS500-50 dataset, we should fine-tune a pre-trained neural network, a process that also has its drawbacks [66].
  • The usual strategy for deep learning segmentation works in the image space; that means the model input is an image I instead of the pixel channels R , G , B as performed in Section 6.5. However, in this case, it is not expected that BSDS500-50 will be enough for training, considering the diversity of patterns observed in Figure 2 and Figure 17. In fact, we need a large training dataset D t r even when using pre-trained neural networks. In summary, the performance verified in Section 6.5 for Algorithm 1 indicates that it is a suitable choice for applications with limited training data.
  • Besides issues involving the training stage, deep neural networks have also several parameters to be set in advance (hyperparameters). There is no systematic theory for hyperparameter optimization. The state-of-the-art approaches, like Bayesian, evolutionary algorithms, and grid search [67], depend on computationally intensive tests, which is another issue for deep learning approaches. According to Table 5, the hyperparameters space of Algorithm 1 has dimension 9, since L and f i x _ p a r are defined by the application requirements. Such a dimension is much smaller compared with deep learning approaches.
With respect to issues related to our proposal, we shall return to Algorithm 1 and notice that there is no guarantee that the criterion in line 24 (or 30) will eventually be satisfied. In such a case, the parameterization of the algorithm must be reconsidered, for instance, by changing the initial guess r ^ , increasing the computational cost. The incorporation of prior knowledge in the segmentation process, a natural procedure for machine learning approaches during training, is not considered in Algorithm 1, which is a disadvantage compared to deep learning models, for instance, despite the computational cost to train the model. In addition, the proposed method deals only with bi-level thresholding. Its extension to the multilevel case will be addressed in further work.   

Author Contributions

Conceptualization and Methodology: P.E.F.J. and V.M.M. Original draft preparation: G.A.G. Conceptualization; Methodology; Software; Validation: E.P.S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by INCT-MACC (National Institute of Science and Technology), the Brazilian agency CNPq, and National Laboratory for Scientific Computing.

Informed Consent Statement

Not applicable.

Data Availability Statement

Image data used in this research can be found at https://drive.google.com/file/d/1LzQNCjK4YeTzqGf-hYQ7Q_osi2Ery9m8/view?usp=sharing (accessed on 15 September 2025).

Acknowledgments

The authors would like to thank Congjie Ou for supplying your images and their ground-truth. Thanks are also extended to ZuoYong Li for supplying several images and the ground-truth of them.

Conflicts of Interest

Authors declare that they have no conflicts of interest.

Abbreviations

    The following abbreviations are used in this manuscript:
F α F-measure
IRInfrared Images
JSJaccard Similarity
MEMisclassification Error
NDTNondestructive Images
PDEsPartial Differential Equations
RAERelative Foreground Area Error

Appendix A. Masi Thresholding Algorithm

This section presents a variation of Algorithm 1, obtained by using just the Masi entropy, aiming to allow ablation analysis of Algorithm 1. The new procedure incorporates the same SA methodology of Algorithm 1 and the possibility of using it just to compute the threshold t o p t given a fixed r value (set f i x _ p a r = 1 ).
Algorithm A1 Pseudo-code for Masi thresholding based on SA.
1:
Input: grayscale image I with N pixels; image intensity range: 0 , L ; T i n i t i a l T m i n ; α ; m a r k o v l e n g t h ; m a x i t e r ; groud truth segmentation I g t ; Reference error E r e f ; Reference uniformity measure U M r e f ; d; f i x _ p a r 0 , 1 ;
2:
Calculate image histogram h = h t ;
3:
Initialization of entropic parameter: r ^ ;
4:
T 0 T i n i t i a l ;
5:
Objective function:
F 3 t , r ^ = S r ^ A + B ( t ) ; 0 t L ;
6:
Solve the optimization problem:
t o p t r ^ = arg max 0 t L F 3 t , r ^ ;
7:
Binarization of input image: B i n a r y r ^ = T h r e s h o l d i n g I , t o p t r ^ ;
8:
if  f i x _ p a r = = 0  then
9:
    go to line 41;
10:
end if
11:
if ground truth I g t is available then
12:
    Calculate Misclassification Error: E 0 = E r r o r I g t , B i n a r y r ^ ;
13:
    If E 0 E r e f then go to Line 41;
14:
else
15:
    Calculate U M ( j ) = U M ( t o p t r ^ ) ;
16:
    If U M j U M r e f then go to Line 41;
17:
end if
18:
i 0 ;
19:
while  T i > T m i n and i < M a x i t e r  do
20:
    for  j = 1 , j + + , while j < m a r k o v l e n g t h ; do
21:
        Update r j r ^ + r a n d d , d ;
22:
        Solve the optimization problem:
t o p t r j = arg max 0 t L F 3 t , r j ;
23:
        Binarization of input image: B i n a r y r j = T h r e s h o l d i n g I , t o p t r j ;
24:
        if ground truth I g t is available then
25:
           Calculate Misclassification Error: E j = E r r o r I g t , B i n a r y r j ;
26:
           If E j E r e f then go to Line 41;
27:
           Calculate difference: Δ E = E j E j 1 ;
28:
           Acceptance criterion: If Δ E 0 then r ^ r j else p r o b = exp Δ E / T ;
29:
           If Δ E > 0 and r a n d 0 , 1 < p r o b then r ^ r j ;
30:
        else
31:
           Calculate U M ( j ) = U M ( t o p t r j ) ;
32:
           If U M j U M r e f then go to Line 41;
33:
           Calculate difference: Δ U M = U M j U M j 1 ;
34:
           Acceptance criterion: If Δ U M 0 then r ^ r j else p r o b = exp Δ U M / T ;
35:
           If Δ U M < 0 and r a n d 0 , 1 < p r o b then r ^ r j ;
36:
        end if
37:
    end for
38:
     i i + 1 ;
39:
     T i T i 1 · α ;
40:
end while
41:
Output: Threshold t o p t and parameter r ^ .

References

  1. Sezgin, M.; Sankur, B. Survey over image thresholding techniques and quantitative performance evaluation. J. Eletronic Imaging 2004, 13, 146–165. [Google Scholar]
  2. Pun, T. Entropic Thresholding: A New Approach. Comput. Graph. Image Process. 1981, 16, 210–239. [Google Scholar] [CrossRef]
  3. Albuquerque, M.P.; Esquef, I.A.; Gesualdi Mello, A.R.; Albuquerque, M.P. Image Thresholding Using Tsallis Entropy. Pattern Recognit. Lett. 2004, 25, 1059–1065. [Google Scholar] [CrossRef]
  4. Lin, A.; Wu, L.; Zheng, B.; Zan, H. Self-Adaptive Parameter Selection in One-dimensional Tsallis Entropy Thresholding with PARTICLE Swarm Optimization Algorithm. In Proceedings of the 3rd International Congress on Image and Signal Processing—CISP, Yantai, China, 16–18 October 2010; pp. 1460–1464. [Google Scholar]
  5. Lin, Q.; Ou, C. Tsallis Entropy and the Long-range Correlation in Image Thresholding. Signal Process. 2012, 92, 2931–2939. [Google Scholar] [CrossRef]
  6. Nie, F.; Zhang, P.; Li, J.; Ding, D. A novel generalized entropy and its application in image thresholding. Signal Process. 2017, 134, 23–34. [Google Scholar] [CrossRef]
  7. Li, Z.; Liu, C.; Liu, G.; Yang, X.; Cheng, Y. Statistical Thresholding Method for Infrared Images. Pattern Anal. Appl. 2011, 14, 109–126. [Google Scholar] [CrossRef]
  8. Feng, D.; Wenkang, S.; Liangzhou, C.; Yong, D.; Zhenfu, Z. Infrared Image Segmentation with 2-D Maximum Entropy Method Based on PARTICLE Swarm Optimization (PSO). Pattern Recognit. Lett. 2005, 26, 597–603. [Google Scholar] [CrossRef]
  9. Sparavigna, A.C. Tsallis Entropy in Bi-level and Multi-Level Image Thresholding. Int. J. Sci. 2015, 4, 40–49. [Google Scholar] [CrossRef]
  10. Khairuzzaman, A.K.; Chaudhury, S. Masi Entropy Based Multilevel Thresholding for Image Segmentation. Multimed. Tools Appl. 2019, 78, 33573–33591. [Google Scholar] [CrossRef]
  11. Shubham, S.; Bhandari, A.K. A Generalized Masi Entropy Based Efficient Multilevel Thresholding Method for Color Image Segmentation. Multimed. Tools Appl. 2019, 78, 17197–17238. [Google Scholar] [CrossRef]
  12. Ishak, A.B. A two-dimensional multilevel thresholding method for image segmentation. Appl. Soft Comput. 2017, 52, 306–322. [Google Scholar] [CrossRef]
  13. Ishak, A.B. Choosing parameters for Rényi and Tsallis entropies within a two-dimensional multilevel image segmentation framework. Phys. A 2017, 466, 521–536. [Google Scholar] [CrossRef]
  14. Zhao, X.; Turk, M.; Li, W.; Lien, K.-C.; Wanga, G. A multilevel image thresholding segmentation algorithm based on two-dimensional K–L divergence and modified pARTICLE swarm. Appl. Soft Comput. 2016, 48, 151–159. [Google Scholar] [CrossRef]
  15. Sathya, P.D.; Kayalvizhi, R. PSO-Based Tsallis Thresholding Selection Procedure for Image Segmentation. Int. J. Comput. Appl. 2010, 5, 39–46. [Google Scholar] [CrossRef]
  16. Wunnava, A.; Naik, M.K.; Panda, R.; Jena, B.; Abraham, A. A novel interdependence based multilevel thresholding technique using adaptive equilibrium optimizer. Eng. Appl. Artif. Intell. 2020, 94, 103836. [Google Scholar] [CrossRef]
  17. Kapur, J.N.; Sahoo, P.K.; Wong, A.K.C. A new method for gray-level picture thresholding using the entropy of the histogram. Comput. Vis. Graph. Image Process. 1985, 29, 273–285. [Google Scholar] [CrossRef]
  18. Sahoo, P.K.; Wilkins, C.; Yeager, J. Thresholding Selection Using Rényi’s Entropy. Pattern Recognit. 1997, 30, 71–84. [Google Scholar] [CrossRef]
  19. Sparavigna, A.C. Shannon, Tsallis and Kaniadakis entropies in bi-level image thresholding. Int. J. Sci. 2015, 4, 35–43. [Google Scholar] [CrossRef]
  20. Zhou, Y.; Gao, M.; Fang, D.; Zhang, B. Tank segmentation of infrared images with complex background for the homing anti-tank missile. Infrared Phys. Technol. 2016, 77, 258–266. [Google Scholar] [CrossRef]
  21. Zhou, Y.; Gao, M.; Fang, D.; Zhang, B. Unsupervised Background-Constrained Tank Segmentation of Infrared Images in Complex Background Based on the Otsu Method; SpringerPlus: Berlin, Germany, 2016; Volume 5. [Google Scholar]
  22. Yu, X.; Zhou, Z.; Gao, Q.; Li, D.; Ríha, K. Infrared image segmentation using growing immune field and clone threshold. Infrared Phys. Technol. 2018, 88, 184–193. [Google Scholar] [CrossRef]
  23. de Oliveira, J.P.S.; Conci, A.; Pérez, M.G.; Andaluz, V.H. Segmentation of infrared images: A new technology for early detection of breast diseases. In Proceedings of the 2015 IEEE International Conference on Industrial Technology (ICIT), Seville, Spain, 17–19 March 2015; pp. 1765–1771. [Google Scholar] [CrossRef]
  24. Fan, S.; Yang, S.; He, P.; Nie, H. Infrared Electric Image Thresholding Using Two-Dimensional Fuzzy Renyi Entropy. Energy Procedia 2011, 12, 411–419. [Google Scholar] [CrossRef]
  25. Liu, Q.; He, Z.; Li, X.; Zheng, Y. PTB-TIR: A Thermal Infrared Pedestrian Tracking Benchmark. IEEE Trans. Multimed. 2020, 22, 666–675. [Google Scholar] [CrossRef]
  26. Masi, M. A step beyond Tsallis and Rényi entropies. Phys. Lett. A 2005, 338, 217–224. [Google Scholar] [CrossRef]
  27. Lei, B.; Fan, J. Image thresholding segmentation method based on minimum square rough entropy. Appl. Soft Comput. 2019, 84, 105687. [Google Scholar] [CrossRef]
  28. Lei, B.; Fan, J. Adaptive granulation Renyi rough entropy image thresholding method with nested optimization. Expert Syst. Appl. 2022, 203, 117378. [Google Scholar] [CrossRef]
  29. Jumiawi, W.A.H.; El-Zaart, A. Gumbel (EVI)-Based Minimum Cross-Entropy Thresholding for the Segmentation of Images with Skewed Histograms. Appl. Syst. Innov. 2023, 6, 87. [Google Scholar] [CrossRef]
  30. Zou, Y.; Zhang, J.; Upadhyay, M.; Sun, S.; Jiang, T. Automatic Image Thresholding Based on Shannon Entropy Difference and Dynamic Synergic Entropy. IEEE Access 2020, 8, 171218–171239. [Google Scholar] [CrossRef]
  31. Wang, S.; Fan, J. Image Thresholding Method Based on Tsallis Entropy Correlation. Multimed. Tools Appl. 2025, 84, 9749–9785. [Google Scholar] [CrossRef]
  32. Hegazy, E.; Gabr, M. Enhancng Image Thresholding with Masi Entropy: An Empirical Approach to Parameter Selection. In Proceedings of the 33rd International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision’2025 (WSCG’2025), Pilsen, Czech Republic, 26–29 May 2025; Volume 3501. [Google Scholar] [CrossRef]
  33. Rodrigues, P.S.; Giraldi, G.A. Computing q-index for Tsallis Nonextensive Image Segmentation. In Proceedings of the XXII Brazilian Symposium on Computer Graphics and Image Processing—SIBGRAPI, Rio de Janeiro, Brazil, 11–15 October 2009; pp. 232–237. [Google Scholar]
  34. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics. In Proceedings of the 8th International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  35. Elaraby, A.; Moratal, D. A generalized entropy-based two-phase threshold algorithm for noisy medical image edge detection. Sci. Iran. 2017, 24, 3247–3256. [Google Scholar] [CrossRef]
  36. Manda, M.P.; Hyun, D.K. Double Thresholding with Sine Entropy for Thermal Image Segmentation. Trait. Du Signal 2021, 38, 1713–1718. [Google Scholar] [CrossRef]
  37. Deng, Q.; Shi, Z.; Ou, C. Self-Adaptive Image Thresholding within Nonextensive Entropy and the Variance of the Gray-Level Distribution. Entropy 2022, 24, 319. [Google Scholar] [CrossRef]
  38. Ferreira Júnior, P.E.; Mello, V.M.; Giraldi, G.A. Image thresholding through nonextensive entropies and long-range correlation. Multimed. Tools Appl. 2023, 82, 43029–43073. [Google Scholar] [CrossRef]
  39. Khan, M. A Survey: Image Segmentation Techniques. Int. J. Future Comput. Commun. 2014, 3, 89–93. [Google Scholar] [CrossRef]
  40. Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef] [PubMed]
  41. Rayed, M.E.; Islam, S.S.; Niha, S.I.; Jim, J.R.; Kabir, M.M.; Mridha, M. Deep learning for medical image segmentation: State-of-the-art advancements and challenges. Inform. Med. Unlocked 2024, 47, 101504. [Google Scholar] [CrossRef]
  42. Rahman, M.M.; Rahman, S.; Bhatt, S.; Faezipour, M. A Systematic Review on Advancement of Image Segmentation Techniques for Fault Detection Opportunities and Challenges. Electronics 2025, 14, 974. [Google Scholar] [CrossRef]
  43. Garcia-Garcia, A.; Orts, S.; Oprea, S.; Villena-Martinez, V.; Martinez-Gonzalez, P.; Rodríguez, J.G. A survey on deep learning techniques for image and video semantic segmentation. Appl. Soft Comput. 2018, 70, 41–65. [Google Scholar] [CrossRef]
  44. Samantaray, L.; Hembram, S.; Panda, R. A New Harris Hawks-Cuckoo Search Optimizer for Multilevel Thresholding of Thermogram Images. Rev. d’Intelligence Artif. 2020, 34, 541–551. [Google Scholar] [CrossRef]
  45. Wachs-Lopes, G.A.; Santos, R.M.; Saito, N.; Rodrigues, P.S.S. Recent nature-Inspired algorithms for medical image segmentation based on tsallis statistics. Commun. Nonlinear Sci. Numer. Simul. 2020, 88, 105256. [Google Scholar] [CrossRef]
  46. Babu, A.; Rajam, V. Water-body segmentation from satellite images using Kapur’s entropy-based thresholding method. Comput. Intell. 2020, 36, 1242–1260. [Google Scholar] [CrossRef]
  47. Xing, Z. An improved emperor penguin optimization based multilevel thresholding for color image segmentation. Knowl. Based Syst. 2020, 194, 105570. [Google Scholar] [CrossRef]
  48. Sukanthi; Murugan, S.; Hanis, S. Binarization of Stone Inscription Images by Modified Bi-level Entropy Thresholding. Fluct. Noise Lett. 2021, 20, 2150054. [Google Scholar] [CrossRef]
  49. Alexander, T.; Kumar, S. A novel binarization technique based on Whale Optimization Algorithm for better restoration of palm leaf manuscript. J. Ambient. Intell. Humaniz. Comput. 2020, 1–8. [Google Scholar] [CrossRef]
  50. Mahmoudi, L.; Zaart, A.E. A survey of entropy image thresholding techniques. In Proceedings of the 2012 2nd International Conference on Advances in Computational Tools for Engineering Applications (ACTEA), Beirut, Lebanon, 12–15 December 2012; pp. 204–209. [Google Scholar]
  51. Aczél, J.; Daróczy, Z. (Eds.) Rényi Entropies. In On Measures of Information and their Characterizations; Elsevier: Amsterdam, The Netherlands, 1975; Volume 115, pp. 134–172. [Google Scholar] [CrossRef]
  52. Tsallis, C. Possible Generalizations of Boltzmann-Gibbs Statistics. J. Stat. Phys. 1988, 52, 480–487. [Google Scholar] [CrossRef]
  53. Tsallis, C. Nonextensive statistics: Theoretical, experimental and computational evidences and connections. Braz. J. Phys. 1999, 29, 1–35. [Google Scholar] [CrossRef]
  54. Xu, Y.; Chen, R.; Li, Y.; Zhang, P.; Yang, J.; Zhao, X.; Liu, M.; Wu, D. Multispectral Image Segmentation Based on a Fuzzy Clustering Algorithm Combined with Tsallis Entropy and a Gaussian Mixture Model. Remote Sens. 2019, 11, 2772. [Google Scholar] [CrossRef]
  55. Shannon, C. A mathematical theory of communication. Bell Syst. Echcnical J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  56. Lei, B.; Fan, J. Adaptive Kaniadakis entropy thresholding segmentation algorithm based on particle swarm optimization. Soft Comput. 2020, 24, 7305–7318. [Google Scholar] [CrossRef]
  57. Wunnava, A.; Naik, M.; Panda, R.; Jena, B.; Abraham, A. A differential evolutionary adaptive Harris hawks optimization for two dimensional practical Masi entropy-based multilevel image thresholding. J. King Saud Univ.—Comput. Inf. Sci. 2020, 34, 3011–3024. [Google Scholar] [CrossRef]
  58. Grady, L.; Schwartz, E. Isoperimetric graph partitioning for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 469–475. [Google Scholar] [CrossRef]
  59. Cáceres, M.O. Non-Markovian processes with long-range correlations: Fractal dimension analysis. Braz. J. Phys. 1999, 29, 125–135. [Google Scholar] [CrossRef]
  60. Wang, H.; Zhang, H.; Ray, N. Adaptive shape prior in graph cut image segmentation. Pattern Recognit. 2013, 46, 1409–1414. [Google Scholar] [CrossRef]
  61. Lei, B.; Fan, J. Infrared pedestrian segmentation algorithm based on the two-dimensional Kaniadakis entropy thresholding. Knowl.-Based Syst. 2021, 225, 1–12. [Google Scholar] [CrossRef]
  62. Liu, T.; Sun, J.; Zheng, N.; Tang, X.; Shum, H. Learning to Detect A Salient Object. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 33, 353–367. [Google Scholar]
  63. Martin, D.R.; Fowlkes, C.C.; Malik, J. Learning to Detect Natural Image Boundaries Using Local Brightness, Color, and Texture Cues. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 530–549. [Google Scholar] [CrossRef] [PubMed]
  64. Lei, B.; Li, J.; Wang, N.; Yu, H. An efficient adaptive Masi entropy multilevel thresholding algorithm based on dynamic programming. J. Vis. Commun. Image Represent. 2024, 98, 104008. [Google Scholar] [CrossRef]
  65. Wang, G.; Li, Z.; Weng, G.; Chen, Y. An overview of industrial image segmentation using deep learning models. Intell. Robot. 2025, 5, 143–180. [Google Scholar] [CrossRef]
  66. Sánchez-Medina, M.A.; Batres-Mendoza, P.; Solís-Arrazola, M.A.; Guerra-Hernández, E.I.; Sotelo-Figueroa, M.A.; Espinal, A. A Comparative Study of Transfer Learning and Fine-Tuning for Facial Paralysis Image Classification. In Modern Artificial Intelligence Based on Soft Computing Techniques; Springer: Berlin, Germany, 2025; pp. 89–101. [Google Scholar]
  67. Ibrahim, M.; Harkouss, F.; Biwole, P.; Fardoun, F.; Ouldboukhitine, S.E. Multi-objective hyperparameter optimization of artificial neural network in emulating building energy simulation. Energy Build. 2025, 337, 115643. [Google Scholar] [CrossRef]
Figure 1. Plots of functions (22) and (23) showing short-range correlation in green and long-range correlation in red, respectively (Source [38]).
Figure 1. Plots of functions (22) and (23) showing short-range correlation in green and long-range correlation in red, respectively (Source [38]).
Entropy 27 01203 g001
Figure 2. Some images of BSDS500-50 dataset with ground truth.
Figure 2. Some images of BSDS500-50 dataset with ground truth.
Entropy 27 01203 g002
Figure 3. Thresholding results of 000280 image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 91 ), (e) Albuquerque’s method (entropic parameter q = 0.8 and T = 91 ), (f) Lin & Ou’s method ( q = 0.8 and threshold threshold T = 169 ), (g) Nie et al.’s method (entropic parameter r = 1.35 and T = 173 ), and (h) the proposed method ( r = 1.21 and threshold T = 170 ).
Figure 3. Thresholding results of 000280 image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 91 ), (e) Albuquerque’s method (entropic parameter q = 0.8 and T = 91 ), (f) Lin & Ou’s method ( q = 0.8 and threshold threshold T = 169 ), (g) Nie et al.’s method (entropic parameter r = 1.35 and T = 173 ), and (h) the proposed method ( r = 1.21 and threshold T = 170 ).
Entropy 27 01203 g003
Figure 4. Thresholding results of Airplane image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 117 ), (e) Albuquerque’s method (entropic parameter q = 0.8 and threshold value T = 117 ), (f) Lin & Ou’s method ( q = 0.8 and T = 117 ), (g) Nie et al. method (entropic parameter r = 1.22 and threshold value T = 121 ), and (h) the proposed method ( r = 0.797 and T = 123 ).
Figure 4. Thresholding results of Airplane image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 117 ), (e) Albuquerque’s method (entropic parameter q = 0.8 and threshold value T = 117 ), (f) Lin & Ou’s method ( q = 0.8 and T = 117 ), (g) Nie et al. method (entropic parameter r = 1.22 and threshold value T = 121 ), and (h) the proposed method ( r = 0.797 and T = 123 ).
Entropy 27 01203 g004
Figure 5. Thresholding results of Tank image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 140 ), (e) Albuquerque’s method (entropic parameter q = 0.9 and threshold value T = 140 ), (f) Lin & Ou’s method ( q = 0.9 and T = 220 ), (g) Nie et al.’s method (entropic parameter r = 1.25 and T = 194 ), and (h) the proposed method ( r = 0.99 and T = 191 ).
Figure 5. Thresholding results of Tank image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 140 ), (e) Albuquerque’s method (entropic parameter q = 0.9 and threshold value T = 140 ), (f) Lin & Ou’s method ( q = 0.9 and T = 220 ), (g) Nie et al.’s method (entropic parameter r = 1.25 and T = 194 ), and (h) the proposed method ( r = 0.99 and T = 191 ).
Entropy 27 01203 g005
Figure 6. Thresholding results of Panzer image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 154 ), (e) Albuquerque’s method (entropic parameter q = 0.1 and threshold value T = 157 ), (f) Lin & Ou’s method ( q = 0.1 and T = 153 ), (g) Nie et al.’s method (entropic parameter r = 1.231 and T = 174 ), and (h) the proposed method ( r = 0.997 and T = 174 ).
Figure 6. Thresholding results of Panzer image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 154 ), (e) Albuquerque’s method (entropic parameter q = 0.1 and threshold value T = 157 ), (f) Lin & Ou’s method ( q = 0.1 and T = 153 ), (g) Nie et al.’s method (entropic parameter r = 1.231 and T = 174 ), and (h) the proposed method ( r = 0.997 and T = 174 ).
Entropy 27 01203 g006
Figure 7. Thresholding results of Car image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method ( T = 81 ), (e) Albuquerque’s method (entropic parameter q = 0.9 and threshold value T = 81 ), (f) Lin & Ou’s method ( q = 0.9 and T = 80 ), (g) Nie et al.’s method (entropic parameter r = 1.28 and T = 89 ), and (h) the proposed method ( r = 0.95 and T = 90 ).
Figure 7. Thresholding results of Car image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method ( T = 81 ), (e) Albuquerque’s method (entropic parameter q = 0.9 and threshold value T = 81 ), (f) Lin & Ou’s method ( q = 0.9 and T = 80 ), (g) Nie et al.’s method (entropic parameter r = 1.28 and T = 89 ), and (h) the proposed method ( r = 0.95 and T = 90 ).
Entropy 27 01203 g007
Figure 8. Thresholding results of Sailboat image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method ( threshold value T = 155 ), (e) Albuquerque’s method (entropic parameter q = 0.6 and T = 155 ), (f) Lin & Ou’s method ( q = 0.6 and T = 141 ), (g) Nie et al.’s method (entropic parameter r = 0.9 and T = 155 ), and (h) the proposed method ( r = 1.03 and T = 193 ).
Figure 8. Thresholding results of Sailboat image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method ( threshold value T = 155 ), (e) Albuquerque’s method (entropic parameter q = 0.6 and T = 155 ), (f) Lin & Ou’s method ( q = 0.6 and T = 141 ), (g) Nie et al.’s method (entropic parameter r = 0.9 and T = 155 ), and (h) the proposed method ( r = 1.03 and T = 193 ).
Entropy 27 01203 g008
Figure 9. Infrared image sequence and their thresholding results: (a) original, (b) ground-truth image, (c) Albuquerque’s method (entropic parameter q = 0.8 ), the thresholds from top to bottom: 91, 93, 91, 97, 90, 98, 93, 96, 94, 96, (d) Lin & Ou’s method (entropic parameter q = 0.8 ), the thresholds from top to bottom: 169, 178, 185, 177, 179, 178, 179, 173, 175, 192, (e) Nie et al.’s method, the parameters r and thresholds from the top to bottom: ( 1.35 , 173 ) , ( 1.341 , 177 ) , ( 1.31 , 171 ) , ( 1.34 , 176 ) , ( 1.334 , 175 ) , ( 1.32 , 170 ) , ( 1.334 , 177 ) , ( 1.35 , 176 ) , ( 1.35 , 178 ) , ( 1.32 , 182 ) , (f) the proposed method, the parameters r and thresholds from top to bottom: ( 1.21 , 170 ) , ( 1.19 , 174 ) , ( 1.15 , 170 ) , ( 1.205 , 177 ) , ( 1.192 , 175 ) , ( 1.19 , 174 ) , ( 1.19 , 177 ) , ( 1.21 , 175 ) , ( 1.22 , 178 ) , ( 1.17 , 181 ) .
Figure 9. Infrared image sequence and their thresholding results: (a) original, (b) ground-truth image, (c) Albuquerque’s method (entropic parameter q = 0.8 ), the thresholds from top to bottom: 91, 93, 91, 97, 90, 98, 93, 96, 94, 96, (d) Lin & Ou’s method (entropic parameter q = 0.8 ), the thresholds from top to bottom: 169, 178, 185, 177, 179, 178, 179, 173, 175, 192, (e) Nie et al.’s method, the parameters r and thresholds from the top to bottom: ( 1.35 , 173 ) , ( 1.341 , 177 ) , ( 1.31 , 171 ) , ( 1.34 , 176 ) , ( 1.334 , 175 ) , ( 1.32 , 170 ) , ( 1.334 , 177 ) , ( 1.35 , 176 ) , ( 1.35 , 178 ) , ( 1.32 , 182 ) , (f) the proposed method, the parameters r and thresholds from top to bottom: ( 1.21 , 170 ) , ( 1.19 , 174 ) , ( 1.15 , 170 ) , ( 1.205 , 177 ) , ( 1.192 , 175 ) , ( 1.19 , 174 ) , ( 1.19 , 177 ) , ( 1.21 , 175 ) , ( 1.22 , 178 ) , ( 1.17 , 181 ) .
Entropy 27 01203 g009
Figure 10. Thresholding results of Gear image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 185 ), (e) Albuquerque’s method (entropic parameter q = 0.1 and T = 128 ), (f) Lin & Ou’s method ( q = 0.6 and T = 87 ), (g) Nie et al.’s method (entropic parameter r = 1.22 and T = 75 ), and (h) the proposed method ( r = 1.7 and T = 80 ).
Figure 10. Thresholding results of Gear image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 185 ), (e) Albuquerque’s method (entropic parameter q = 0.1 and T = 128 ), (f) Lin & Ou’s method ( q = 0.6 and T = 87 ), (g) Nie et al.’s method (entropic parameter r = 1.22 and T = 75 ), and (h) the proposed method ( r = 1.7 and T = 80 ).
Entropy 27 01203 g010
Figure 11. Thresholding results of Pcb image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 158 ), (e) Albuquerque’s method (entropic parameter q = 0.8 and T = 158 ), (f) Lin & Ou’s method ( q = 0.8 and T = 78 ), (g) Nie et al.’s method (entropic parameter r = 1.297 and T = 82 ), and (h) the proposed method ( r = 0.84 and T = 99 ).
Figure 11. Thresholding results of Pcb image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 158 ), (e) Albuquerque’s method (entropic parameter q = 0.8 and T = 158 ), (f) Lin & Ou’s method ( q = 0.8 and T = 78 ), (g) Nie et al.’s method (entropic parameter r = 1.297 and T = 82 ), and (h) the proposed method ( r = 0.84 and T = 99 ).
Entropy 27 01203 g011
Figure 12. Thresholding results of Cell image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 172 ), (e) Albuquerque’s method (entropic parameter q = 0.7 and T = 171 ), (f) Lin & Ou’s method ( q = 0.7 and T = 214 ), (g) Nie et al.’s method (entropic parameter r = 1.2 and T = 222 ), and (h) the proposed method ( r = 1.25 and T = 213 ).
Figure 12. Thresholding results of Cell image: (a) original, (b) Histogram, (c) ground-truth image, (d) Shannon method (threshold value T = 172 ), (e) Albuquerque’s method (entropic parameter q = 0.7 and T = 171 ), (f) Lin & Ou’s method ( q = 0.7 and T = 214 ), (g) Nie et al.’s method (entropic parameter r = 1.2 and T = 222 ), and (h) the proposed method ( r = 1.25 and T = 213 ).
Entropy 27 01203 g012
Figure 13. Results of convergence test: ME versus number of executions.
Figure 13. Results of convergence test: ME versus number of executions.
Entropy 27 01203 g013
Figure 14. ME behavior against variance in noisy images for Algorithm 1.
Figure 14. ME behavior against variance in noisy images for Algorithm 1.
Entropy 27 01203 g014
Figure 15. ME and UM variation with respect to r ^ calculated through Algorithm 1.
Figure 15. ME and UM variation with respect to r ^ calculated through Algorithm 1.
Entropy 27 01203 g015
Figure 16. ME and UM variation with respect to r ^ calculated through Algorithm A1.
Figure 16. ME and UM variation with respect to r ^ calculated through Algorithm A1.
Entropy 27 01203 g016
Figure 17. Images used for SVM training (black rectangle) and inference (red rectangle).
Figure 17. Images used for SVM training (black rectangle) and inference (red rectangle).
Entropy 27 01203 g017
Figure 18. Comparison between Algorithm 1 and SVM results.
Figure 18. Comparison between Algorithm 1 and SVM results.
Entropy 27 01203 g018
Table 1. Comparison between the techniques. Ideal values: M E = 0 , J S = 100 % , R A E = 0 , F α = 1 . In black, the best results are highlighted.
Table 1. Comparison between the techniques. Ideal values: M E = 0 , J S = 100 % , R A E = 0 , F α = 1 . In black, the best results are highlighted.
ImagesThresholding Methods
Shannon Albuquerque Lin & Ou Nie et al. Our Method
000280Threshold9191169173 170
Misclassified pixels4900490066 4
ME 0.0638021 0.0638021 7.81 × 10 05 7.81 × 10 05 5 . 21 × 10 05
JS (%) 3.94 3.94 97.06 97.02 98 . 01
RAE 0.960596 0.960596 0 0.0298507 0 . 0199005
F α 0.057964 0.057964 0.985075 0.989848 0 . 993277
AirplaneThreshold117117117121 123
Misclassified pixels4444448 4
ME 0.000671387 0.000671387 0.000671387 0.00012207 6 . 10 × 10 05
JS (%) 93.05 93.05 93.05 98.66 99 . 32
RAE 0.0695103 0.0695103 0.0695103 0.0134003 0 . 00679117
F α 0.952561 0.952561 0.952561 0.991026 0 . 997726
TankThreshold140140220194 191
Misclassified pixels39039025521 9
ME 0.0133929 0.013929 0.00875687 0.000721154 0 . 000309066
JS (%) 59.29 59.29 55.11 96.30 98 . 44
RAE 0.407098 0.407098 0.448944 0.0369718 0 . 0155979
F α 0.68599 0.68599 0.786432 0.987365 0 . 989547
PanzerThreshold154157153 174 174
Misclassified pixels574426642 0 0
ME 0.00875854 0.00650024 0.00979614 0 0
JS (%) 75.23 80.36 73.08 100 . 00 100 . 00
RAE 0.247734 0.196404 0.269182 0 0
F α 0.819978 0.859891 0.802856 1 1
CarThreshold81818089 90
Misclassified pixels49249258155 11
ME 0.00750732 0.00750732 0.00886536 0.000839233 0 . 000167847
JS (%) 60.17 60.16 56.12 93.11 98 . 54
RAE 0.398381 0.398381 0.438822 0.0689223 0 . 0145889
F α 0.693744 0.693744 0.657328 0.952971 0 . 990227
SailboatThreshold155155141155 193
Misclassified pixels1295129524471295 98
ME 0.0197601 0.0197601 0.0373383 0.0197601 0 . 00149536
JS (%) 41.46 41.46 27.26 41.46 90 . 35
RAE 0.585443 0.585443 0.727408 0.585443 0 . 0965517
F α 0.515072 0.515072 0.359843 0.515072 0 . 933492
Table 2. The thresholding accuracy (JS, Equation (12)) of methods. Ideal value: JS = 100%. In black, the best results are highlighted.
Table 2. The thresholding accuracy (JS, Equation (12)) of methods. Ideal value: JS = 100%. In black, the best results are highlighted.
ImagesAlbuquerqueLin & OuNie et al.Our
(%) (%) (%) (%)
0002803.9497.0696.0297.51
0003004.2594.4794.0495.87
0003203.7590.0594.9194.93
0003404.8595.0294.5995.02
0003604.2691.2192.0892.08
0003805.9889.7890.3892.36
0004004.2398.20100.00100.00
0004203.9697.5198.9899.49
0004403.8793.6797.6197.61
0004605.1294.0596.8698.05
Average 4 . 42 94 . 10 95 . 55 96 . 29
Table 3. Comparison between the techniques concerning misclassified pixels and ME. Ideal value: M E = 0 . In black, the best results are highlighted.
Table 3. Comparison between the techniques concerning misclassified pixels and ME. Ideal value: M E = 0 . In black, the best results are highlighted.
ImagesThresholding Methods
Albuquerque Lin & Ou Nie et al. Our Method
000280Threshold91169173 170
Misclassified pixels490066 4
ME 0.0638021 7.81 × 10 05 7.81 × 10 05 5 . 21 × 10 05
000300Threshold93178177 174
Misclassified pixels48951213 9
ME 0.063737 0.00015625 0.000169271 0 . 000117188
000320Threshold91185171 170
Misclassified pixels541321 11 11
ME 0.0704818 0.000273437 0 . 000143229 0 . 000143229
000340Threshold97 177 176 177
Misclassified pixels4339 11 12 11
ME 0.0564974 0 . 000143229 0.00015625 0 . 000143229
000360Threshold90179 175 175
Misclassified pixels535521 19 19
ME 0.0697266 0.000273437 0 . 000247396 0 . 000247396
000380Threshold98178170 174
Misclassified pixels43092828 21
ME 0.0561068 0.000364583 0.000364583 0 . 000273437
000400Threshold98179 177 177
Misclassified pixels50334 0 0
ME 0.0655339 5.21 × 10 05 0 0
000420Threshold96173176 175
Misclassified pixels475652 1
ME 0.0619271 6.51 × 10 05 2.60 × 10 05 1 . 30 × 10 05
000440Threshold94175 178 178
Misclassified pixels514214 5 5
ME 0.0669531 0.000182292 6 . 51 × 10 05 6 . 51 × 10 05
000460Threshold96192182 181
Misclassified pixels4654158 5
ME 0.060599 0.000195313 0.000104167 6 . 51 × 10 05
Table 4. Comparison between the—NDT images. Ideal values: M E = 0 , J S = 100 % , R A E = 0 , F α = 1 . In black, the best results are highlighted.
Table 4. Comparison between the—NDT images. Ideal values: M E = 0 , J S = 100 % , R A E = 0 , F α = 1 . In black, the best results are highlighted.
ImagesThresholding Methods
Shannon Albuquerque Lin & Ou Nie et al. Our Method
GearThreshold185128877580
Misclassified pixels820015531881390
ME 0.125122 0.0236969 0.00286865 0.00212097 0
JS (%) 73.05 94.90 99.38 99.55 100.00
RAE 0.269506 0.0510419 0.00617893 0.00454769 0
F α 0.890489 0.982387 0.997932 0.996964 1
PcbThreshold158158788299
Misclassified pixels138521385228082166145
ME 0.211365 0.211365 0.0428467 0.0330505 0.00221252
JS (%) 36.01 36.01 88.52 90.90 99.33
RAE 0.639904 0.639904 0.114823 0.0909587 0.00669839
F α 0.628004 0.628004 0.920405 0.937465 0.997757
CellThreshold172171214222213
Misclassified pixels7276737527126340
ME 0.111023 0.112534 0.00413513 0.0401917 0
JS (%) 88.26 88.12 99.50 95.19 100.00
RAE 0.117376 0.118783 0.00495312 0.0481421 0
F α 0.918563 0.917547 0.998343 0.983421 1
Table 5. Input parameters used in the executions of Algorithm 1.
Table 5. Input parameters used in the executions of Algorithm 1.
ParameterValue
r ^ 1.5
d 0.1
L255
T i n i t i a l 100.0
T m i n 10 3
α 0.99
m a r k o v _ l e n g t h 20
m a x _ i t e r 100
E r e f 0.0
U M r e f 1.0
f i x _ p a r 0 or 1
Table 6. Sensitivity test of the r parameter—panzer image.
Table 6. Sensitivity test of the r parameter—panzer image.
kME ValuesSensitivity (S)
3 0.008758544921875 0.00823975
2 0.0035400390625 0.00302124
1 0.0013885498046875 0.00086975
0 0.000518798828125 0
1 0.0021820068359375 0.00166321
2 0.0035247802734375 0.00300598
3 0.0050811767578125 0.00456238
Table 7. Mean of ME values and standard deviations (Adapted from [38]).
Table 7. Mean of ME values and standard deviations (Adapted from [38]).
Thresholding MethodsME Mean: μ Std Deviat.: σ [ μ σ , μ + σ ]
Albuquerque [3]0.320.24[0.08, 0.56]
Nie et al. [6]0.310.25[0.06, 0.56]
Lin & Ou [5]0.380.25[0.13, 0.63]
Deng [37]0.490.26[0.23, 0.75]
Manda & Hyun [36]0.480.25[0.23, 0.73]
Elaraby & Moratal [35]0.270.19[0.08, 0.46]
Proposal 1 of [38]0.310.22[0.09, 0.53]
Proposal 2 of [38]0.440.26[0.18, 0.70]
Our Algorithm 1 0.005 0.0282 [−0.0232, 0.0332]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ferreira Júnior, P.E.; Mello, V.M.; Ribeiro, E.P.S.; Giraldi, G.A. A Masi-Entropy Image Thresholding Based on Long-Range Correlation. Entropy 2025, 27, 1203. https://doi.org/10.3390/e27121203

AMA Style

Ferreira Júnior PE, Mello VM, Ribeiro EPS, Giraldi GA. A Masi-Entropy Image Thresholding Based on Long-Range Correlation. Entropy. 2025; 27(12):1203. https://doi.org/10.3390/e27121203

Chicago/Turabian Style

Ferreira Júnior, Perfilino Eugênio, Vinícius Moreira Mello, Enzo P. Silva Ribeiro, and Gilson Antonio Giraldi. 2025. "A Masi-Entropy Image Thresholding Based on Long-Range Correlation" Entropy 27, no. 12: 1203. https://doi.org/10.3390/e27121203

APA Style

Ferreira Júnior, P. E., Mello, V. M., Ribeiro, E. P. S., & Giraldi, G. A. (2025). A Masi-Entropy Image Thresholding Based on Long-Range Correlation. Entropy, 27(12), 1203. https://doi.org/10.3390/e27121203

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop