Next Article in Journal
A Reversible Data Hiding Method in Encrypted Images for Controlling Trade-Off between Hiding Capacity and Compression Efficiency
Previous Article in Journal
Off-The-Grid Variational Sparse Spike Recovery: Methods and Algorithms
Previous Article in Special Issue
Tree Crowns Segmentation and Classification in Overlapping Orchards Based on Satellite Images and Unsupervised Learning Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation

by
Giacomo Aletti
,
Alessandro Benfenati
*,† and
Giovanni Naldi
Environmental Science and Policy Department, Università degli Studi di Milano, 20133 Milan, Italy
*
Author to whom correspondence should be addressed.
The three authors are members of the Italian Group “Gruppo Nazionale per il Calcolo Scientifico” of the Italian Institute “Istituto Nazionale di Alta Matematica” and of the ADAMSS (ADvanced Applied Mathematical and Statistical Sciences) Centre.
J. Imaging 2021, 7(12), 267; https://doi.org/10.3390/jimaging7120267
Submission received: 18 October 2021 / Revised: 26 November 2021 / Accepted: 1 December 2021 / Published: 7 December 2021
(This article belongs to the Special Issue Advances in Multi/Hyperspectral Imaging)

Abstract

:
The development of the hyperspectral remote sensor technology allows the acquisition of images with a very detailed spectral information for each pixel. Because of this, hyperspectral images (HSI) potentially possess larger capabilities in solving many scientific and practical problems in agriculture, biomedical, ecological, geological, hydrological studies. However, their analysis requires developing specialized and fast algorithms for data processing, due the high dimensionality of the data. In this work, we propose a new semi-supervised method for multilabel segmentation of HSI that combines a suitable linear discriminant analysis, a similarity index to compare different spectra, and a random walk based model with a direct label assignment. The user-marked regions are used for the projection of the original high-dimensional feature space to a lower dimensional space, such that the class separation is maximized. This allows to retain in an automatic way the most informative features, lightening the successive computational burden. The part of the random walk is related to a combinatorial Dirichlet problem involving a weighted graph, where the nodes are the projected pixel of the original HSI, and the positive weights depend on the distances between these nodes. We then assign to each pixel of the original image a probability quantifying the likelihood that the pixel (node) belongs to some subregion. The computation of the spectral distance involves both the coordinates in a features space of a pixel and of its neighbors. The final segmentation process is therefore reduced to a suitable optimization problem coupling the probabilities from the random walker computation, and the similarity with respect the initially labeled pixels. We discuss the properties of the new method with experimental results carried on benchmark images.

1. Introduction

Hyperspectral imaging systems have gained a great amount of attention from researchers in the past few years. The sensors of these systems allow the simultaneous acquisition of hundreds of spectral wavelengths for each image pixel. This detailed spectral information increases the possibility of more accurately discriminating objects, materials, or regions of interest. Furthermore, the fine spatial resolution of the sensors enables the analysis of small spatial structures in the image. The main property of the Hyperspectral images is the strong resolving power for fine spectra, then they have a wide range of applications in agriculture [1,2], food industry [3], geosciences [4,5], biomedical applications [6,7], document image processing [8], environment [9,10], and others. However, the analysis of HSI requires developing specialized methods and algorithms for data processing [11,12,13,14]. The main methods for the processing of the hyperspectral remote sensing images include image correction [11], noise reduction [15], dimensionality reduction [12], and classification [16,17,18,19].
In principle, the spectral information from the available hundreds of narrow bands collected by hyperspectral sensors can help discriminate among spectrally similar object pixels. Then, the accurate discrimination of different regions in the image is possible and the hyperspectral image classification is one of the most active part of the research in the hyperspectral field [18,20,21,22]. However, the HSI technology still faces a series of challenges, mainly including the following problems that need to be solved. The high dimensionality of the hyperspectral data, for example, the spectral reflectance values of hyperspectral images collected by airborne or space-borne imaging spectrometers, is up to hundreds of dimensions. Moreover, factors such as sensors, atmospheric conditions, surrounding environment, and composition and distribution of ground features affect the spatial variability of spectral information. The interference of noise (e.g., Poisson noise [23,24]) and background factors also seriously degrades the quality of the collected data and the corresponding classification accuracy of HSI. Finally, in practical applications, it is extremely difficult to obtain labeled samples to be used in the classification work of a hyperspectral image. The intrinsic properties of hyperspectral images need to be addressed specifically because conventional classification algorithms made for multispectral images do not adapt well to the analysis of hyperspectral images [25].
Several approaches have been proposed for classification of HSI. A subclass of classifiers is based on probabilistic approaches by using statistical tools to find the best class for a given pixel providing a probability of the pixel being a member of each of the possible classes. For instance, the multinomial logistic regression (MLR) classifier [26] supplies a degree of plausibility for such classes. In the sparse version of MLR a Laplacian prior to enforce sparsity is adopted which leads, with some computational limitations, to good generalization capabilities in HSI classification. More recently, an improved version of this classifier has also been proposed [27] using a subspace based method. The idea of applying subspace projection approach relies on the assumption that the samples within each class can approximately lie in a lower-dimensional subspace.
Due to their successful application in several problems of pattern recognition neural networks have also attracted many researchers in the field of the classification of hyperspectral images [28,29,30,31]. The main advantage of these approaches comes from the fact that neural networks do not need prior knowledge about the statistical distribution of the classes. They need the availability of feasible training techniques for nonlinearly separable data and their use has been affected by their training complexity as well as by the number of parameters that need to be tuned. Several neural network-based classification methods have been proposed in the literature that consider both supervised and unsupervised nonparametric approaches [32,33]. Recently, the extreme learning machine (ELM) algorithm has been successfully applied as nonlinear classifiers for hyperspectral data [34,35], and have shown remarkable efficiency in terms of accuracy and computational complexity. Some deep models have also been proposed for hyperspectral data feature extraction and classification [16]. The architecture design is the crucial part of a successful deep learning model together with the availability of an appropriate broad training set.
Another example of a supervised classification approach is support vector machines (SVMs). They have been widely used for the classification of hyperspectral data due to of their ability to handle high-dimensional data with a limited number of training samples [36]. To generalize the SVM for nonlinear classification problems, the so called kernel trick was introduced [37]. However, the sensitivity to the choice of the regularization parameters and the kernel parameters is the most important disadvantage of a kernel SVM. Other methods for HSI classification involve decision trees [38,39], random forests [40], and sparse representation classifiers with dictionary based generative models [41,42].
In this work we present a novel local/global method for semiautomatic multilabel classification of HSI. Semisupervised methods rely on limited information on the objects to recognize inside the data. Graph Clustering CNNs [43] consists of a two-stage clustering strategy in order to reduce the burden of graph convolution computation. The authors in [44] apply a coupled spatial-spectral approach for approximating the convolution on graphs. Autoencoder models are widely employed for HSI classification tasks (Ref. [45] and references therein). As a first step, we consider a preprocessing process using novel feature selection approaches in order to address the curse of dimensionality and reduce the redundancy of high-dimensional data. We adopt a linear discriminant analysis based on the labeled regions in order to project the high dimensional feature space to a lower dimensional subspace. Then we development a similarity index/distance between pixel reduced features of pixel using reduced features of pixel and involving pixels in a neighborhood. We point out that the so-called pixel based or spectral classifier only treats the HSI data as a list of spectral measurement without considering spatial relations of adjacent pixels, thus discarding important information. In a real image, neighboring pixels are related or correlated, both because imaging sensors acquire significant amount of energy from adjacent pixels and because homogeneous structures which are generally large compared to the size of a pixel frequently occurred in the image scene. This spatial contextual information should help for an accurate scene interpretation. Therefore, in order to improve classification results, spectral-spatial classification methods must be developed, which assign each image pixel to one class based on its own spectral values (the spectral information) and information extracted from its neighborhood (the spatial information). Then, the new similarity index includes the contextual spatial information provided in the HSI data considering features of adjacent pixels together. The same strategy was recently consider in [46] for the segmentation of color images.
Using the new similarity index we represent the image with an undirected graph where the set of vertices is the set of pixels present in the image, and the set of the edges consist of the pairs of neighboring pixels in the image. The weight of an edge can be represented by a function based on the difference between the features of each pixel. The vertices set can be partitioned into two sets: the “labeled vertices”, and the rest of the image pixels. In order to label the last vertices we develop a random walker approach which improves the quality of the image segmentation and involves only the minimization of a quadratic function. We remark that our method considers in a different nonlinear way two terms and it is not a simple thresholding step. In fact, the distances between features affect at the same time the similarity between labeled and unlabeled pixels and the construction of the graph for the random walk. We point out that we could consider the random walk method as a Laplacian-based manifold method with a solid theoretical background [47].
The remaining sections of the paper are organized as follows. Section 2 introduces the new method and the improved random walker segmentation algorithm. Moreover, we will also discuss a new definition of non-local distance between the features of the pixels. Section 3 is devoted to the numerical experiments. In this section the proposed method is tested on some benchmark images, in order to evaluate its performance. Following the findings of the case study, the conclusions are presented in the last Section 4 with a discussion about the properties and the possible developments of the approach.

2. A Spatial-Spectral Classifier Method for Hyperspectral Images

The detailed spectral information collected by the available hyperspectral sensors improves the capability of discriminating between different objects/sub-regions in an image by providing a division into classes with increased accuracy. This accurate capacity for discrimination makes hyperspectral data a valuable source of information to be fed to advanced classifiers. The output of the classification step is generally known as the classification map.
However, several particular problems should be considered in the classification task of hyperspectral data, among which are the following: the spatial variability of the spectral characteristics, the high number of spectral channels, the high cost of true sample labeling. In this paper we introduce a novel classification method in the semi-automatic approach framework. The general goal is to segment an image into two or more separate regions, each corresponding to a particular object (or the background), based on some user input. The proposed method reduces the interaction to the minimum, asking the user to just choose the object of interest by selecting a subregion inside it. The proposed approach does not attempt to estimate a specific model for a texture region, or an object, instead it tests for the homogeneity of a given feature pattern in a region: the features are the bands for each pixel in the image. In order to identify this homogeneity, the following assumptions about the image are made:
(a)
the image contains a set of approximately homogeneous regions;
(b)
spectral variables are also highly correlated that their dimension can be reduced without losing important information;
(c)
the features between two neighbouring regions are distinguishable.
The first assumption requires images that present several regions with similar features, and we require a noise level that allows to distinguish the different regions. See [48,49,50] for a discussion on noise affecting data. Based on the second assumption we include a dimension reduction as a first processing step. For classification, in the lower dimensional space, we expect that data are well separated. Then, in the low dimensional space nearby points or points on the same structure (cluster or manifold) are likely to have the same label. In our case, nearby points are those pixels with similar features. Assumption (a) refers to a local property of the image: in an homogeneous region there is an high probability that a random walker remains in such region, whilst it is unlikely that it can reach a far but similar region. On the other hand, Assumption (b) refers to a global property of the image: similar regions may share similar spectral signature even if they are distant.
The classification results could be improved by using the contextual spatial information provided in the HS data in addition to the spectral information. In order to capture the spatial variations of the pixel features we consider closest fixed neighborhoods and a distance which involves all the pixels in these neighborhoods. Finally, the classification map is obtained by combining a random walk based model [51,52,53] with a direct label assignment: both are computed using this distance. The resulting energy changes the energy related to the random walker approach and improves the quality of the image classification. Moreover, the algorithm involves only the minimization of a quadratic function. We point out that the graph-based methods have been paid attention because of their solid mathematical background [47], relationship with sparseness properties, kernel methods, model visualization, and reliable results in many areas.

2.1. Regularized Linear Discriminant Analysis

When dealing with hyperspectral image analysis, one usually has a large number of spectral features m and a relative small number of training marked pixels N T , divided in n M marked regions. A simple linear discriminant analysis (LDA) would then result in an ill-posed problem. In fact, from a theoretical point of view, one should deal with an infinite dimension Hilbert space to be linearly and compactly projected onto a finite dimensional one (see [54]). From a computational point of view, different strategies have been tested to overcome the ill-posedness of the problem. In [55] it is shown that an efficient version of the regularized LDA (RLDA) proposed in [56] performs better in segmenting hyperspectral images than support vector machine (SVM) classifiers and other LDA-based classifiers, such as penalized LDA, orthogonal LDA, and uncorrelated LDA.
The aim of RLDA is to find a projection matrix G ^ R m × l to reduce the high-dimensional feature x R m to a lower dimensional vector a = G ^ x R l , l m . More precisely (see [55], Equation (10)),
G ^ = arg max G R m × l trace G ( S + λ I ) G 1 G S b G ,
where S is the total scatter matrix, S b is the between-class variance matrix, I is the identity matrix, and λ is a regolarization parameter. Note that RLDA reduces to a classical LDA when λ = 0 .
The efficient Algorithm 1 for RLDA (see [55]) needs two singular values decompositions (SVD).
Algorithm 1 Efficient RLDA given in [55]
Require:  λ 0
Ensure:  G ^ as in (1)
[ U , D , V ] S V D ( H )                 ▹ H = U D V
r rank ( H ) = rank ( D )
D s D 2 + λ I r
[ U b , D b , V b ] S V D ( D s 1 / 2 U H b )     ▹ D s 1 / 2 U H b = U b D b V b
G ^ U D s 1 / 2 U b
The first one computes the SVD of the normalized data matrix
H = 1 n L [ X T μ 1 ] R m × N T ,
where X T R m × N T is the matrix data of the marked pixels and μ R m is the feature mean vector of such data. The second SVD involves the normalized between-class data matrix
H b = 1 N T [ n 1 ( μ 1 μ ) , n 2 ( μ 2 μ ) , , n n M ( μ n M μ ) ] R m × n M ,
where n j (resp. μ j ) denotes the sample size (res. the feature mean vector) for each marked region indexed by j 1 , , n M . The challenging calculation of a ϵ -approximation of such SVDs may be faced with the new efficient randomized algorithms given in [57] (to this aim, Matlab introduced the function svdsketch since R2020b, that we decided to use).
Finally, the matrix G ^ , output of the Algorithm 1, is the matrix that projects the n M marked regions in a subspace that optimizes the Fisher score in a robust way [55]. We expect that this low-dimensional subspace will carry the relevant information to segment the entire figure. Hence, the original image X 0 R m × n P is mapped to Y = G ^ X 0 R l × n P .

2.2. A Spectral/Spatial Similarity Measure

After reducing the data, it is necessary to define an appropriate metric to establish the similarity between pixels in the mapped image Y. Each pixel in y j in Y has l m features (projected bands) that represent a vector in the linear space R l . Instead, to focus on a pixel-wise similarity we fix a system of neighbourhoods with N b pixels, for each pixel, for example 8–neighborhoods, and a feature space, subset of R l . Finally we collect all the entries of the feature vectors in a single vector f j R l · N b . We fix a distance d l · N b in the space R l · N b . Then, for a couple of pixels y j and y i , we define the similarity index S ( y j , y i ) = d l · N b ( f j , f i ) + ε 1 , ε 1 , that will be used in our experiments as a weight function, see (8) below. Moreover, the notion of similarity may be extended to a comparison between a pixel y j and a label k { 1 , , n M } . In fact, the region marked by label k is formed by pixels that have their features, from which we may extract a “centroid feature” f ^ ( k ) , that best represents the labeled region. Then, for any pixel y i and any region k, the quantity
S ( i , k ) = d l · N b ( f i , f ^ ( k ) ) + ε 1
will represent the similarity between the pixel i and the region k. This index can be interpreted as the first step of a k-means algorithm, whose starting centroids are computed with the user-marked regions as seeds.
This similarity, or the distance, could be used in a clustering algorithm because it represents a measure and/or comparison between pixels. Moreover, this distance allows to see the feature of the pixel in relation to its neighbors. This non-local method is inspired by recent works in signal analysis [58,59,60].
Moreover, using information from the patches instead of single pixels induces a smoothing effect [61] which may be useful in presence of noise. Indeed, consider a pixel y j : denote with f j * R l · N b its corresponding feature vector and with f j f j * + N ( 0 , σ 2 I d ) its noisy version when Gaussian noise with zeros mean and covariance matrix σ 2 I d is considered, being I d the identity matrix. Consider another pixel y i such that the intersection of the neighbourhoods of y j and y i is empty. We can give an estimation of the expected value E d l · N b ( f j , f i ) 2 of the square euclidean distance as
E d l · N b ( f j , f i ) 2 = f j * f i * 2 + 2 · ( l · N b ) σ 2 = d l · N b ( f j * , f i * ) 2 + ( 2 l · N b ) σ 2
A similar estimation can be given when the noise affecting the pixels is not addictive but signal-dependent: for the case of Poisson noise f j Poiss ( f j * ) one obtains
E d l · N b ( f j , f i ) 2 = f j * f i * 2 + | f j * + f i * | 1 = d l · N b ( f j * , f i * ) 2 + f j * + f i * 1
where · 1 is the 1 norm in R l · N b and d l · N b ( f j * , f i * ) is the euclidean distance between the clean pixels. The above estimations are based on the fact that for a random variable X one has E [ X 2 ] = E [ X ] 2 + σ 2 ( X ) , with σ 2 ( X ) the variance of X.
Figure 1 depicts the while procedure for features compression and similarity computation. The first row is devoted to explain with a small example how the reduction is pursued. Consider just three pixels and their spectral distribution, i.e., the red, blue and green curves. The RLDA projection compresses each distribution in just a 2D vector: having at disposal these reduced vectors, one can compute the relative distance among these.
The second row depicts the procedure for a larger image: the grid on the left represents the pixels of the image, while the curves correspond to the spectral distributions for (some of) these pixels. Suppose that the dimension of the image is m × n × , where m × n is the dimension of the grid while refers to the number of recorded hyperspectral frequencies. The RDLA algorithm then compresses the distribution of each pixel in just two principal components ( P C 1 and P C 2 ), providing a smaller data with size m × n × 2 . Consider now two pixels, namely P 1 and P 2 , and their 8–neighborhoods in each component: in the previous notation, this hence amounts to have l = 2 and N b = 8 . Then, the vectors f 1 and f 2 that collect all the feature values of P 1 and P 2 and relative neighborhoods, respectively, belong to R 16 : we can compute the similarity index as
S ( P 1 , P 2 ) = 1 d 16 ( f 1 , f 2 ) + ε , ε 1
with d 16 being a distance in R 16 (for example, the classical Euclidean distance).

2.3. The Random Walker Method

The random walker (RW) method (see [51]) is a segmentation method based on a “path distance”. It belongs to the classes of probabilistic methods, that allow to address several scientific tasks, such as optimization problems via Consensus-Based Optimization [62], Opinion Formation [63] or Particle Swarm Optimization [64].
More precisely, the RW method works as follows: in each image there are subregions of pixels which are marked with labels. Then, starting from a unmarked pixel, a random walk is performed along the whole image. Obviously, the subregions closer to the starting points are more likely to be first reached than the others. The random walk is then biased in order to promote those paths that involve similar pixels: a longer path with similar pixels may have a shorter “path distance” than a longer one that crosses non homogeneous regions.
More formally, the framework here involves an undirected graph G = ( V , E ) , where V = { v i | v i is a pixel in the image } is the set of vertexes and E is the set of edges. Some pixels have been marked by the user, and will be denoted by V m , the set of “marked vertices” in the graph. V m consists of n M marked regions. Define the set of labels for the marked vertices as a function Q ( v j ) = k , v j V m . The other “unmarked vertices” of the image are denoted by V u , so that V = V m V u . The set of edges E will contain all the pairs of pixels which are neighbors in the original image, and e i j will denote the edge linking the vertexes v i and v j .
The weight of an edge e i j encodes node similarity, and is represented by a function ω ( v i , v j ) . If d ( v i , v j ) represents some distance between two pixels, then common choices of ω are
ω ( v i , v j ) = e β d ( v i , v j ) 2 or ω ( v i , v j ) = 1 ε + σ d ( v i , v j )
where the value of the parameters β , σ , ε > 0 can be tuned accordingly. Starting form an unmarked vertex, a random walk is performed. At each step, when the walker is in the vertex v i , the probability of reaching the neighbor v j depends on ω ( v i , v j ) proportionally: the highest values of the distance d will imply lower probabilities of reaching that neighbor. The algorithm then computes the probability of reaching any one of the labeled vertices belonging to one of the n M marked regions. Formally, for any vertex v i V u and k 1 , , n M , we denote by x i k this probability. For a labeled node v j V m that is associated to the label Q ( v j ) { 1 , , n M } , we have x j k = δ i , Q ( v i ) . Image segmentation is then made on these probabilities, and the algorithm will tend to observe the weights of the image edges during the segmentation.
In [51], the computation of the probabilities { x i k , i V u , k = 1 , , n M } are calculated by solving a sparse linear system of equations, that involves the graph Laplacian matrix L associated to ω , that is defined as
L i j = v k adjacent to v i ω ( v i , v k ) , if i = j ω ( v i , v j ) , if v i and v j are adjacent nodes 0 , otherwise .
In particular, for each label k, the probabilities x k = x 1 k , x 2 k , , x | V | k are found by solving the minimization of
E ( x k ) = 1 2 ( v i , v j ) E ω ( v i , v j ) ( x i k x j k ) 2 = 1 2 x k T L x k .
Note that, since L is positive semi-definite, the only critical points of E will be minima. In addition, since the corresponding continuous problem leads to the minimization of the Dirichlet integral via harmonic functions then the minimization problem (3) is also called combinatorial harmonic function (see [65]). The following problem
x D k = argmin x E ( x k ) , k = 1 , , n M
is also called combinatorial Dirichlet problem.
We recall that V = V m V u and V m V u = . Without loss of generality, it is assumed that the nodes in L and x are ordered so that marked nodes comes first and unmarked nodes after. Therefore, for each k { 1 , , n M } , we may decompose the above formula (3) into
E ( x m k , x u k ) = 1 2 x m k , x u k L m B B L u x m k x u k ,
where x m k and x u k correspond to the probabilities of the labeled and unlabeled nodes, respectively, while B represents the anti-diagonal blocks of the Laplacian matrix. The problem could be interpreted as an interpolation of missing data (the unmarked points), while we have defined some (numerical) values for a subset of the vertices (our labeled nodes).
Omitting the superscript k for the ease of notation, the Equation (5) reads
E ( x m , x u ) = 1 2 x m L m x m + x u B x m + 1 2 x u T L u x u
and the unknowns are the entries of the vector x u . Differentiating E with respect to x u and finding the critical point yields
L u x u = B x m
which is a system of linear equations with | V u | unknowns. If every connected component contains at least a labeled vertex, then the Equation (6) is non-singular. Define X u = ( x u 1 , x u 2 , , x u n M ) and X m = ( x m 1 , x m 2 , , x m n M ) , then the solution to the combinatorial Dirichlet problems (4) may be found by solving
L u X u = B X m .
where X m is the labeling matrix with values in { 0 , 1 } such that X m 1 = 1 and L u is invertible. Thus, each unlabeled pixel v i gets n M probabilities ( x i 1 , x i 2 , , x i n M ) . Eventually, the label assigned to v i V u corresponds to the index of maximum-by-rows in the solution of (7). For example, suppose that an image contains only n M = 4 marked regions. Consider just one pixel v i V u : the solution of (7) for this pixel reads as ( x i 1 , x i 2 , x i 3 , x i 4 ) = ( 0.15 , 0.5 , 0.25 , 0.1 ) : this means that a random walker starting from v i has a probability of reaching the first region equal to 0.15 , it has a probability of reaching the region labeled with k = 2 of 0.5 and to reach the the region marked as k = 3 or k = 4 with probability of 0.25 or 0.1 , respectively. This pixel is then labeled with as belonging to the second region, since a random walker is more likely attracted from that region. This approach is adopted also in [53].
Finally, in the next section we will test a combination of RW probabilities and similarity index, where the weights in Equation (3) are chosen as
ω ( v i , v j ) = S ( v i , v j ) = 1 d l · N b ( f i , f j ) + ε , ε = 10 3 .

2.4. A Local/Global Classification Method

Now, we consider the vertex labeling function, for simplicity we will consider labels represented by integers,
F L : V S L = { 1 , 2 , , n M } , n M N , n M > 1
which associates a label in a certain set to each vertex (pixel). Combining the RW approach, with the new distance defined above, and the new similarity measure, let F L as
F L ( v i ) = argmax k S L ( S ( i , k ) α ( x i k ) 1 α )
where S ( i , k ) is given in (2), x i k are solutions of (7), and 1 α 0 is a parameter introduced for adding flexibility to the algorithm and to provide different weights to the two components of the labeling function. Due to the concavity of the logarithm function, and the positivity of S ( i , k ) and x i k , we can rewrite the labeling problem in an equivalent way as follows
F L ( v i ) = argmax k S L α log ( S ( i , k ) ) + ( 1 α ) log ( x i k ) .
The proposed method therefore can be summarized in Algorithm 2.
Algorithm 2 Hyperspectral Random Walk by Similarity Index algorithm (HyperRaWaSI)
  • Set the parameters α , λ , the neighbor system and the similarity function.
  • Acquisition of user-marked pixels.
  • Compute the projection of the original image X 0 , Y = G ^ X 0 , where G ^ is given by Algorithm 1.
  • Compute the global similarity index S ( i , k ) for any v i V u and v k V m .
  • Solve system (7) for V m , where the Laplacian matrix uses the index computed at step 3.
  • Evaluate the labeling function as in (10) for the image X 0 .
Remark 1.
The two terms in the functional in (10) could be considered as a “fidelity term”, the α log ( S ) part, and a regularizing term, the ( 1 α ) log ( x ) part.
Remark 2.
The approach employed in our work is not simple post-processing: the similarity index with spatial information plays a role inside the classification decision. The optimization viewpoint depicted in Equation (10) raises a different framework from a post-processing one. Furthermore, the features distances affect at the same time the similarity between labeled and unlabeled pixels and on the construction of the graph for the part of the random walk.
Figure 2 provides a visual representation of the steps for Algorithm 2 and how they are related. The very first step is to acquire the user-marked region in the original image: this labels are the building blocks of the entire procedure, since
  • they are employed to get the optimal projection of the hyperspectral image and hence the feature image;
  • they are used to compute the centroids from the feature image;
  • they represent the seeds for the random walker method.
The final segmentation is obtained by solving (10), which involves a convex combination of the similarity indexed image, obtained by the feature image and the centroids, and the output of the random walk method.

3. Results

In this section we present the numerical experiments done in order to assess the performance of the proposed algorithm. We used four different datasets, publicly available at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (last visited: 17 October 2021):
  • University of Pavia: this image was taken by the German Aerospace Agency (DLR) using the airborne ROSIS (Reflective Optics System Imaging Spectrometer) sensor. The spatial dimensions of the slices are 610 × 340 and the number of spectral band is 103: then the dataset size is 103 × 207,400. The ground resolution is 1.3 m and the spectral gamma is 430–860 nm. The number g of ground truth labels is 9: Asphalt, Meadows, Gravel, Trees, Painted metal sheets, Bare soil, Bitumen, Self-blocking bricks, and Shadow.
  • Pavia Center: this image refers to the center of the city of Pavia, but some samples in this dataset contain no information: the spatial size is 1096 × 1096 and it is then reduced to 1096 × 715 . The considered spectral bands are 102, leading to a final size of 102 × 783,640. The spectral bands lie in the interval 430–860 nm and the ground resolution is 1.3 m. The g = 9 ground truth labels are Water, Trees, Asphalt, Self-Blocking Bricks, Bitumen, Tiles, Shadows, Meadows, Bare Soil.
  • KSC: this image refers to the Kennedy Space Center, Florida (US) and it was acquired by the airborne AVIRIS (Airborne Visible/Infrared Imaging Spectrometer) NASA instrument. The spatial dimensions of the slices are 512 × 614 and the number of the spectral bands is 176: the size of the dataset is hence 176 × 314,368. The ground resolution is 18 m and the spectral gamma is 400–2500 nm. The g ground truth labels are 13: Scrub, Willow swamp, CP hammock, CP/Oak, Slash pine, Oak/Broadleaf, Hardwood swamp, Graminoid marsh, Spartina marsh, Cattail marsh, Salt marsh, Mud flats, Water.
  • Indian Pines: this image refers to the Indian Pines test site in North-Western Indiana, taken by AVIRIS sensor. The spatial dimensions are 145 × 145 and the employed bands are 220: The dataset size is hence 220 × 21,025: the spectral gamma is 0.4 2.5 nm. The g ground truth labels are 16: Alfalfa, Corn-notill, Corn-mintill, Corn, Grass-pasture, Grass-trees, Grass-pasture-mowed, Hay-windrowed, Oats, Soybean-notill, Soybean mintill, Soybean-clean, Wheat, Woods, Building-grass-trees-drives, Stone-trees-drives.
  • Salinas HSI: this image refers to an area including agricultural fields in the Salinas Valley, California, acquired again by AVIRIS sensor. The spatial size is 512 × 217 and the number of bands is 204: the dataset size is hence 204 × 111,104 the ground resolution is 3.7 m, while the spectral gamma is 0.4 2.5 nm. The g = 16 ground truth labels are: Broccoli gr. wds 1, Broccoli gr. wds, Fallow, Fallow rough plow, Fallow smooth, Stubble, Celery, Grapes untrained, Soil vineyard develop, Corn sen. gr. wds, Lettuce romaine 4 wk, Lettuce romaine 5 wk, Lettuce romaine 6 wk, Lettuce romaine 7 wk, Vineyard untrained, Vineyard vert. trellis.
The numerical experiments are organized as follows: the datasets Indian Pines, Pavia University and Salinas HSI are employed to assess the performance of the proposed algorithm, using suitable indexes and segmentation quality measurements. The datasets Pavia Center and KSC are used as real-world datasets: the labels contained in the ground truth are employed as seed for our segmentation technique.
All the experiments were carried on a laptop equipped with Linux 19.04, with an Intel(R) Core(TM) i5–8250U CPU (1.60 GHz), 16 GiB RAM memory (Intel, Santa Clara, CA, USA) and under MatLab R2020b environment (MathWorks, Natick, MA, USA). The code is available at https://github.com/AleBenfe/RaWaCs (last access: 17 October 2021).

3.1. Performance Measurements

This section makes use of the datasets Indian Pines, Pavia University, and Salinas HSI. As previously mentioned, the proposed method belongs to the class of semi-supervised techniques: the initial seeds, i.e., the user-marked regions, are in a lower number then the original ones, that are 16, 9, and 16 for Indian Pines, Pavia University, and Salinas HSI, respectively. We denote with G t the ground truth, which contains N G labels, and with { g i } i = 1 , , g the marked subsets. We select manually n M labels on the dataset: this will produce a segmentation S t result with n M marked regions { s j } j = 1 , , n M . In order to assess the performance when the ground truth contains a number g of labels greater than n M , we make use of the following indexes.
  • Rand Index (RI) [52]: this index measures if the two partitions of the image, namely G t and S t , are coherent. For any couple of pixels ( p 1 , p 2 ) in the ground truth G t , the RI measures the coherence between the partitions: it checks if p 1 and p 2 belongs to the same subset g G t and at the same time they belongs to the same subset s S t . It checks also if p 1 and p 2 belongs to the two different subsets g 1 , g 2 G t and at the same time they belongs to two different subsets s 1 , s 2 S t . Denote with n s the number of couples that belongs to the same subset in G t and that belong to the same subset in S t , while denote with n d the number of couples that do not belong to the same subset in G t and that do not belong to the same subset in S t , then
    R I = n s + n d n G 2
    where n G is the total number of pixels labeled in G t and n G 2 refers to the number of all possible couples. The best performances are obtained when the RI is close to 1.
  • Overall Accuracy (OA): for each true label i = 1 , , n G , we compute
    ı ^ = arg max j = 1 , , n { # ( g i s j ) } .
    This classifies the label i as belonging to the region of s ı ^ S t , marked with ı ^ . The Overall Accuracy is defined as
    O A = i = 1 N g # ( g i s ı ^ ) i = 1 g # ( g i ) = i = 1 N g max j = 1 , , n { # ( g i s j ) } n G ,
    which measures the average number of pixels of the ground truth G t that are truly classified together. The best performances are obtained when the OA is close to 1.
The first experiment refers to the Indian Pines dataset: the original image is depicted in Figure 3a, the segmentation result in Figure 3b whilst the ground truth labels are in Figure 3b. These results are obtained by setting λ = 0.01 and α = 0.8 in Algorithm 2. The Rand Index for this particular experiment is 0.8505, while the OA is 0.9029: this means that each user-marked label contains the most part of a region marked in the ground truth. This result is clearly shown in Figure 3d: the element in ( i , j ) position shows the percentage of the pixels in ground truth label j which are marked as belonging to the user-marked label i.
The Pavia University and Salinas HSI datasets present very similar results, with the same setting for the parameters ( α = 0.8 , λ = 0.01 ): the segmentation on the former provides a RI of 0.8060 and a OA of 0.9393, and the proposed procedure reaches a reliable performance on the latter too, with RI = 0.8704 and OA = 0.9696. These experiments show that the proposed procedure is able to properly cluster several regions, according to common hyperspectral properties, even when the initial seeds are selected via a visual inspection of some of the bands presented in the dataset. Figure 4 presents the results for the Pavia University and Salinas HSI datasets.
Table 1 collects the computational times of the entire procedure (excluding the marking process) for each image: even for very large datasets such as Pavia University the proposed algorithm is very fast. We employed the MatLab function svdsketch which implements the RLDA procedure depicted in Algorithm 1.
We now assess the dependence of the performance on the parameters α and λ : the plots in Figure 5 depict the behavior of the Rand Index with respect to α in correspondence of several choice for λ , namely from λ = 1 to λ = 10 6 . These plots show that the proposed procedure is quite stable with respect to this settings, showing an increasing of the performance around α = 0.8 . Moreover, the dimensionality reduction seems to be independent on the choice of λ . As already noted in [46], the best performances are achieved when α 0 and α 1 , meaning that the coupling (in this case, via a convex combination) of the RW approach and of the similarity index is the optimal strategy with respect to selecting only one of the two methods.

3.2. Comparison with State–of–the–Art Methods

This section is devoted to compare the proposed strategy with state-of-the-art methods, namely K-MBC [21], FCM [66], FDPC [67], and GWEEN [68], when the true number of clusters is known. The seeds employed for our method are randomly chosen in the ground truth mask: for each marked region, two small squares with size 7 pixels are selected. See Figure 6 for a visual explanation of this process and for example of different square size.
Table 2 presents the comparison between our method and the other ones using the Overall Accuracy and the Purity indexes as evaluations. The Purity index measures how much pure a cluster is, i.e., its tendency to contain a sole class. The performance indexes regarding the other methods are taken from ([21], Tables 4 and 5). The setting for the proposed procedure are α = 0.93 , 0.92 and α = 0.91 for Salinas, Pavia, and KSC datasets, respectively, while λ = 0.1 for all 3 datasets.
The proposed procedure overcomes all the methods used in the comparison, both on overall accuracy and on purity indexes.
We test the performance of the strategy with respect to the size and shape of the user-marked regions. We consider the salinas and Pavia datasets, and we employ as user-marked regions two randomly chosen squares with size 3, 5, and 7 (see Figure 6). The results are depicted in Table 3.
As one expects, the larger the marking size, the higher the performances: even with a very small marking ( 3 × 3 square), the overall accuracy is above 80% and the purity is greater than 90% in the case of salinas dataset, meaning that the recognized clusters contain mostly one class.

3.3. Segmentation as Atlas

In this section the Pavia Center and KSC datasets are segmented using the ground truth labels as seeds. The idea beyond this strategy relies on the concept of atlas (see [69]): an atlas is a collection of objects whose labels are known. The strategy consists in merging the images of the atlas with the one which has to be segmented, hence the resulting image in this way contains a labeled part (from the atlas) which allows the classification of the rest of the data.
We apply this strategy to the Pavia Center and KSC datasets: we adopted this approach since the subsets labeled in the ground truth data are low in number in both cases, despite the dimension of the images.
Figure 7 refers to the Pavia center dataset: in the left panel a colored image of the landscape is depicted, while the central panel shows the labels contained in the ground truth. The strategy consists then of using these labeled regions as an atlas, providing exact information for the classification, without requiring user intervention. The final result, obtained with α = 0.9 and λ = 0.2 is given in the last panel of Figure 7: a visual inspection of the achieved segmentation suggests that the recognition of the various regions is remarkable and precise. Indeed, all the red roofs, together with their projected shadow, are recognized in a remarkable way. The trees are well segmented even if few markings of them are contained in the atlas. Note that there is an abrupt interruption inside the image, as previously mentioned in the description at the beginning of this section: despite this discontinuity, the proposed procedure does not suffer from the presence of such issue in the dataset, providing a segmentation that can take into account that the image is not continuous.
We apply the same strategy (using the ground truth as an atlas) to the KSC dataset, where the main aim is to segment the different type of vegetation. Figure 8 presents the visual representation of the segmentation result, obtained by setting α = 0.97 and λ = 10 4 . The right part of the image present a reliable segmentation, since the different parts of terrain and vegetation are well recognized. On the other hand, the segmentation result on the left part classifies some buildings as part of vegetation or water, since the ground truth does not present any label referring to human constructions, such as buildings, bridges or streets. Figure 9 is devoted to showing the reason why some buildings are included in the Water label: indeed, the river contains several white parts which the segmentation process correctly includes in the Water label. On the other hand, there are several buildings that present a spectral signature (both in the visible and in the infrared region) very similar to these white regions inside the river. The picture shown in Figure 9 also better exposes the reliable performance on the right part of the image.
We finally report that the computational times for the segmentation of Pavia Center and KSC datasets, which present both large dimensions: 102 × 783,640 the former and 176 × 314,368 the latter. In both cases, the computational time is close to being negligible with respect to the size (see Table 4).

4. Conclusions

In this work we proposed a new spectral/spatial semi-automatic classification method coupling a Regularized Linear Discriminant analysis for dimension reduction, a suitable similarity index, and a random walker approach. The final computational problem involves a new energy function obtained by a new definition of similarity/distance between pixel using a reduced features space and pixels in a neighborhood. The experimental results showed that the proposed approach is very robust with respect to the presence of noise and with good accuracy. In our approach we have the flexibility of a system of different neighborhoods for the calculation of the distance between pixels and in the construction of the graph corresponding to a given image.
Regarding the hyper-parameters λ and α , we observed that in the regularization process the choice of the λ parameter does not seem to influence the numerical results even varying it by several orders of magnitude. The RW part encourages the selection of convex zones, consequently changing the α parameter in Equation (10) can give more weight to these sub-regions in the final classification. In a future paper we will consider appropriate learning methods for the optimal choice of these parameters for some classes of images.
From the computational point of view, the most expensive steps concern the calculation of the new distance between pixels and the preliminary reduction step (the computation of the projection operator G ^ , see Algorithm 1). However, we observe that these operations can be performed efficiently in parallel, for example with an appropriate implementation through the use of GPUs. Moreover, we adopted a fast state-of-the-art SVD algorithm. The computation of the probabilities of RW requires the numerical solution of linear systems which may be large, but sparse and well structured at the same time, consequently, efficient algorithms can be used. Then, our method is reliable and efficient also for high-definition images.
We point out that it is possible to introduce some set of suitable atlas through the projection matrix G ^ , see Algorithm 1. We plan to explore the combination of atlas and our classifier using an adaptive approach in order to learn better weights to be employed in the features distance. Moreover, we will consider the possibility to replace some or all the labeled pixels with atlases well adapted to the specific image. Preliminary results were shown and discussed in Section 3.3. Further comparisons will be made with other semi-automatic methods, identifying suitable quality measures of the segmentation obtained.

Author Contributions

Conceptualization, G.A. and G.N.; methodology, G.A., A.B. and G.N.; software, G.A.; validation, G.A. and A.B.; formal analysis, G.A., A.B. and G.N.; investigation, G.A. and G.N.; resources, G.A., A.B. and G.N.; data curation, G.A. and A.B.; writing—original draft preparation, G.A., A.B. and G.N.; writing—review and editing, A.B.; visualization, A.B.; supervision, G.A., A.B. and G.N.; project administration, G.N.; funding acquisition, G.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received fundings from PRECISION—PRecision crop protECtion: deep learnIng and data fuSION project.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

On 3 December 2021, the code for the MatLab implementation of RaWaCs is available at https://github.com/AleBenfe/RaWaCs. Publicly available datasets were analyzed in this study. On 17 October 2021, data were found at: http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Elmasry, G.; Mandour, N.; Al-Rejaie, S.; Belin, E.; Rousseau, D. Recent applications of multispectral imaging in seed phenotyping and quality monitoring—An overview. Sensors 2019, 19, 1090. [Google Scholar] [CrossRef] [Green Version]
  2. Lu, B.; Dao, P.; Liu, J.; He, Y.; Shang, J. Recent advances of hyperspectral imaging technology and applications in agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  3. Ma, J.; Sun, D.W.; Pu, H.; Cheng, J.H.; Wei, Q. Advanced Techniques for Hyperspectral Imaging in the Food Industry: Principles and Recent Applications. Annu. Rev. Food Sci. Technol. 2019, 10, 197–220. [Google Scholar] [CrossRef] [PubMed]
  4. Ghamisi, P.; Plaza, J.; Chen, Y.; Li, J.; Plaza, A. Advanced Spectral Classifiers for Hyperspectral Images: A review. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–32. [Google Scholar] [CrossRef] [Green Version]
  5. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J. Feature Extraction for Hyperspectral Imagery: The Evolution from Shallow to Deep: Overview and Toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  6. Gutiérrez-Gutiérrez, J.; Pardo, A.; Real, E.; López-Higuera, J.; Conde, O. Custom scanning hyperspectral imaging system for biomedical applications: Modeling, benchmarking, and specifications. Sensors 2019, 19, 1692. [Google Scholar] [CrossRef] [Green Version]
  7. Halicek, M.; Fabelo, H.; Ortega, S.; Callico, G.; Fei, B. In-vivo and ex-vivo tissue analysis through hyperspectral imaging techniques: Revealing the invisible features of cancer. Cancers 2019, 11, 756. [Google Scholar] [CrossRef] [Green Version]
  8. Qureshi, R.; Uzair, M.; Khurshid, K.; Yan, H. Hyperspectral document image processing: Applications, challenges and future prospects. Pattern Recognit. 2019, 90, 12–22. [Google Scholar] [CrossRef]
  9. Muller-Karger, F.; Hestir, E.; Ade, C.; Turpie, K.; Roberts, D.; Siegel, D.; Miller, R.; Humm, D.; Izenberg, N.; Keller, M.; et al. Satellite sensor requirements for monitoring essential biodiversity variables of coastal ecosystems. Ecol. Appl. 2018, 28, 749–760. [Google Scholar] [CrossRef]
  10. Goddijn-Murphy, L.; Peters, S.; van Sebille, E.; James, N.; Gibb, S. Concept for a hyperspectral remote sensing algorithm for floating marine macro plastics. Mar. Pollut. Bull. 2018, 126, 255–262. [Google Scholar] [CrossRef] [Green Version]
  11. Yokoya, N.; Grohnfeldt, C.; Chanussot, J. Hyperspectral and multispectral data fusion: A comparative review of the recent literature. IEEE Geosci. Remote Sens. Mag. 2017, 5, 29–56. [Google Scholar] [CrossRef]
  12. Li, W.; Feng, F.; Li, H.; Du, Q. Discriminant Analysis-Based Dimension Reduction for Hyperspectral Image Classification: A Survey of the Most Recent Advances and an Experimental Comparison of Different Techniques. IEEE Geosci. Remote Sens. Mag. 2018, 6, 15–34. [Google Scholar] [CrossRef]
  13. Signoroni, A.; Savardi, M.; Baronio, A.; Benini, S. Deep learning meets hyperspectral image analysis: A multidisciplinary review. J. Imaging 2019, 5, 52. [Google Scholar] [CrossRef] [Green Version]
  14. Khan, M.; Khan, H.; Yousaf, A.; Khurshid, K.; Abbas, A. Modern Trends in Hyperspectral Image Analysis: A Review. IEEE Access 2018, 6, 14118–14129. [Google Scholar] [CrossRef]
  15. Rasti, B.; Scheunders, P.; Ghamisi, P.; Licciardi, G.; Chanussot, J. Noise reduction in hyperspectral imagery: Overview and application. Remote Sens. 2018, 10, 482. [Google Scholar] [CrossRef] [Green Version]
  16. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  17. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  18. Shah, S.; Qureshi, S.; Rehman, A.; Shah, S.; Hussain, J. Classification and Segmentation Models for Hyperspectral Imaging—An Overview. Commun. Comput. Inf. Sci. 2021, 1382, 3–16. [Google Scholar] [CrossRef]
  19. Gu, Y.; Chanussot, J.; Jia, X.; Benediktsson, J. Multiple Kernel Learning for Hyperspectral Image Classification: A Review. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6547–6565. [Google Scholar] [CrossRef]
  20. Borzov, S.; Potaturkin, O. Spectral-Spatial Methods for Hyperspectral Image Classification. Review. Optoelectron. Instrum. Data Process. 2018, 54, 582–599. [Google Scholar] [CrossRef]
  21. Le Moan, S.; Cariou, C. Minimax bridgeness-based clustering for hyperspectral data. Remote Sens. 2020, 12, 1162. [Google Scholar] [CrossRef] [Green Version]
  22. Cheng, G.; Li, Z.; Han, J.; Yao, X.; Guo, L. Exploring Hierarchical Convolutional Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6712–6722. [Google Scholar] [CrossRef]
  23. Bonettini, S.; Benfenati, A.; Ruggiero, V. Primal-dual first order methods for total variation image restoration in presence of poisson noise. In Proceedings of the 2014 IEEE International Conference on Image Processing, ICIP 2014, Paris, France, 27–30 October 2014; pp. 4156–4160. [Google Scholar] [CrossRef]
  24. Bonettini, S.; Benfenati, A.; Ruggiero, V. Scaling techniques for ϵ-subgradient methods. SIAM J. Optim. 2016, 26, 1741–1772. [Google Scholar] [CrossRef] [Green Version]
  25. Landgrebe, D. Multispectral land sensing: Where from, where to? IEEE Trans. Geosci. Remote Sens. 2005, 43, 414–421. [Google Scholar] [CrossRef]
  26. Böhning, D. Multinomial logistic regression algorithm. Ann. Inst. Stat. Math. 1992, 44, 197–200. [Google Scholar] [CrossRef]
  27. Li, J.; Bioucas-Dias, J.; Plaza, A. Spectral-spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  28. Benediktsson, J.; Swain, P.; Ersoy, O. Conjugate-gradient neural networks in classification of multisource and very-high-dimensional remote sensing data. Int. J. Remote Sens. 1993, 14, 2883–2903. [Google Scholar] [CrossRef]
  29. Yang, H.; Van Der Meer, F.; Bakker, W.; Tan, Z. A back-propagation neural network for mineralogical mapping from AVIRIS data. Int. J. Remote Sens. 1999, 20, 97–110. [Google Scholar] [CrossRef]
  30. Zhou, P.; Han, J.; Cheng, G.; Zhang, B. Learning Compact and Discriminative Stacked Autoencoder for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4823–4833. [Google Scholar] [CrossRef]
  31. Cheng, G.; Yang, C.; Yao, X.; Guo, L.; Han, J. When Deep Learning Meets Metric Learning: Remote Sensing Image Scene Classification via Learning Discriminative CNNs. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2811–2821. [Google Scholar] [CrossRef]
  32. Merényi, E.; Farrand, W.; Taranik, J.; Minor, T. Classification of hyperspectral imagery with neural networks: Comparison to conventional tools. Eurasip J. Adv. Signal Process. 2014, 2014, 71. [Google Scholar] [CrossRef] [Green Version]
  33. Del Frate, F.; Pacifici, F.; Schiavon, G.; Solimini, C. Use of neural networks for automatic classification from high-resolution images. IEEE Trans. Geosci. Remote Sens. 2007, 45, 800–809. [Google Scholar] [CrossRef]
  34. Zhou, Y.; Peng, J.; Chen, C. Extreme Learning Machine with Composite Kernels for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2351–2360. [Google Scholar] [CrossRef]
  35. Santos, A.; Araujo, A.; Menotti, D. Combining multiple classification methods for hyperspectral data interpretation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 1450–1459. [Google Scholar] [CrossRef]
  36. Pal, M.; Mather, P. Assessment of the effectiveness of support vector machines for hyperspectral data. Future Gener. Comput. Syst. 2004, 20, 1215–1225. [Google Scholar] [CrossRef]
  37. Li, J.; Marpu, P.; Plaza, A.; Bioucas-Dias, J.; Benediktsson, J. Generalized composite kernel framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  38. Friedl, M.; Brodley, C.; Strahler, A. Maximizing land cover classification accuracies produced by decision trees at continental to global scales. IEEE Trans. Geosci. Remote Sens. 1999, 37, 969–977. [Google Scholar] [CrossRef]
  39. Pal, M.; Mather, P. An assessment of the effectiveness of decision tree methods for land cover classification. Remote Sens. Environ. 2003, 86, 554–565. [Google Scholar] [CrossRef]
  40. Xia, J.; Falco, N.; Benediktsson, J.; Du, P.; Chanussot, J. Hyperspectral Image Classification with Rotation Random Forest via KPCA. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1601–1609. [Google Scholar] [CrossRef] [Green Version]
  41. Chen, Y.; Nasrabadi, N.; Tran, T. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  42. Xing, Z.; Zhou, M.; Castrodad, A.; Sapiro, G.; Carin, L. Dictionary learning for noisy and incomplete hyperspectral images. SIAM J. Imaging Sci. 2012, 5, 33–56. [Google Scholar] [CrossRef]
  43. Zeng, H.; Liu, Q.; Zhang, M.; Han, X.; Wang, Y. Semi-supervised Hyperspectral Image Classification with Graph Clustering Convolutional Networks. arXiv 2020, arXiv:2012.10932. [Google Scholar]
  44. Qin, A.; Shang, Z.; Tian, J.; Wang, Y.; Zhang, T.; Tang, Y.Y. Spectral–Spatial Graph Convolutional Networks for Semisupervised Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 241–245. [Google Scholar] [CrossRef]
  45. Hanachi, R.; Sellami, A.; Farah, I.R.; Mura, M.D. Semi-supervised Classification of Hyperspectral Image through Deep Encoder-Decoder and Graph Neural Networks. In Proceedings of the 2021 International Congress of Advanced Technology and Engineering (ICOTEN), Istanbul, Turkey, 22–23 December 2021; pp. 1–8. [Google Scholar] [CrossRef]
  46. Aletti, G.; Benfenati, A.; Naldi, G. A Semiautomatic Multi-Label Color Image Segmentation Coupling Dirichlet Problem and Colour Distances. J. Imaging 2021, 7, 208. [Google Scholar] [CrossRef]
  47. Belkin, M.; Niyogi, P. Towards a theoretical foundation for Laplacian-based manifold methods. J. Comput. Syst. Sci. 2008, 74, 1289–1308. [Google Scholar] [CrossRef] [Green Version]
  48. Benfenati, A.; Ruggiero, V. Image regularization for Poisson data. J. Phys. Conf. Ser. 2015, 657, 012011. [Google Scholar] [CrossRef] [Green Version]
  49. Bertero, M.; Boccacci, P.; Ruggiero, V. Inverse Imaging with Poisson Data; IOP Publishing: Bristol, UK, 2018. [Google Scholar] [CrossRef]
  50. Benfenati, A.; Ruggiero, V. Inexact Bregman iteration for deconvolution of superimposed extended and point sources. Commun. Nonlinear Sci. Numer. Simul. 2015, 20, 882–896. [Google Scholar] [CrossRef]
  51. Grady, L. Random walks for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1768–1783. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Bampis, C.G.; Maragos, P.; Bovik, A.C. Graph-Driven Diffusion and Random Walk Schemes for Image Segmentation. IEEE Trans. Image Process. 2017, 26, 35–50. [Google Scholar] [CrossRef] [PubMed]
  53. Casaca, W.; Gois, J.P.; Batagelo, H.C.; Taubin, G.; Nonato, L.G. Laplacian Coordinates: Theory and Methods for Seeded Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2665–2681. [Google Scholar] [CrossRef]
  54. Shin, H. An extension of Fisher’s discriminant analysis for stochastic processes. J. Multivar. Anal. 2008, 99, 1191–1216. [Google Scholar] [CrossRef]
  55. Bandos, T.; Bruzzone, L.; Camps-Valls, G. Classification of hyperspectral images with regularized linear discriminant analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  56. Ye, J.; Xiong, T.; Li, Q.; Janardan, R.; Bi, J.; Cherkassky, V.; Kambhamettu, C. Efficient model selection for regularized linear discriminant analysis. In Proceedings of the International Conference on Information and Knowledge Management, Arlington, VA, USA, 6–11 November 2006; pp. 532–539. [Google Scholar] [CrossRef]
  57. Yu, W.; Gu, Y.; Li, Y. Efficient randomized algorithms for the fixed-precision low-rank matrix approximation. SIAM J. Matrix Anal. Appl. 2018, 39, 1339–1359. [Google Scholar] [CrossRef]
  58. Cagli, E.; Carrera, D.; Aletti, G.; Naldi, G.; Rossi, B. Robust DOA estimation of speech signals via sparsity models using microphone arrays. In Proceedings of the 2013 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, New Paltz, NY, USA, 20–23 October 2013; pp. 1–4. [Google Scholar]
  59. Aletti, G.; Naldi, G.; Parigi, G. Around the image analysis of the vessels remodelling during embryos development. In Proceedings of the 19th European Conference on Mathematics for Industry, Santiago de Compostela, Spain, 13–17 June 2016; p. 225. [Google Scholar]
  60. Aletti, G.; Moroni, M.; Naldi, G. A new nonlocal nonlinear diffusion equation for data analysis. Acta Appl. Math. 2020, 168, 109–135. [Google Scholar] [CrossRef] [Green Version]
  61. Wang, X.; Zhu, C.; Bichot, C.E.; Masnou, S. Graph-based image segmentation using weighted color patch. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 4064–4068. [Google Scholar] [CrossRef] [Green Version]
  62. Carrillo, J.A.; Choi, Y.P.; Totzeck, C.; Tse, O. An analytical framework for consensus-based global optimization method. Math. Model. Methods Appl. Sci. 2018, 28, 1037–1066. [Google Scholar] [CrossRef] [Green Version]
  63. Benfenati, A.; Coscia, V. Nonlinear microscale interactions in the kinetic theory of active particles. Appl. Math. Lett. 2013, 26, 979–983. [Google Scholar] [CrossRef]
  64. Kennedy, J. Particle Swarm Optimization. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2010; pp. 760–766. [Google Scholar] [CrossRef]
  65. Grady, L.; Polimeni, J.R. Discrete Calculus: Applied Analysis on Graphs for Computational Science; Springer: London, UK, 2010. [Google Scholar]
  66. Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Springer Science & Business Media: Boston, MA, USA, 2013. [Google Scholar]
  67. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  68. Cariou, C.; Chehdi, K. A new k-nearest neighbor density-based clustering method and its application to hyperspectral images. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 6161–6164. [Google Scholar] [CrossRef]
  69. Iglesias, J.; Sabuncu, M. Multi-atlas segmentation of biomedical images: A survey. Med. Image Anal. 2015, 24, 205–219. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Visual representation of the RLDA algorithm and distance computation of the neighborhoods. First row: example with just 3 pixels and their hyperspectral distribution, reduced in their 2D principal components. Second row: RLDA applied to a larger image. Third row: extraction of the 8-neighborhoods of two pixels in each principal component.
Figure 1. Visual representation of the RLDA algorithm and distance computation of the neighborhoods. First row: example with just 3 pixels and their hyperspectral distribution, reduced in their 2D principal components. Second row: RLDA applied to a larger image. Third row: extraction of the 8-neighborhoods of two pixels in each principal component.
Jimaging 07 00267 g001
Figure 2. Flow chart of Algorithm 2.
Figure 2. Flow chart of Algorithm 2.
Jimaging 07 00267 g002
Figure 3. Indian Pines segmentation results. (a): false gray scale image. (b): segmentation result. (c): true labels. (d): accuracy over the classes. The colormaps in (b,c) are different because the segmentation process takes into account 5 labels, while the ground truth contains 16 labels. Each region in the ground truth falls almost entirely in one of the 5 manually selected labels.
Figure 3. Indian Pines segmentation results. (a): false gray scale image. (b): segmentation result. (c): true labels. (d): accuracy over the classes. The colormaps in (b,c) are different because the segmentation process takes into account 5 labels, while the ground truth contains 16 labels. Each region in the ground truth falls almost entirely in one of the 5 manually selected labels.
Jimaging 07 00267 g003
Figure 4. (ac): false grayscale image, segmentation result and ground truth labels for the Pavia University dataset. (df): false grayscale image, segmentation result and ground truth labels for the Salinas HSI dataset. The former experiments achieves an Overall Accuracy of 0.93, whilst the latter achieves an OA of 0.9696.
Figure 4. (ac): false grayscale image, segmentation result and ground truth labels for the Pavia University dataset. (df): false grayscale image, segmentation result and ground truth labels for the Salinas HSI dataset. The former experiments achieves an Overall Accuracy of 0.93, whilst the latter achieves an OA of 0.9696.
Jimaging 07 00267 g004
Figure 5. (ac): Dependence of the RI wrt α for several values of λ , from 1 to 10 6 for Indian Pines, Pavia University and Salinas HSI datasets, respectively. The RI remains high, there is a peak around α = 0.85 in the former cases, while the Salinas HSI datasets achieve its best performance for α = 0.65 when λ is lower than 1. The number of bands obtained by the dimensionality reduction is stable wrt λ .
Figure 5. (ac): Dependence of the RI wrt α for several values of λ , from 1 to 10 6 for Indian Pines, Pavia University and Salinas HSI datasets, respectively. The RI remains high, there is a peak around α = 0.85 in the former cases, while the Salinas HSI datasets achieve its best performance for α = 0.65 when λ is lower than 1. The number of bands obtained by the dimensionality reduction is stable wrt λ .
Jimaging 07 00267 g005
Figure 6. (a) Ground truth labels. (bd) Random seeds for the proposed procedure, chosen among the ground truth mask. From left to right: square seeds of dimension 3, 5, and 7 pixels. The squares are clipped in order to refer to the correct region.
Figure 6. (a) Ground truth labels. (bd) Random seeds for the proposed procedure, chosen among the ground truth mask. From left to right: square seeds of dimension 3, 5, and 7 pixels. The squares are clipped in order to refer to the correct region.
Jimaging 07 00267 g006
Figure 7. Result of the segmentation of Pavia Center using the ground truth labels as atlas. From left to right: image of the landscape in false colors, labels of the ground truth, final result. The latter panel shows that all the roofs are remarkably recognized, as the shadows they project on the ground. The vegetation is segmented with a very high level of precision. We reported the labels as reported in the database we used for these experiments: there are clearly some errors, since some classes (such as Shadows, Meadows and Bare Soils) refers to objects that are not the ones described by these labels.
Figure 7. Result of the segmentation of Pavia Center using the ground truth labels as atlas. From left to right: image of the landscape in false colors, labels of the ground truth, final result. The latter panel shows that all the roofs are remarkably recognized, as the shadows they project on the ground. The vegetation is segmented with a very high level of precision. We reported the labels as reported in the database we used for these experiments: there are clearly some errors, since some classes (such as Shadows, Meadows and Bare Soils) refers to objects that are not the ones described by these labels.
Jimaging 07 00267 g007
Figure 8. Segmentation results on the KSC dataset. From left to right: false color image, ground truth labels employed as seeds, and segmentation result. On the bottom of the images the legend associates the color to the labels.
Figure 8. Segmentation results on the KSC dataset. From left to right: false color image, ground truth labels employed as seeds, and segmentation result. On the bottom of the images the legend associates the color to the labels.
Jimaging 07 00267 g008
Figure 9. Overlay between the segmentation and original image (in grayscale) for the KSC dataset.
Figure 9. Overlay between the segmentation and original image (in grayscale) for the KSC dataset.
Jimaging 07 00267 g009
Table 1. Computational time in seconds.
Table 1. Computational time in seconds.
DatasetSizeTime
Pavia University 103 × 207,4001.08
Indian Pines 220 × 21,0250.13
Salinas HSI 204 × 111,1040.45
Table 2. Comparison between methods.
Table 2. Comparison between methods.
SalinasPaviaKSC
MethodO.A.PurityO.A.PurityO.A.Purity
FCM54.60.6341.90.4752.50.56
FDPC62.20.7144.20.5447.60.64
GWEEN65.30.7747.90.6249.80.69
K–MBC76.50.9365.90.9158.20.74
HyperRaWaSI93.80.9491.10.9097.80.98
Table 3. Performance behavior with respect to the label size.
Table 3. Performance behavior with respect to the label size.
SalinasPavia
SizeO.A.PurityO.A.Purity
389.90.9280.10.80
592.70.9482.70.81
793.80.9491.10.90
Table 4. Computational time in seconds.
Table 4. Computational time in seconds.
DatasetSizeTime
Pavia University 102 × 783,6405.24
KSC 176 × 314,3683.31
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Aletti, G.; Benfenati, A.; Naldi, G. A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation. J. Imaging 2021, 7, 267. https://doi.org/10.3390/jimaging7120267

AMA Style

Aletti G, Benfenati A, Naldi G. A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation. Journal of Imaging. 2021; 7(12):267. https://doi.org/10.3390/jimaging7120267

Chicago/Turabian Style

Aletti, Giacomo, Alessandro Benfenati, and Giovanni Naldi. 2021. "A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation" Journal of Imaging 7, no. 12: 267. https://doi.org/10.3390/jimaging7120267

APA Style

Aletti, G., Benfenati, A., & Naldi, G. (2021). A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation. Journal of Imaging, 7(12), 267. https://doi.org/10.3390/jimaging7120267

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop