Next Article in Journal
Oil Spill Detection in Quad-Polarimetric SAR Images Using an Advanced Convolutional Neural Network Based on SuperPixel Model
Previous Article in Journal
Mobile Laser Scanning Data for the Evaluation of Pavement Surface Distress
Previous Article in Special Issue
An Object-Based Strategy for Improving the Accuracy of Spatiotemporal Satellite Imagery Fusion for Vegetation-Mapping Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Source and Multi-Temporal Image Fusion on Hypercomplex Bases

1
Geoinformatics Department, Munich University of Applied Sciences, Karlstraße 6, D-80333 Munich, Germany
2
German Aerospace Center (DLR), Earth Observation Center (EOC), Oberpfaffenhofen, D-82234 Weßling, Germany
3
Karlsruhe Institute of Technology (KIT), Institute of Photogrammetry and Remote Sensing (IPF), Englerstr. 7, D-76131 Karlsruhe, Germany
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(6), 943; https://doi.org/10.3390/rs12060943
Submission received: 15 January 2020 / Revised: 9 March 2020 / Accepted: 10 March 2020 / Published: 14 March 2020
(This article belongs to the Special Issue Advances in Remote Sensing Image Fusion)

Abstract

:
This article spanned a new, consistent framework for production, archiving, and provision of analysis ready data (ARD) from multi-source and multi-temporal satellite acquisitions and an subsequent image fusion. The core of the image fusion was an orthogonal transform of the reflectance channels from optical sensors on hypercomplex bases delivered in Kennaugh-like elements, which are well-known from polarimetric radar. In this way, SAR and Optics could be fused to one image data set sharing the characteristics of both: the sharpness of Optics and the texture of SAR. The special properties of Kennaugh elements regarding their scaling—linear, logarithmic, normalized—applied likewise to the new elements and guaranteed their robustness towards noise, radiometric sub-sampling, and therewith data compression. This study combined Sentinel-1 and Sentinel-2 on an Octonion basis as well as Sentinel-2 and ALOS-PALSAR-2 on a Sedenion basis. The validation using signatures of typical land cover classes showed that the efficient archiving in 4 bit images still guaranteed an accuracy over 90% in the class assignment. Due to the stability of the resulting class signatures, the fuzziness to be caught by Machine Learning Algorithms was minimized at the same time. Thus, this methodology was predestined to act as new standard for ARD remote sensing data with an subsequent image fusion processed in so-called data cubes.

Graphical Abstract

1. Introduction

The introduction addresses the necessity for Analysis Ready Data (ARD), presents the Multi-SAR System, an evaluated framework for the production of polarimetric ARD data, and how the Multi-SAR System is expanded with a new approach for image fusion for optical and SAR data. Finally, we formulates the research questions which will be answered in the following.

1.1. Necessity for Analysis Ready Data

Earth Observation (EO) is turning more and more to ‘Big Earth Data’ [1]. By the end of 2018, the Copernicus Sentinel products had a total data volume of 10 PiB with a daily average growth of 15 TiB in November 2018 [2]. The complexity of data processing, computational power, and storage increases with the data volume. This means that even sophisticated users of EO data invest more time and effort into data preparation than into data interpretation. The idea is to support EO users to transform the immense amounts of data into useful information and decision-ready products [3]. This challenge can be met with the provision of pre-processed EO data to a minimum set of requirements and their organization in a standardized, consistent, and transparent manner. It requires the possibility of an immediate analysis without additionally user effort, interoperability, and processing time as well as open and easy data access. The specifications are regulated within the Committee on Earth Observation Satellites (CEOS) for ARD [4]. ARD products representing surface reflectance are commonly used and their custom pre-processing workflow and provision is already standardized. Since 2008, the United States Geological Survey (USGS) provides the Landsat data as L 1 T analysis-ready products in a single, universally accessible, centralized global archive [5]. The focus of the Swiss Data Cube (SDC) is a consistent and comparable data processing and storage of different sensors. Therefore, they use the same framework for data discovery, downloading, and processing of Landsat as well as of Sentinel-2 ARD products [3,6]. The provision of ARD products representing radar backscatter started with the continuous and global coverage of Sentinel-1 imagery. The cloud based platform of Google Earth Engine provides processed Level-1 Ground Range Detected (GRD) Sentinel-1 data. The pre-processing contains border and thermal noise removal, radiometric and geometric calibration using Shuttle Radar Topography (SRTM) and Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) Digital Elevation Model (DEM), and conversion to decibel. It is implemented by the Sentinel Toolbox and allows a subsequent processing in Python and JavaScript languages [7]. Truckenbrodt et al. [8] analysed the capability of possible software solutions like the Sentinel Application Platform (SNAP) and GAMMA for the automatic production of Sentinel-1 ground range detected ARD products. He concluded that SNAP is a convenient and user-friendly graphical user interface and additionally, it is open source. Anyway, the error messages were difficult to interpret without prior knowledge and especially the complex Java structure complicates error search. The basic command line interface of licensed GAMMA required experiences in application and development of new workflows. Therefore, it was very robust and fast. The analyses of Ticehurst et al. [9] went beyond the usage of Sentinel-1 radar backscatter data and involved equally polarimetric radar data. She developed ARD products from SAR including normalized radar backscatter ( γ 0 ), eigenvector-based dual-polarization decomposition, and interferometric coherence based on SNAP. Her feasibility study showed a good agreement between SNAP and GAMMA with minor systematic geometric and radiometric differences which can be neglected.
The above mentioned backscatter ARD products bases mostly on Sentinel-1 GRD data. Though, CEOS demands Level-1 Single Look Complex (SLC) ARD products in order to exploit similarly the polarimetric information. This requires a standardized provision of polarimetric radar ARD data including speckle filtering, polarimetric decomposition, and normalization of the data in order to reduce their storage demand [10]. We present the Multi-SAR System, an evaluated framework for the production of polarimetric ARD data. The Multi-SAR System contains ortho-rectification, Kennaugh decomposition, multi-looking, radiometric calibration, image enhancement using the multi-scale multi-looking (MSML) approach or Schmittlets, and normalization based on hyperbolic tangent function for the final (almost) lossless data compression according to References [11,12]. The system can be used for the processing of all SAR data in level 1 (focused, slant range, single-look, complex) independent of their geometric, radiometric, and polarimetric resolution. Its main benefit is the usage of the same workflow and the same parameter configuration file independent of the different SAR input data formats and interfaces. The final result is a SAR image in level 2 represented in a consistent ARD format. It is implemented at the German Remote Sensing Data Center (DFD) within the German Aerospace Center (DLR) at Oberpfaffenhofen.

1.2. Applications Based on Multi-SAR System

With respect to the preprocessing of SAR images acquired by different sensors in different modes a first idea how to establish a consistent frame for the ARD provision was mentioned in Reference [13], where the Kennaugh elements were chosen to perform a robust change detection on wetlands. Already two years later, the normalized Kennaugh elements were introduced as a consistent frame for the fusion of multi-sensor and multi-polarized SAR data allowing for simple data analysis with standard and open-source software, effective data compression, and even change detection in combination with advanced locally adaptive filtering techniques [11]. Refining these filtering techniques by multi-directional kernels led to the Schmittlet approach that even enabled the analysis of image texture in the best available (total) intensity image [12]. The change detection capability was used for mapping wind throw areas in mid-European forests from TerraSAR-X images [14]. The usefulness of the normalized Kennaugh elements in the context of wetland remote sensing especially in semi-arid regions was proven several times using TerraSAR-X, RADARSAT-2, and even archived data of RADARSAT-1 and ENVISAT-ASAR [15,16,17]. Due to its flexibility concerning the input data, different polarization combinations could be evaluated at the same time for surface water monitoring purposes [18]. The advanced preprocessing finally enabled an almost daily estimation of the water gauge in the Forggensee in southern Germany via flooded area [19]. Furthermore, the Kennaugh decomposition was successfully utilized in coastal applications for the mapping of bivalve beds on exposed intertidal flats from dual-co-polarized TerraSAR-X data [20]. Going further north, the normalized Kennaugh elements excelled as input feature for classification of Tundra environments in northern Canada. The relation between snow depth, topography, and vegetation was studied on TerraSAR-X time series preprocessed to Kennaugh elements [21]. Even for the understanding of snow covered sea ice processes from TerraSAR-X the Kennaugh preprocessing showed up to be essential [22]. Regarding glaciered areas, the polarimetric information captured in Kennaugh elements helped to detect the supraglacial meltwater drainage on the Devon Ice Cap, Canada. Alpine glaciers were also analysed by normalized Kennaugh elements—a study evaluated the correlation between ground-based radar measurements and space-borne multi-polarized SAR images over the Vernagtferner in the Oetztaler Alps, Austria, [23], the other study evaluated time series of Sentinel-1 data over several summers in order to discriminate wet snow from firn as a first step towards a mass balance estimation from space [24]. Apart from natural landscapes the Kennaugh elements provided a valuable input for the mapping of slums from polarimetric SAR data [25]. In combination with the Schmittlet index and a highly sophisticated, multi-scale classification approach—called histogram classification—the discrimination of formal from informal settlements by multi-aspect SAR images became feasible [26]. The histogram classification which bases on the similarity measure for the comparison of two discrete local histograms was further studied for typical land cover classification based on reflectance data only [27]. Recently, a very promising new approach was published how to generate Kennaugh-like elements from optical multi-spectral data [28] in order to perform a SAR-sharpening gathering typical characteristics of optical and SAR images in one layer. Though, the sharpening of SAR by the fusion of multi-mode images delivered excellent results, the fusion of SAR and Optics still showed some weaknesses, for example, the combined use of SAR and Optics without radiometric fusion still was superior to the fused data set [29]. Additionally, it was restricted to four channel images like the 10 m -layers of Sentinel-1 or the typical channels of an aerial image. Thus, the fusion of four channels from SAR and four channels from Optics resulted in four channels again. Although the fused channels showed the typical characteristics of both inputs, they did not provide more degrees of freedom which seems to be essential. Finally, the multiplicative fusion ignores the signal-to-noise-ratio of the input channels and thus, possibly amplifies noise contained in the input channels instead of reducing it. A parallel study [30] proved that the Kennaugh-like provision of multi-spectral data from Optics has advantages over the standard data preparation in reflectance values when using the special multi-scale histogram classification algorithm. In summary, the remaining problem can be seen in the reduction of the image channels when the images are fused instead of an orthogonal transform into Kennaugh-like elements.

1.3. Multi-SAR System with Image Fusion

A step forward to the production of ARD data would be the image fusion of surface reflectance and polarimetric ARD data. Image fusion creates new images that are more suitable for the purposes of human visual perception and object recognition. The detailed overview of existing image fusions in References [31,32] showed impressively that the development of new methods is an ongoing process. The image fusion techniques can be categorized in spatial domain and transform domain fusion [33]. Ha et al. suggested two additional groups for categorization containing optimization approaches like Bayesian data fusion and other techniques like adaptive methods, smoothing filter-based intensity modulations, and spatial and temporal adaptive reflectance fusion models [31,34]. Due to the simplicity and the majority use of the first two mentioned groups, we neglect the two latter groups. The spatial domain fusion contains among others the Intensity-hue-saturation transform [35,36], component substitution [37,38], colour composition, principal component analysis (PCA) [39,40], and high-pass filtering [39,41]. It bases on pixel level processing [42]. Their characteristic is the usage of simple fusion rules and hence a low computing time. The fused images contain their high spatial quality and preserve their spectral information content. Though, the disadvantage is the spectral degradation which reduces the contrast levels in the fused image and hence could affect negatively the classification results [43,44]. The transform domain fusion comprises multi-resolution [45], wavelet transform [46,47], Laplacian and Gaussian pyramid [48], Bandelet Transform, Contourlet [49,50], and Curvelet [51,52]. The Fourier transform of the entire image into the frequency domain overcomes the limitations of the spatial domain methods. As it enhances the properties of the image like brightness, contrast, or distribution of gray levels, it provides its higher spectral content. Additionally, these methods preserve more spatial features and have a better signal to noise ratio compared to pixel-based approaches. Though, the transform domain methods are more complex than approaches of the spatial domain and leads to inadequately fusion images when applied on images from different sensors [43,44,53,54].
Since few years, deep learning has achieved great success in image fusion and can be regarded as a new categorization besides the spatial and transform domain. Its beginning were already in the 1960s as Ivakhnenko and Lapa published their first general, working learning algorithm for supervised deep feedforward multilayer perceptrons [55]. The term ’deep learning’ was for the first time introduced to the machine learning community in 1986 [56]. But first the increase and the availability of amount of labelled data helped deep learning to its breakthrough. Since then, it is used more and more in computer vision and image processing for classification and segmentation [32]. The approach is based on artificial neural networks like Convolutional Neural Networks (CNN) [57,58,59], Convolutional Sparse Coding [57,60], or Stacked Autoencoders [61]. The main advantage of this approach lies in its ability to outperform nearly every other machine learning algorithm. It supplies improvements for unstable and complex problems [32,57]. In addition, the same neural network can be applied to many different applications and data types. It is executed in the pixel level processing such as the spatial domain fusion [42]. In the same way, multi-source data can be fused using the CMIR-NET (e.g.) that learns from two separate, but labelled data sets [62]. Compared to other techniques, a high performance can only be achieved with a very large amount of data. As the model of the neural network must be trained, its accuracy depends on quality and size of the training data. Though, the most important disadvantage is that the output is not explainable as the prediction is not traceable. In a nutshell, the strength of the machine learning methods can be seen in the robust handling of more or less stochastic variations in labelled data. In contrast to that, our preprocessing approach for data fusion is simple, deterministic, robust, fast, and completely independent from labelled data.
The aim of the Multi-SAR System is to reduce stochastic variations in SAR data combinations to a minimum so that simple threshold classification is already enough to detect temporal changes [11]. With respect to the up-to-date machine learning approaches the introduction of Kennaugh-like elements—according to the Kennaugh elements of the Multi-SAR System—is expected to produce most stable physical measures in order to reduce the stochastic in the class assignment. In consequence, the remaining variation to be captured by the machine learning methods is significantly lower and therefore, the results based on Kennaugh-like elements are expected to be even better than using standard reflectance values. Therefore, the Multi-SAR System is expanded to the fusion of SAR and optical images. The current method uses hypercomplex bases which produces Kennaugh-like elements for optical data and hence enables the fusion of SAR and optical data to one image in a second step. It is performed on pixel level with an optional spatial or temporal filtering as known from the Schmittlets [12] implemented in the Multi-SAR System. We summarize the main points explained above in five research question to be answered in the article:
  • Is there a flexible descriptor for multi-spectral data with more than four channels that delivers Kennaugh-like elements as utilized in the Multi-SAR System?
  • How can multi-source images (e.g., from SAR and Optics) which are transformed to a common basis be fused without loss of information?
  • Is it possible to consider not only spectral or polarimetric, but also further dimensions such as “temporal” or “spatial” in a similar way?
  • Does the fusion of multi-source data really produce more stable class signatures than the simple channel combination?
  • How does a reduction of the image bit depth due to storing reasons effects the stability of the class signatures?
The methodology section will introduce the novel data fusion approach as well as the methods for its evaluation. The result section illustrates selected examples which are interpreted in the discussion section before the main conclusion are formulated.

2. Methodology

This section comprises the methodology needed for the generation of a consistent data frame, for the subsequent data fusion, low-loss compression, and finally, the generation and validation of stable class signatures as input to machine learning algorithms.

2.1. Consistent Data Frame

The basis for any data fusion was a consistent data frame in which each set of layers can be projected. With respect to polarimetric SAR data, the Kennaugh decomposition was first renewed in Reference [13] and then extended to any set of multi-polarized SAR data available nowadays in Reference [11]. The calculation of Kennaugh-like elements even from multi-spectral data in order to sharpen SAR images was recently published in Reference [28]. The core of this technique is an orthogonal transform by a matrix A with following characteristics:
  • a i , j = 1 n i , j [ 1 , n ]
    All entries of A share the same absolute value given by the dimension of A which is needed to fulfil the orthogonality requirement. This also implies an equal weight of all input channels corresponding to a uniform look factor.
  • j = 1 n a 1 , j = 1
    The first row of A denotes an equally weighted sum over all input channels giving the total intensity (SAR) or reflectance (Optics), respectively. This entry will be necessary for the normalization of the complete set of Kennaugh-like elements later on.
  • j = 1 n a i , j = 0 i [ 2 , n ]
    All other rows composes of as many negative as positive entries, thus their sum and their expectation value consequently equal to zero. Zero means “no information” in this element, whereas any deviation from zero indicated spectral or temporal information, respectively.
  • A T = A 1
    The matrix A is orthogonal. The inverse transform thus is simply given by its transposed matrix of A. As the matrix A is predefined and not estimated for each data set separately as usual in PCA (e.g.), it is guaranteed that the transform can simply be inverted.
  • A T · A = E
    The transposed matrix neutralizes the transform. This characteristic directly results from the orthogonality stated above and underlines that the basis change preserves information. The transform therefore only introduces new coordinate axes in the feature space.
For instance, the transformation matrix to derive two Kennaugh elements from twin-polarized radar measurements can be described as follows according to [11]:
K t w i n = 1 2 R H H + R V V R H H R V V = 1 2 1 2 1 2 1 2 · R H H R V V = A · R H H R V V
Similar formulations can be found for dual-cross-polarizations with intensity measurements exclusively. The quantum leap in comparison to other decomposition strategies lies in its flexibility and the strict distinction between reliable reflectance (or intensity) and often noisy phase information, thus in its radiometric stability. From a mathematical point of view, the matrix A defined above is most interesting: due to its symmetry, the matrix is its own inverse, that is, A = A T = A 1 and therewith, A 2 = A · A = E produces the neutral element in multiplication and A A 1 = A A = 0 likewise in addition. These characteristics result from the special choice of the new axes, cf. Figure 1 in the following.

2.1.1. The Complex Basis

More generally, the transform can be expressed via a linear transform of a two-element vector, where the two columns of the transformation matrix span one basis of the complex plane C .
K 2 = C · R 2 = 1 2 1 1 1 1 · R 2
Due to that reason the transformation matrix is now denoted by C. It linearly transforms the vector of reflectance measures with r i 0 into Kennaugh elements with the well-known properties, c.f., [11]. Thanks to its orthogonality, this transform restricts to a rotation and an inversion of the incident axes as shown in Figure 1. The scale remains untouched, that is, the transform does not amplify or damp the original signal. Even the relative angles between the channels are equal in order to prevent additional correlations.

2.1.2. The Quaternion Basis

The transform given above is only applicable to two-dimensional spaces. Regarding dual-polarized acquisitions with fixed phase measurements or the up-to-date compact-hybrid mode, four channels are available in Kennaugh elements. Beyond SAR, most airborne camera or scanning systems provide a four-channel image containing Blue, Green, Red, and Infrared. For this case, an extended approach was already published in Reference [28] that formulated a four-by-four matrix with the required characteristics. For conformity reasons, an equivalent matrix is now created by the substitution of each element in C (see Equation (2)) by another complex number. Referring the coefficients to numbers, the four-by-four matrix Q is created, which constitutes the Quaternion basis Q .
K 4 = Q · R 4 = 1 2 C C C C · R 4 = 1 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 · R 4
The only difference to the A matrix published in Reference [28] is the order and orientation of the single dimensions which is—by the way—irrelevant for its application. This matrix has already been used for the fusion of images from Sentinel-1 and Sentinel-2. The inspection of the polarimetric information content after the image fusion proofed that the polarimetric information of the input images was retained or even complemented [28].

2.1.3. The Octonion Basis

In order to serve higher-dimensional problems in the same way, the elements 1 or 1 of matrix Q (see Equation (3)) again are replaced by complex entries leading to the Octonion basis O with eight dimensions.
K 8 = O · R 8 = 1 2 Q Q Q Q · R 8 = 1 2 C C C C C C C C C C C C C C C C · R 8
Eight channels are especially needed for the fusion of Sentinel-1 (dual-polarized) and Sentinel-2 ( 10 m pixel) images when both, the polarimetric and the spectral information, shall be preserved. The article cited above was titled “SAR sharpening” and thus, only cared for the polarimetric information content [28]. Due to the reduction of the dimensionality from eight to four, a minimum data loss is indisputable. Using the new Equation (4) the eight input channels again result in eight output channels, thus the information preserving transform is guaranteed.

2.1.4. The Sedenion Basis

When exploiting the complete information contained in a Sentinel-2 image, thirteen channels are involved whereas channel 8 and 8 A only differ in their spectral and spatial resolution, that is, twelve independent channels. In combination with the four Kennaugh elements of Sentinel-1, the need for a sixteen channel transform arises which can be fulfilled by a fusion on the Sedenion basis S .
K 16 = S · R 16 = 1 2 O O O O · R 16 = 1 2 Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q · R 16
Thus, the Sedenion basis is able to reproduce the whole information content of a combined Sentinel-1 and Sentinel-2 image which will be of highest interest for monitoring purposes in the near future due to the high temporal sampling (at least once a week) and the easy data access at no charge.

2.1.5. Higher Order Bases

With view to coming satellite mission like EnMAP with a hyperspectral sensor on board and therewith, a much higher number of channels, the extension to high-dimensional spaces could be of interest. Due to the consistent spanning of higher-order vector spaces by simple substitution of the input elements, the extension to 32 dimensions (Equation (6)), to 64 dimensions (Equation (7)), to 128 dimensions (Equation (8)), or even more is possible without (theoretical) limitations.
K 32 = Z · R 32 = 1 2 S S S S · R 32
K 64 = V · R 64 = 1 2 Z Z Z Z · R 64
K 128 = H · R 128 = 1 2 V V V V · R 128
In the case of a minimally lower number of distinct input channels, the respective channel can simply be set to zero leading to a equally lower number of independent output elements as already stated in the context of the Kennaugh decomposition in Reference [11]. Thus, a very flexible, but completely consistent feature space for the fusion of arbitrary reflectance and even further rasterized GIS layers is given. At this point we can answer the first research question: multi-layer data sets can easily be transformed to Kennaugh-like elements using extensions of the complex plane defined in (2) to so-called hypercomplex bases.

2.2. Multi-Source Image Fusion

The following section summarizes the strategies for data fusion on the hypercomplex bases introduced above subdivided to the typically available data sets.

2.2.1. Polarimetric Fusion

The fusion of multi-polarized data sets to pseudo-polarimetric images was realized by the completion of redundant Kennaugh elements in one acquisition by the distinct Kennaugh elements of another acquisition [28]. This additive approach was successfully applied to data acquired in the pursuit-monostatic phase of den TanDEM-X mission for the generation of pseudo-quad-polarized images from two dual-polarized acquisitions [63].
K fused = K dual - co - pol + K dual - cross - pol
This symbolic formulation can be further refined by introducing the individual spatial resolution [28], that is, the look factor as weight of the input channels. For clarity reasons, only the equally weighted, and therewith the orthogonal case, was considered.

2.2.2. SAR Sharpening

Multi-polarized images generally suffer from a reduced spatial coverage and resolution. To avoid the reduced spatial sampling, an intensity image K pol of higher spatial resolution can be introduced in the same way pan-sharpening is done on optical imagery. In the simplest case, the total intensity of the multi-polarized image K pol is completely replaced.
K sharpenend = 0 K 1 K 2 K 3 pol + K 0 0 0 0 int = 0 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 · K pol + 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 · K int
Though exemplarily shown by a four-channel image, it can be transferred to any polarimetric combination without any limitation. The intensity layer is not restricted to SAR intensities, but can even be given by the panchromatic channel of an airborne camera system as proven in Reference [28].

2.2.3. Spectral Fusion

The fusion of multi-spectral data requires the primary transform from reflectance values to Kennaugh-like elements as already presented in Reference [28]. If the channels are drawn from different sensors R a and R b , the matrix multiplication can be divided in two additive parts.
K fused = 1 2 C 0 C 0 R a R b + 1 2 0 C 0 C R a R b = 1 2 C C R a + 1 2 C C R b
Due to simplicity, the approach is demonstrated in four-dimensional space, but without any limitation of the generality.

2.2.4. SAR-Optical Fusion

The problem occurring when fusing multi-polarized SAR and multi-spectral optical data lies in the combination of already transformed Kennaugh elements K SAR in the case of SAR and reflectance values R OPT in the case of optical images. We simply circumvented this problem by replacing C 2 · R a = K SAR in Equation (11) and got following equation.
K SAR & OPT = K SAR K SAR + 1 2 C C · R OPT = K SAR K SAR + K OPT K OPT
The consistent derivation explained in Section 2 still guarantees an orthogonal transform of the original features space even if the transform is split into two independent steps as described in Equation (12).

2.2.5. Image Fusion with Arbitrary Layers

The property reflectance in Optics or intensity with respect to SAR essentially denoted that this input layer was subject to a ratio scale, that is, there was a well-defined origin and the ratio between entries was playing the key role. It further implied that the data range was restricted to positive values. Therewith, any layer in ratio scale could be transformed—in combination with others—to the Kennaugh-space. If the ratio scale was not assured and only a cardinal scale can be accepted because of negative entries (e.g.) the inclusion of this layer has to be done in a scaled version of the Kennaugh-like elements which will be presented in the following Section 2.4. The summary of this section answered the second research question: simple matrix multiplication allowed us to separate the calculation of Kennaugh elements from SAR and Kennaugh-like elements from Optics and to fuse the transformed elements afterwards. The information retain was guaranteed because of the orthogonality of the transform. The fusion was performed in an additive way which considers the signal-to-noise ratio. Furthermore, the Kennaugh elements from SAR were well-known for their robustness towards noise.

2.3. Multi-Temporal Image Fusion

The multi-source image fusion mentioned above was designed to fuse images of different sensors, but taken at (approximately) the same time. This section introduces a methodology how to generate multi-temporal Kennaugh-like matrices from multi-source Kennaugh-like elements by a matrix multiplication from the right with one of the orthogonal matrices given above. Instead of multi-temporal and multi-spectral dimensions, multi-row and multi-column measurements could be processed in the same way in order to represent spatial characteristics, for example, as supplement or even substitute to spatial convolutions in machine learning algorithms.

2.3.1. Change Detection

The simplest case is the comparison of two images, in general a pre-event image at time t 1 and a post-event image at time t 2 . The multi-temporal Kennaugh matrix was calculated by a right multiplication with the basis of the complex plane C which results in the sum and the difference of the two input images.
K 4 × 2 = Q · R t 2 R t 1 · C = K t 2 K t 1 · C = 1 2 · K t 2 + K t 1 K t 2 K t 1 T
This conformed to the normalized intensity difference in Reference [11] and the additive formulation of the differential Kennaugh elements envisaged in Reference [64]. As the first element of K referred to an intensity measure, the requirements for a subsequent normalization were fulfilled. The quaternion basis Q for the multi-source fusion was chosen without limiting the generality: the number of dimensions in temporal and spectral direction were completely independent.

2.3.2. Gradients and Curvature

Considering more than two images, for example, four dates t 1 , t 2 , t 3 , and t 4 , more descriptors than sum and difference could be evaluated. The temporal combination then was realized by the quaternion basis Q likewise resulting in a four by four matrix of "differential" Kennaugh-like elements.
K 4 × 4 = Q · R t 4 R t 3 R t 2 R t 1 · Q = 1 2 · K t 4 + K t 3 + K t 2 + K t 1 K t 4 K t 3 + K t 2 K t 1 K t 4 + K t 3 K t 2 K t 1 K t 4 K t 3 K t 2 + K t 1 T
The matrix K 4 × 4 could be interpreted as follows: the first column refers to the sum of all Kennaugh elements, that is, to a most stable mean image or a low pass filter. The second column highlighted high frequency gradients or a high pass filter. The third column calculated the mean temporal gradient or a band pass filter. The fourth column produced the temporal curvature. All descriptors were given in critical sampling, but representing the whole information contained in the source products due to the orthogonality of the transformation matrix.

2.3.3. General Time Series Analysis

The Sentinel-1 mission with its repeat pass acquisitions every six days delivers sixteen images for a quarter of a year. In this case, the temporal fusion could simply be replaced by the Sedenion basis S without limitation of the generality.
K 4 × 16 = Q · R t 16 R t 1 · S = K t 16 K t 1 · S
With respect to longer time spans, hypercomplex bases of higher order were also possible, for example, the Z matrix for half a year or the V matrix for a whole year, see Section 2.1.5. The inclusion of further image layers would require the substitution of Q at the left of Equation (15) and therefore, led to an increase of the number of rows, that is, longer K vectors. In any case, the concept guaranteed that the first element always consists of an intensity measure that could be used for the normalization of all other elements. To conclude, we could affirm the third research question: it was possible to introduce the time as new dimension in the same way that multi-source images were combined. Thus, the consistency of the framework was guaranteed. The identical procedure allows equally the inclusion of spatial features like multi-row and multi-column image chips for a robust description of spatial patterns.

2.4. Data Scaling Approach

This section treats the possible scaling variants of the defined elements which can be linear, logarithmic, or even normalized to a closed range with respect to the application. The origin of the elements—vectors from multi-source or matrices from multi-temporal data—is completely irrelevant.

2.4.1. Linear Scale

All image data from optical remote sensing satellites were delivered in reflectance values, mostly multiplied by a calibration factor in order to guarantee integer value. SAR data in contrast was given as real- and imaginary part in single look slant range products or amplitude data in the case of ground range or geocoded products. The conversion from the delivered scaling to intensity values could be found in the respective product specifications. The data range of linearly scaled reflectance data is given by R l i n e a r [ 0 , ] whereas the mode of the exponential distribution could be found slightly above zero. The adequate storing in digital numbers therefore was quite difficult which resulted in a high radiometric sampling of usually at least 16 bit nowadays.

2.4.2. Hyperbolic Tangent Normalized Scale

In order to reduce the bit depth without noteworthy loss of information a normalization approach was introduced in Reference [11]. The core of his hyperbolic tangent (TANH) normalization was a division of all Kennaugh elements except K 0 by the total reflectance contained in K 0 —the first element of the vector or the first element of the main diagonal, respectively,—and the reduction of K 0 to a reference intensity I R e f similar to differential Kennaugh elements [13] where the pre-event intensity served as reference for the post-event intensity.
k = K 0 I R e f K 0 + I R e f K 1 K 0 K m K 0 T
The resulting normalized elements shared a closed data range of k [ 1 , + 1 ] which simplified not only the interpretation based on Support Vector Machines (e.g.), but also the conversion to digital numbers. The standard sampling of 16 bit enables a data range of ±48 dB with a maximum radiometric sampling of 0.0003 dB. The reduction to 8 bit still covered a range of ±24 dB with a maximum radiometric sampling of still 0.07 dB. The range where the radiometric sampling was better than 1 dB is given by ±41.7 dB and ±17.5 dB, respectively. Thus, the loss of information should be negligible in comparison to a conversion of the linearly scaled variables. The possible reduction of the radiometric sampling to only a few bins enabled new classification approaches like the histogram classification [26].

2.4.3. Logarithmic Scale

The reason why TANH normalized values could be expressed in decibel, thus in logarithmic scale was pointed out in Reference [11]. It could be shown that the area hyperbolic tangent of the normalized elements gave the variables in logarithmic scale which differed from the unit decibel only by a constant factor which was the same for all normalized Kennaugh-like elements including K 0 .
k d B = a t a n h k d B · 20 d B ln 10 = a t a n h k d B · 8.68588963806504 d B
The logarithmic scale was suitable when layers in cardinal scale should be included on the one hand, see Section 2.2.5. On the other hand, it resulted in layers with more or less a normal distribution which was required by many parameter-based classification approaches like the maximum likelihood classification. Regarding data compression, the logarithmic scaling was superior to the linear scaling because of the variable sampling rate (high around zero and lower towards infinity), but inferior to the normalized scaling because of its open data range k d B [ , + ] .

2.5. Evaluation Approach

The evaluation of data fusion approaches was not a trivial task. Most scientists conducted a whole classification for each data set as well as for the fused one separately and compared the classification results. The improvement was expected to show up in a lower number of misclassifications using the fused data set. This was truly an adequate approach for a predefined combination of data set, classification task, and algorithm. As we were aiming at a universal approach for the data preparation, it was neither reasonable to focus on one application case nor possible to cover all possible applications. Therefore, we decided in favour of following validation strategy composed of traditional and novel methods.

2.5.1. Visual Inspection

Though very subjective, the visual inspection of imagery still was the most reliable interpretation technique with view to extremely complex or high-dimensional tasks. The high dimensionality was present in the multi-source and multi-temporal fusion. Regarding the Quaternion transform, four spectral (or polarimetric) dimensions met four temporal dimensions, which resulted in sixteen elements. The semantic subsampling to a few land cover classes was also very difficult, because the classes had to represent single temporal developments which could be highly variable [15]. Therefore, the multi-temporal fusion was validated by a visual inspection of RGB (Red-Green-Blue) compositions of the single multi-spectral elements along the temporal dimension which is provided by the single columns of the matrix. To support the conclusion drawn from visual inspection, the first five moments of the resulting distribution were calculated along the temporal and spectral dimension separately. They were normalized to a linear deviation, that is, the mean absolute deviation for the first moment, the standard deviation for the second moment, the third root of the kurtosis, and so forth.

2.5.2. Class Signatures

The multi-source image fusion resulted in a feature vector which was much easier to interpret. The signatures of different land cover classes were expected to by quite stable within the region of interest and thus could be described by the distribution of the belonging pixels in the normalized Kennaugh space. Correlations between the single Kennaugh elements which are taken into account in some parametric classifiers like the Maximum Likelihood Estimation (MLE) are of minor importance to modern classifiers like Random Forest (e.g.) or the multi-scale histogram classification [30], that is, due to the immense computing demand when evaluating multi-variable histograms with common statistical measures like mutual information. Our study will consider both: standard MLE and similarity estimation.

2.5.3. Maximum Likelihood Contingency Evaluation

One traditional method to compare class signatures with individual signatures was given within MLE. This parametric approach described the class signatures by multi-variate normal distributions defined via mean vector μ and covariance matrix ς . The discriminator for each pixel vector ψ in logarithmic scale [65] then unfolded to
S M L E = l n ς ψ μ T · ς 1 · ψ μ
Each pixel in the reference data set was attributed to the class with the maximum likelihood. After this step, a contingency matrix of the training areas—similar to the confusion matrix for classifications—was calculated. As this procedure had to be performed for varying radiometric resolutions, that is, different binnings, only the total accuracy and the κ -coefficient were reported as indicators for the quality of the class description.

2.5.4. Similarity of Signatures

As parametric approaches like the MLE always imply a certain distribution, non-parametric methods evaluating the empirically derived distributions exclusively such as the Correlation Coefficient, the Kullback-Leibler-Divergence, the Cross-Entropy, or the similarity were preferred nowadays. The similarity was designed especially for the comparison of discrete histograms from normalized remote sensing data with or without consideration of the inter-channel correlations [26]. For the case of normalized elements without noteworthy correlation, it unfolded to
1 S p , q = j = 1 m i = 1 n q i , j 2 p i , j m
where n denotes the maximum number of bins and m the number of dimensions compared, that is, Kennaugh-like elements. The root does not result directly from the derivation, but it balances the varying number of input elements when testing scenarios with varying numbers of input channels. The similarity had a closed data range of S p , q [ 0 , 1 ] like any probability value. For visualization purposes the unit decibel S p , q [ d B ] = 10 · log 10 S p , q [ , 0 ] was preferable and will be used in Section 3.

2.5.5. Intra- and Inter-Class Similarities

Using the signatures defined in Section 2.5.2 and the similarity measure given in Section 2.5.4 intra-class and inter-class similarities were derived and compared in order to estimate the impact of the data fusion and scaling on the class separability. As the similarity could only be calculated for two signatures under study, the arithmetic mean of all possible combinations was chosen as representative value for the intra- and inter-class similarities. We expected high intra-class similarities and low inter-class similarities. In the optimal case, they do not converge until the radiometric sampling is very low, for example, binary. In addition to several diagrams that depicted the influence for each class separately, the mean gain over all classes was given in numbers. It composed of the difference between the mean intra-class and the mean inter-class similarity indicating the relative increase in class similarity. The mean similarities were calculated as arithmetic mean over all samples of the single classes. This equals the geometric mean of the incident class similarities, which means that it was very sensitive towards possible low outliers.

2.5.6. Gain of Intra- vs. Inter-Class-Similarity

In favour of a more comprehensive interpretation, the manifold results from the preceding section were rearranged and combined to a gain in similarity, that is, to which extent the mean similarity of signatures was increased or decreased. The gain was simply defined as difference between the mean logarithmic similarity of objects belonging to the same class over the mean logarithmic similarity of objects belonging to other classes. This reflected the logarithmic quotient of the respective similarities and could again be given in the unit decibel. It provided a measure equivalent to the total accuracy and κ -coefficient evaluated for the MLE comparison with mean class signatures, but derived from the similarity of individual signatures exclusively this time.

3. Results

The capability of the transforms shown in the previous section will by demonstrated in the following by selected examples. Understandably, the choice of examples was not sufficient to cover all possible applications, but enough to provide a rough impression of the potential inherent to the novel technique.

3.1. Multi-Temporal Image Fusion

The multi-temporal image fusion presented in Section 2.3 was applied to a data set consisting of four Sentinel-2 geocoded and atmospherically corrected L2A images over the “Murnauer Moos” near the Alps in southern Germany. The area is characterized by restored, quasi natural moor areas, agricultural land, forests, and built-up areas. Only the bottom-of-atmosphere (BOA) reflectance values contained in the 10 m -layers are considered in this example for reasons of simplicity. The almost cloud-free images were acquired on 2019-07-23, 2019-08-27, 2019-09-21, and 2019-10-26. True Colour and Colour Infrared images of the data stack were depicted in Figure 2. The reflectance measures were transformed to multi-spectral Kennaugh elements according to Equation (3) and then combined to multi-temporal elements according to Equation (14). The single rows of the resulting multi-spectral and multi-temporal Kennaugh matrix are displayed in Figure 3 gathered to multi-temporal RGB images, that is, each channel refers to one spectral Kennaugh element according to the indication right of the images.
The normalized moments (e.g., the standard deviation as compared in Reference [66]) along the temporal and the spectral dimension are listed in Table 1 divided in the four spectral or temporal channels remaining in the respective case.

3.2. Multi-Source Image Fusion

This section summarizes the results of the multi-source image fusion to one data set consisting of Sentinel-1, Sentinel-2, and ALOS-PALSAR-2 acquisitions for the three regions of interest (ROI) in Germany, namely the “Starnberg Lake” (ROI 1) in southern Germany, the “Ruhr Metropolitan Area” (ROI 2) in western Germany, and the “Potsdam Lake Land” (ROI 3) near the capital in eastern Germany. Detailed descriptions of the input data can be found in Table 2. All images were projected into the earth-fixed UTM coordinate system with a uniform pixel spacing of 10 m by 10 m resulting in slightly varying look factors according to [28] which is ignored in the current study for reasons of simplicity. The following figures are all reprojected to latitude and longitude in WGS84.
The SAR data were preprocessed by the Multi-SAR System [11] exclusively available at DLR (Oberpfaffenhofen) to geocoded, calibrated, and normalized Kennaugh elements. The multi-spectral data were provided in geocoded and radiometrically calibrated BOA reflectance values and transformed into Kennaugh-like elements according to the formulas given above. The study considers two cases: a fusion on the Octonion basis including Sentinel-1 ( K 0 , K 1 , K 5 , K 8 ) and Sentinel-2 ( B 2 , B 3 , B 4 , B 8 ) and a fusion on the Sedenion basis including ALOS-PALSAR-2 ( K 0 K 7 ) and Sentinel-2 ( B 2 B 8 , B 11 ). In all three test sites sample segments were manually selected in order to represent following land cover classes: (1) Water, (2) Forest, (3) Field, (4) Meadow, and (5) Urban. The class signatures were derived from labelled segments generated in a previous study [29].

3.2.1. Fused Images

The visualization of Kennaugh-like elements was not an easy task because of their central distribution where the deviation from zero in both directions indicated information. The common visualization comprises three reflectance channels: Red, Green, and Blue with a non-negative exponential distribution. Therefore, the input data sets were presented in the accustomed way: True Colour Images for multi-spectral images of Sentinel-2 and Bourgeaud decomposition (i.e., the intensity channels) for SAR images of Sentinel-1 and ALOS-PALSAR-2. Out of the collection of fused Kennaugh-like elements only a selection of three elements (indicated on the right site of each image) are depicted in logarithmic scale. A subset of the images for the first region of interest can be found in Figure 4. Selected subsets of the second and the third test site are illustrated in the same way in Figure 5 and Figure 6.

3.2.2. Signature Stability

The illustration of multi-spectral data with more than three channels was quite difficult. In optical remote sensing the four-channel information typically was divided in one true colour image (TCI) and one colour infra-red image (CIR) as illustrated in Figure 2 well being aware of the high redundancy. In order to generate an illustration of all eight or sixteen channels of the fused data set, histograms of three different scaling variants were computed. The linear scaling in the left column of Figure 7 and Figure 8 equals the standard methodology working on reflectance values (optical remote sensing) and covariance matrices (radar remote sensing), respectively. Due to the orthogonality of the transform between reflectance channels and Kennaugh-like [28] elements as well as between the covariance matrix and the Kennaugh elements [64], the statement drawn from the histograms was transferable. Elements in linear scale consist of one reflectance value (the total intensity) and several reflectance differences ranging from to + . The second variant considered in the middle column of Figure 7 and Figure 8 was the logarithmic scaling in the unit decibel (dB) which is quite usual in radar remote sensing and electronic engineering in general. This scaling adopted a multiplicative characteristic of the input data that was supposed to be composed of intensity quotients. The typical range of decibel values—though theoretically unlimited—restricted to −30 dB to +30 dB. The normalized Kennaugh-like elements displayed in the right column of Figure 7 and Figure 8 shared the unique property of a closed value range [ 1 , + 1 ] , which allowed for the most effective radiometric sampling which will be studied in the following section. In order to show the advantages of our approach two classes out of the five classes of the study which were not easy to describe with traditional methods due to their high variability were chosen for demonstration, namely “Forest” in Figure 7 and “Urban” in Figure 8.

3.2.3. Maximum Likelihood Contingency

The total accuracy and the κ -coefficient achieved by an MLE of the class assignment within the reference data are depicted in Figure 9 with respect to the ROI (line style), the input data (line colour), and the binning (right axis). The total accuracy attained 20 % for only one bin which was the accuracy to be expected when five equally distributed classes were assigned randomly. The κ -coefficient lowered to zero in this case underlining that the class assignment became purely random.

3.2.4. Signature Similarity

The similarity of two signatures was calculated via Equation (19). In default of symmetry, the similarity measure was not commutative, that is, the similarity of class 1 to class 2 in general differed from the similarity of class 2 to class 1. Therefore, both cases were considered. For clarity reasons only the mean similarities of the respective sample class were kept in Figure 10 for the Octonion fusion and Figure 11 for the Sedenion fusion. The diagrams were split in four parts:
  • upper left: the mean intra-class similarities of the individual object signatures
  • upper right: the mean intra-class similarities of the mean class signatures
  • lower left: the mean inter-class similarities of the individual object signatures
  • lower right: the mean inter-class similarities of the mean class signatures
All similarities were provided in logarithmic scaling in the decibel unit. The searched class was indicated by the colour according to legends contained in the respective figures.

3.2.5. Similarity Gain

The mean gain in class similarity by comparing the intra-class to the inter-class similarities is numbered in Table 3 for the fusion on the Octonion basis and in Table 4 for the fusion on the Sedenion basis. Though, the analysis was performed separately for the three test sites, the results were quite similar. Obviously, the Sedenion fusion with its 16 channels even raised the gain in comparison to the Octonion fusion with its 8 channels. As expected the normalized Kennaugh-like elements showed the highest gain for low radiometric sampling. The gain in the logarithmic scale was remarkably variable. The linear scale was only applicable with higher radiometric sampling. In the case of the Sedenion fusion (see Table 4) it only seldom exceeded the gain of normalized scaling. Regarding the Octonion fusion in Table 3, the linear scaling produced higher values from 20 bins upwards. In order to provide a more detailed view to the single classes under study, selected examples of the similarity gain per class are plotted in Figure 12.

4. Discussion

This section interprets the results shown above and draws the conclusions concerning the application of the novel data fusion approach. In order to avoid interfering influences from varying classification methods and parameters as studied by [67], we base the whole interpretation on statistical measures and the identical set of class signatures.

4.1. Multi-Temporal Image Fusion

The benefit from the multi-temporal image fusion becomes obvious when comparing Figure 3 with Figure 2. The standard method based on reflectance measurements in several spectral channels suffers from the high correlation of the input channels. Therefore, mostly brightness changes are clearly visible, see Figure 2. The changes in brightness are summarized in the temporal variation of the total intensity K 0 in Figure 3f. The multi-temporal combination of the spectral properties in Figure 3d indicates that the mean spectral information content is quite low what was expected when looking at the single images in Figure 2. The really characteristic changes for single land cover classes are all captured in the temporal variation of spectral Kennaugh elements. Agriculture land (e.g.,) shows up in diverse colours. Especially in Figure 3f,g the moor areas reveal a fine structure which was not visible at all in the standard display of Figure 2. The statistical evaluation in Table 1 reveals three points: (1) The spectral variability in the four acquisitions (left part of the table) is equally low for reflectance values, gets variable with the transformation to linear Kennaugh-like elements, and increases significantly with the scaling of the Kennaugh-like elements. (2) The temporal variability in the right part of the table is even lower for the single spectral channels. The transformation to Kennaugh-like elements causes an increase which is again intensified when applying the typical scaling. (3) All channels show comparably high values over all five moments, that is, the underlying distributions seem to be very complex which inhibits the use the simple parametric estimators. Thus, the use of non-parametric approaches like the similarity is highly recommended. A similar validation method was already used by [68] in order to evaluate the gain by the fusion of hyperspectral and panchromatic data. Due to the complexity of the landscape, the clustering to multi-temporal class signatures is a quite difficult task and therefore, will be subject to future studies. The use of enhanced clustering methods like self-organizing (e.g.) would go far beyond the scope of this article. According to [69], the multi-temporal combination promises up to 91% overall accuracy and 89% for the κ -coefficient with a relatively high thematic sampling of nine land cover classes when using enhacned classification methods. To sum up, the combination to multi-temporal Kennaugh-like elements highlights phenomena that were not apparent in the original reflectance data.

4.2. Multi-source Image Fusion

This section interprets the results of the image fusion on the Octonion and the Sedenion basis shown above. The quantitative evaluation will focus on the multi-source Kennaugh elements and take the signature stability and similarity over the classes provided in the reference data set into account.

4.2.1. Visual Inspection

The subset of the “Starnberg Lake” test site in Figure 4 depicts relatively smooth water and meadow areas in the optical true colour image whereas settlements appear much more heterogeneous. The SAR intensity images share the typical characteristics: smooth surfaces with respect to the wavelength are dark. The ALOS-PALSAR-2 image in Figure 4f reveals some wind effect that cause a higher surface scattering in the cross-polarized channel. Regarding the meadows, Sentinel-1 shows a strong texture (Figure 4d), but ALOS-PALSAR-2 relatively smooth areas (Figure 4f) due to the respective wavelength. The quite clear segments reported by the optical image are preserved and overlaid with the SAR texture when fusing both image sources on the Octonion basis in (Figure 4e) or the Sedenion basis (Figure 4g). Even the open water receives a certain texture that reflects the local wind situation affecting the polarimetric scattering behaviour as described above.
Figure 5 illustrates a subset of the “Ruhr Metropolitan Area” test site. The settlements are clearly distinguishable by their blue tone in the optical true colour image (Figure 5b). In the Sentinel-1 intensity image they cannot be discriminated from the surrounding forests and meadows. The ALOS-PALSAR-2 enables the distinction between meadows that appear dark due to their smoothness with respect to the wavelength and higher objects. The appearance of settlements depends on the orientation, that is, there are very bright spots, but also areas with a mean intensity that can hardly be discriminated from forests. The fused image unites characteristics of SAR and Optics. Settlements are mapped in yellow to red in both the Octonion (Figure 5e) and the Sedenion (Figure 5g) fusion according to their orientation. Agricultural land receives a slight texture that allows for a more detailed analysis of the crop. Water surfaces again appear in pale blue with a clear shape.
The most impressive example for the data fusion can be found in Figure 6 where a subset of the “Potsdam Lake Land” is illustrated. The clear boundaries visible in the optical image are combined with a refined SAR texture. Higher vegetation (e.g., forest) receives a fine and very diverse texture possible allowing for a more detailed analysis. Low vegetation is coloured in green tones according to the wavelength: with a higher variation in C-band (Figure 6e) and a lower variation in L-band (Figure 6g). Even the infrastructure is clearly mapped in the Sedenion fusion. Smooth water is exactly shaped in pale blue with only a slight pink texture in the cross-polarized channel of L-band, cf. Figure 6f,e. In summary, the fused images illustrate scattering characteristics which where not visible at all in the single acquisitions. As the visual impression is often misleading [65], the three test sites are validated quantitatively in the following.

4.2.2. Mean Class Signatures

The mean class signatures of the land cover classes “Forest” and “Urban” are provided in Figure 7 and Figure 8 for the fusion on the Octonion basis (a–c) and on the Sedenion basis (d–f), respectively. The elements in linear scale show only minor deviations from zero over all classes resulting in very slim peaks in the probability density functions (PDF), but with few very high and very low values. This effect even intensifies with an increase of input channels number from Octonion to Sedenion basis. The logarithmic scale provides diverse, but almost normally distributed elements, see middle column. The limitation to the range of [−30 dB, +30 dB] concerns only few values. Regarding typical noise levels of around 25 dB, the values exceeding the limits can be seen as noise. Some undulations become visible that indicate a higher radiometric sampling than provided by the input data. These undulations even increase in the normalized scaling, see Figure 8. The closed data range and the width of the distributions allow a lower radiometric sampling with only negligible loss of information which will be studied in the following section. The signatures over the three test sites appear quite similar. The fusion on Octonion basis provides more distinct histograms than the Sedenion fusion which can be referred to the different input data, see Section 3.

4.2.3. Maximum Likelihood Contingency

The total accuracy and κ -coefficient of the contingency matrix derived from a MLE class assignment proofs the advantage of the presented methodology. For a high radiometric sampling, the three methods deliver identical results. But as soon as the radiometric sampling is reduced to less than 10,000 bins, the class assignment based on reflectance values becomes more or less random (blue lines in Figure 9). The logarithmic Kennaugh-like elements deliver stable results even for a radiometric sampling with only 50 bins in the Octonion case (cf. Figure 9a,b) and even 25 in the Sedenion case (Figure 9c,d). Regarding the normalized scaling, the accuracy of the class assignment is still clearly over 80% both in the total accuracy and the κ -coefficient for 3 bit sampling with 8 bins. For lower sampling rates, the location of the bins seems to play a key role because the neighbouring entries vary by 50% and more. A binary image (1 bit) still provides a class distinction with a total accuracy of 30% to 40% and a κ -coefficient of 20% in some ROIs. This high variability for low sampling rates is confirmed by the signature similarity analysis in the following. With respect to the fourth research question we have to admit that Kennaugh-like elements seem to provide the same accuracy in the class assignment as traditional methods as long as the bit depth is sufficient, that is, 16 bit at least. With reduced bit depths, the novel descriptor surpasses reflectance values by far.
Due to the high diversity and possibility of data combinations different methods for individual data sets and varying accuracy parameters are reported in the literature. A combination of Sentinel-1 and Sentinel-2 using a object-based classifier promises an overall accuracy of 91% and a κ -coefficient of 89%. The classification of a THAICHOTE and Cosmo-Skymed fusion performed by SVM reaches an overall accuracy of 77% and a κ -coefficient of 67% whereas an object-based method delivers 82% for the overall accuracy and 75% for the κ -coefficient [70]. The combination of TerraSAR-X and Quickbird data fused by different methods leads to overall accuracies from 75% (reflectance layers) and 82% (including texture measures) to 91% using refined methods [71]. The latter value is confirmed by a study that uses an INFOFUSE-algorithm on WorldView-2 images, texture measures, and a digital surface model [72]. With respect to a simple maximum likelihood classification, the values drop down to 82% total accuracy and 80% for the κ -coefficient. Higher accuracy values are exclusively reported to the fusion of multi-scale data acquired by the same sensor, in general for the purpose of pan-sharpening. For the Enhanced Thematic Mapper (ETM) of Landsat, 82% overall accuracy and 79% for the κ -coefficient are reported in Reference [65] when using the maximum likelihood approach. When fusing the multispectral channels with the panchromatic channel of IKONOS by the SFIM or Gram-Schmidt approach, the overall accuracy reaches almost 99% with a κ -coefficient of 98% according to [73]. Comprehensively, the calculated accuracy measure depends on many different parameters [67]. One difference could be the usage of the contingency from the training samples in our case and supposably in the latter article in contrast to the confusion matrix from control samples.

4.2.4. Signature Similarity

The diagrams in Figure 10 and Figure 11 show the influence of the radiometric sampling on the class distinction. High values near zero in the intra-class similarity (upper part) and very low values in the inter-class similarity (lower part) should be guaranteed even with a very low number of bins. The distinction using linear elements on the Octonion basis is possible starting with three bins, see Figure 10a. On the Sedenion basis even a radiometric sampling with only two bins enables a weak distinction of the classes. Figure 11c shows an extraordinary high intra-class deviation of the mean signatures of the class “Water” which is not acceptable, but a clear effect of the generally low back scattering. Apart from that, the intra-class similarities seldom exceed −20 dB whereas the inter-class similarities can reach up to −90 dB which is excellent. Looking at the similarity in logarithmic scale in Figure 11b, the high sampling around zero results in several lobes in the inter-class similarity of the individual signatures for less than nine bins. An even number of bins obviously leads to a better class distinction. This effect was already observed in the MLE contingency analysis discussed above. The class assignment now is even possible to a certain degree with only two bins. The same applies to the normalized elements that can be found in Figure 10c for the Octonion basis and in Figure 11c for the Sedenion basis. The main difference to the similarity in linear elements becomes visible when comparing Figure 10a to Figure 11a: the similarity of linearly scaled signatures decreases with an increasing number of bins quite symmetrically in intra- and inter-class direction. This behaviour is independent from the class. The signatures of normalized elements show a stronger increase of around 10 dB in the inter-class direction than in intra-class direction proofing that the signatures similarity to signatures of the same class is ten times higher than the similarity to other classes which is essential for the classification based on normalized elements with reduced radiometric sampling. At this point, we can answer the last research question: the reduction from typically 16 bit to 8 bit images still assures excellent class signature using the new consistent multi-sensor framework. The radiometric sampling of 8 bit equals 256 bins. In this study, 100 bins already promise a class distinction of at least 10 dB and maximum 75 dB using individual signatures, see left edge of Figure 10c. The reduction to 4 bits (or 16 bins) still allows a maximum class discrimination power of about 40 dB. Switching to 3 bit images with only 8 bins further reduced the class distinction. But it is still measurable reaching 5 dB (or 80% in the MLE analysis) on individual signatures in Figure 11c which means that the inter-class similarity is three times higher than the intra-class similarity.
Only one study is reported in the literature that evaluates the classification accuracy in dependence on the compression [66]. This study states an overall accuracy of 45% using a maximum likelihood classification and a compression ratio of 1:50. In our case, this means that the original sampling of 16 bit is reduced to one fifth, that is, reduction from 65536 bins to 1311 bins or approximately 10 bit. The usage of a neural network classification method still promises 49%. The preparation in Kennaugh-like elements in logarithmic or normalized scale already reaches comparable similarity values with only two bins for most classes. With view to the maximum likelihood method whose accuracy is plotted in Figure 9, the total accuracy exceeds the border of 50% with 3 bits, thus already 8 bins. In terms of compression, this equals 8:65536 or 1:8192 for the radiometric compression exclusively.

4.2.5. Similarity Gain

The similarity gain of intra-class over inter-class similarity was not studied in the literature so far. Most publications of data fusion for land cover classification compare the accuracy before and after the fusion. For instance, this was done by [74] for the fusion of two SAR and one optical image in a neural network. The compound classification produced an increase from 84% to 92% in average. Regarding the ETM+ fusion, the accuracy of 62% of the original image reaches 66% at highest for different fusion algorithms [75]. Another study on the ETM+ fusion in combination with a supervised classification [76] reports on a rise from 83% to 90% using the fused image. Looking at higher spatial resolution, it seems to amplify the impact of the classification algorithm. The combination of Quickbird multispectral layers with its panchromatic layer classified by a maximum likelihood approach allows for a total accuracy of 58% before and 59% after the fusion which is an almost negligible gain of. Using the Euclidean Distance instead promises 91% after image fusion in comparison to 87% before [77]. The values provided by these studies can be compared to the accuracy values plotted in Figure 9. For a high radiometric sampling, we cannot state any gain using our fused and normalized images. But, with reduction of the radiometric sampling to 1000 bins, the overall accuracy when classifying the original images falls to values around 20% which is the expectation value for the random assignment of five equally frequent classes. Hence, for images with bit depths in-between 50 bins and 1000 bins we can state an increase from 20% to more than 90% which equals an increase by more than 70 percent points.
The gain of the intra-class over the inter-class similarity, that is, the class discrimination, is will be studied quantitatively in the following. The Octonion fusion in Table 3 indicates that the linear scaling is superior only for high radiometric sampling. For less than 20 bins, the gain is higher using logarithmic and normalized elements. With respect to 16 bins that equals a 4 bit image, the values for normalized elements still range around 20 dB which means that the similarity of samples belonging to the same class is around one hundred times higher than the similarity of samples belonging to different classes. Hence, the class discrimination is assured. The fusion on the Sedenion basis presented in Table 4 enables a gain in similarity of more than 30 dB which equals a thousand times the inter-class similarity. The gain using linear elements decreases almost continuously with the reduction of the radiometric sampling to zero. The gain achieved using the logarithmic or normalized elements only seldom falls below 10 dB, that is, the intra-class similarity is always ten times higher than the inter-class similarity independently from the radiometric sampling. The additional increase of the Sedenion fusion over the Octonion fusion is due to the higher number of channels and the completely different composition of the channels: Sentinel-1 with Sentinel-2 for the Octonion fusion and Sentinel-2 with ALOS-PALSAR-2 for the Sedenion fusion. The choice of the reference samples has another impact that cannot be denied. Anyway, the SAR-Optics fusion on the Sedenion basis would also allow the reduction of the radiometric sampling to a 3 bit image with 8 bins and still provides a class discrimination of factor ten which is unbeaten so far with respect to the low storage requirement. The images thus can be stored with 3 bit only, which enables the further compression with the typical spatial approaches available for most image formats.
Figure 12 illustrates the similarity gain for certain land cover classes separately. In almost all subfigures, be it for the Octonion fusion (left column) or the Sedenion fusion (right column), the normalized Kennaugh-like elements provide the highest similarity gain, see solid line. Only for the “Water” class in the Sedenion fusion (Figure 12b), linear elements seems to excel the logarithmic and normalized elements. This effect can be attributed to the very low response leading to unstable spectral properties which have to be covered by the classification algorithm. In the case of “Forest” a similar effect can be observed in the Octonion fusion (Figure 12c). The “Meadow” class indicates a negative similarity gain, that is, a loss accuracy using mean signatures in all three scaling variants, see Figure 12h. The high variability of the “Meadow” class in terms of soil conditions, soil moisture, and plant phenology might be one reason for that. With view to classification approaches, this means that traditional parametric methods like MLE are not sufficient to describe the “Meadow” class as one homogeneous syntactic class. Thus, the “Meadow” areas have to be divided in several syntactic classes with a homogeneous appearance or modern machine learning approaches have to be utilized for the automated interpretation. Due to the low radiometric sampling needed for an adequate class distinction, the pre-processing to normalized Kennaugh-like elements even supports the usage of Machine Learning: (1) The images are small and easy to store. (2) The bins replace the quantiles often as extended feature space. (3) Inter-channel correlations play only a minor role as shown above. (4) Data fusion is supported in multi-spectral, multi-sensor, and multi-temporal manner. (5) Even the description of spatial features in image chips as realized in CNNs is possible, though not evaluated in detail by now.

5. Conclusions

This article develops a consistent data frame for the fusion of arbitrary collocated raster data sets. The core of the novel fusion technique is an orthogonal transform on hypercomplex bases producing Kennaugh-like elements. The combination can cover multi-spectral, multi-source, and multi-temporal data sets. The description of spatial characteristics by multi-row and multi-column combination is envisaged. Multi-source image fusion results in vectors of Kennaugh-like elements. Multi-source and multi-temporal image fusion produces Kennaugh matrices with the image channel dimensionality along the rows and the time dimensionality along the columns. These linear intensities and intensity differences, respectively, can be further scaled to logarithmic—almost normally distributed—and even normalized elements characterized by a closed data range and a high resilience towards the reduction of the radiometric sampling to 4 bit or even less than that. The method is exemplary applied to Sentinel-1, Sentinel-2, and ALOS-PALSAR-2 single acquisitions of three different regions of interest in Germany and time series of Sentinel-2 over a wetland near the Alps. Due to its flexibility regarding the scaling, any data set with additive or multiplicative characteristics can be used as input layers. A qualitative assessment of the multi-temporal image fusion indicated that the Kennaugh-like elements highlight characteristics that were not visible at all in the standard preparation. Furthermore, the quantitative analysis of multi-source Kennaugh-like elements underlined the stability of the signatures describing typical land cover classes that are distinguishable even with a extremely reduced radiometric sampling of only two bins. The latter point is of importance with respect to modern classification techniques like histogram classifications and Random Forest approaches. The Random Forest method (e.g.) combines a large number of random decision trees in order to stabilize the decision. Each node is a binary sampling which is now simply possible using normalized Kennaugh-like elements in a binary radiometric sampling. This article can only presents the basic idea of a data fusion approach and delivers first hints on its suitability explained by the help of selected experiments. Its practical use has to be evaluated in future studies covering more different landscapes, different sensors, and different classification algorithms. With respect to sharpening approaches of image data fusion with different spatial resolution, the inclusion of weights representing the look factors might be one technical further development. One application-oriented focus will be the analysis of polarimetric SAR time series over Mid-European forests in order to detect and map calamities in a timely manner. The presented methodology further shall be implemented as extension of the Multi-SAR System in order to support ARD data provision to scientific users. The long-term goal will be to establish the normalized Kennaugh-like elements on hypercomplex bases as future standard for multi-dimensional ARD sets in so-called data cubes.

Author Contributions

A.S. developed the methodology, conducted the research, and wrote most part of the paper, A.W. preprocessed the data and wrote the introduction, R.K. and M.H. contributed preliminary studies and provided the training data set, A.R. initiated the idea of a consistent preprocessing, and S.H. supervised the study scientifically. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors like to thank the Japan Aerospace Exploration Agency (JAXA) and the European Space Agency (ESA) for the provision of the satellite data. The polarimetric SAR data were acquired under the research proposal 3258 entitled “Evaluating potential applications of polarimetric L-band SAR data over European temperate forests”. This study presents the first step towards the analysis of L-band full-polarimetric time series which is already applied on forest issues in ongoing studies. ALOS-PALSAR imagery ©JAXA 2017 & 2018. Sentinel-1 imagery ©ESA 2017 & 2018. Sentinel-2 imagery ©ESA 2017, 2018 & 2019.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARDAnalysis Ready Data
ASTERAdvanced Spaceborne Thermal Emission and Reflection Radiometer
BOABottom of Atmosphere
CEOSCommittee on Earth Observation Satellites
CNNConvolutional Neural Networks
DEMDigital Elevation Model
DFDGerman Remote Sensing Data Center
DLRGerman Aerospace Center
EOEarth Observation
ETMEnhanced Thematic Mapper
FPfully polarimetric (HH, HV, VH, VV)
GRDGround Range Detected
L2ALevel 2A images
MLEMaximum Likelihood Estimation
MSMLMulti-Scale Multi-Looking
PCAPrincipal Component Analysis
PDFProbability density function
RGBRed Green Blue
ROIRegion of Interest
SARSynthetic Aperture Radar
SDCSwiss Data Cube
SLCSingle look complex
SRTMShuttle Radar Topography Mission
TANHHyperbolic Tangent
USGSUnited State Geological Survey
UTMUniversal Transverse Mercator Coordinate System
WGS84World Geodetic System from 1984

References

  1. Sudmanns, M.; Tiede, D.; Lang, S.; Bergstedt, H.; Trost, G.; Augustin, H.; Baraldi, A.; Blaschke, T. Big Earth data: Disruptive changes in Earth observation data management and analysis? Int. J. Digit. Earth 2019. [Google Scholar] [CrossRef]
  2. ESA. Sentinel Data Access Annual Report 2019. COPE-SERCO-RP-19-0389 2019, 1, 1–116. [Google Scholar]
  3. Giuliani, G.; Chatenoux, B.; Honeck, E.; Richard, J. Towards Sentinel-2 Analysis Ready Data: A Swiss Data Cube Perspective. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8659–8662. [Google Scholar]
  4. CEOS Committee on Earth Observation Satellites. CEOS Analysis Ready Data Strategy. CEOS Plenary 2019, 1, 1–9. [Google Scholar]
  5. Wulder, M.A.; Whitea, J.C.; Loveland, T.R.; Woodcock, C.E.; Belward, A.S.; Cohen, W.B.; Fosnight, E.A.; Shawf, J.; Masekg, J.G.; Royh, D.P. The global Landsat archive: Status consolidation and direction. Remote Sens. Environ. 2016, 185, 271–283. [Google Scholar] [CrossRef] [Green Version]
  6. Giuliani, G.; Camara, G.; Killough, B.; Minchin, S. Earth Observation Open Science: Enhancing Reproducible Science Using Data Cubes. Data 2019, 4, 147. [Google Scholar] [CrossRef] [Green Version]
  7. Gorelick, N.; Hancher, M.; Dixon, M.; Ilyushchenko, S.; Thau, D.; Moore, R. Google Earth Engine: Planetary-scale geospatial analysis for everyone. Remote Sens. Environ. 2017, 202, 18–27. [Google Scholar] [CrossRef]
  8. Truckenbrodt, J.; Freemantle, T.; Williams, C.; Jones, T.; Small, D.; Dubois, C.; Thiel, C.; Rossi, C.; Syriou, A.; Giuliani, G. Towards Sentinel-1 SAR Analysis-Ready Data: A Best Practices Assessment on Preparing Backscatter Data for the Cube. Data 2019, 4, 93. [Google Scholar] [CrossRef] [Green Version]
  9. Ticehurst, C.; Zhou, Z.-S.; Lehmann, E.; Yuan, F.; Thankappan, M.; Rosenqvist, A.; Lewis, B.; Paget, M. Building a SAR-Enabled Data Cube Capability in Australia Using SAR Analysis Ready Data. Data 2019, 4, 100. [Google Scholar] [CrossRef] [Green Version]
  10. CEOS Committee on Earth Observation Satellites. CEOS Analysis Ready Data For Land Polarimetric data. CEOS Plenary 2019, 2.1, 1–9. [Google Scholar]
  11. Schmitt, A.; Wendleder, A.; Hinz, S. The Kennaugh element framework for multi-scale, multi-polarized, multi-temporal and multi-frequency SAR image preparation. ISPRS J. Photogramm. Remote Sens. 2015, 102, 122–139. [Google Scholar] [CrossRef] [Green Version]
  12. Schmitt, A. Multiscale and Multidirectional Multilooking for SAR Image Enhancement. IEEE Trans. Geosci. Remote Sens. 2013, 54, 5117–5134. [Google Scholar] [CrossRef]
  13. Schmitt, A.; Brisco, B. Wetland Monitoring Using the Curvelet-Based Change Detection Method on Polarimetric SAR Imagery. Water 2013, 5, 1036–1051. [Google Scholar] [CrossRef]
  14. Bauer, O.; Einzmann, K.; Schmitt, A.; Seitz, R. Erfassen von Windwurfflächen im Wald mit Satellitentechnik [Detection of wind throw areas using space-borne techniques]. AFZ – Der Wald 2015, 16, 14–17. [Google Scholar]
  15. Moser, L.; Schmitt, A.; Wendleder, A.; Roth, A. Monitoring of the Lac Bam Wetland Extent Using Dual-Polarized X-Band SAR Data. Remote Sens. 2016, 8, 1–31. [Google Scholar] [CrossRef] [Green Version]
  16. Moser, L.; Schmitt, A.; Wendleder, A. Automated wetland delineation from multi-frequency and multi-polarized SAR Images in high temporal and spatial resolution. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, VIII, 57–64. [Google Scholar] [CrossRef] [Green Version]
  17. Bertram, A.; Wendleder, A.; Schmitt, A.; Huber, M. Long-term Monitoring of water dynamics in the Sahel region using the Multi-SAR-System. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, VIII/8, 313–320. [Google Scholar] [CrossRef]
  18. Irwin, K.; Braun, A.; Fotopoulos, G.; Roth, A.; Wessel, B. Assessing Single-Polarization and Dual-Polarization TerraSAR-X Data for Surface Water Monitoring. Remote Sens. 2018, 10, 2076–3263. [Google Scholar] [CrossRef] [Green Version]
  19. Blocksdorf, A.; Schmitt, A.; Wendleder, A. Der Forggensee im Jahreswandel – auf dem Weg zur täglichen Wasserstandsmeldung aus dem Weltall [The Forggen lake throughout the year – on the way to a daily water-level report from space]. Photogramm. Fernerkund. Geoinf. 2018, 86, 29–33. [Google Scholar]
  20. Wang, W.; Gade, M.; Yang, X. Detection of Bivalve Beds on Exposed Intertidal Flats Using Polarimetric SAR Indicators. Remote Sens. 2017, 9. [Google Scholar] [CrossRef] [Green Version]
  21. Wendleder, A.; Stettner, S.; Roth, A.; Toose, P.; King, J. Relationships of TerraSAR-X Time Series with Snow Depth, Topography and Vegetation height over Arctic Tundra. In Proceedings of the IUGG, Montreal, QC, Canada, 8–18 July 2019. [Google Scholar]
  22. Dufour-Beauséjour, S.; Wendleder, A.; Gauthier, Y.; Bernier, M.; Poulin, J.; Gilbert, V.; Tuniq, J.; Rouleau, A.; Roth, A. Seasonal timeline for snow-covered sea ice processes in Nunavik’s Deception Bay from TerraSAR-X and time-lapse photography. Cryosphere Discuss. 2020. under review. [Google Scholar]
  23. Wendleder, A.; Hintermaier, S.; Schmitt, A.; Mayer, C. Reconstruction of snow stratigraphy using GPR and multi-frequency space-borne SAR data. In Proceedings of the 12th European Conference on Synthetic Aperture Radar, Aachen, Germany, 4–7 June 2018; Volume 9. [Google Scholar]
  24. Heilig, A.; Wendleder, A.; Schmitt, A.; Mayer, C. Discriminating Wet Snow and Firn for Alpine Glaciers Using Sentinel-1 Data: A Case Study at Rofental, Austria. Geosciences 2019, 9, 2076–3263. [Google Scholar] [CrossRef] [Green Version]
  25. Wurm, M.; Taubenböck, H.; Weigand, M.; Schmitt, A. Slum mapping in polarimetric SAR data using spatial features. Remote Sens. Environ. 2017, 194, 190–204. [Google Scholar] [CrossRef]
  26. Schmitt, A.; Sieg, T.; Wurm, M.; Taubenböck, H. Investigation on the separability of slums by multi-aspect TerraSAR-X dual- co-polarized high resolution spotlight images based on the multi-scale evaluation of local distributions. Int. J. Appl. Earth Observ. Geoinf. 2013, 64, 181–198. [Google Scholar] [CrossRef] [Green Version]
  27. Kleynmans, R. Implementierung und Validierung einer Parameterfreien Multiskaligen Maximum-Likelihood Klassifikation auf Basis lokaler Histogramme [Implementation and Validation of a Parameter-Free Multi-Scale Maximum Likelihood Classification Based on Local Histograms]. Bachelor’s Thesis, Munich University of Applied Sciences, Hochschule München, Germany, 2017. 10/OB 08-17 K64. [Google Scholar]
  28. Schmitt, A.; Wendleder, A. SAR-Sharpening in the Kennaugh Framework applied to the Fusion of Multi-Modal SAR and Optical Images. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, IV, 133–140. [Google Scholar] [CrossRef] [Green Version]
  29. Kleynmans, R. Validierung und Quantifizierung des Klassifizierungsmehrgewinns Durch SAR-Opt Datenfusion [Validation and Quantification of the Excess Profit From SAR-OPT-Data Fusion for Classification Purposes]. Master’s Thesis, Munich University of Applied Sciences, Hochschule München, Germany, 2019. 10/OM 08-19 K64. [Google Scholar]
  30. Hell, M. Überwachte, Parameterfreie, Multi-Skalige Klassifikation [Supervised, Parameter-Free, and Multi-Scale Classification]. Bachelor’s Thesis, Munich University of Applied Sciences, Hochschule München, Germany, 2019. 10/OB 08-19 H476. [Google Scholar]
  31. Pohl, C.; Genderen, J. Structuring contemporary remote sensing image fusion. Int. J. Image Data Fusion 2015, 6, 3–21. [Google Scholar] [CrossRef] [Green Version]
  32. Schmitt, M.; Zhu, X. Data Fusion and Remote Sensing: An ever-growing relationship. IEEE Geosci. Remote Sens. Mag. 2016, 4, 6–23. [Google Scholar] [CrossRef]
  33. Zhang, L.; Li, Y.; Lu, H.; Yamawaki, A.; Yang, S.; Serikawa, S. Maximum local energy method and sum modified Laplacian for remote image fusion based on beyond wavelet transform. Int. J. Appl. Math. Inf. Sci. 2013, 7, 149–156. [Google Scholar]
  34. Ha, W.; Gowda, P.H.; Howell, T.A. A review of potential image fusion methods for remote sensing-based irrigation management: Part II. Irrig. Sci. 2013, 31, 851–869. [Google Scholar] [CrossRef]
  35. Siddiqui, Y. The modified IHS method for fusing satellite imagery. In Proceedings of the ASPRS Annual Conference [CD], Baltimore, MD, USA, 7–11 March 2003. [Google Scholar]
  36. Chen, C.-M.; Hepner, G.F.; Forster, R.R. Fusion of hyperspectral and radar data using the IHS transformation to enhance urban surface features. ISPRS J. Photogramm. Remote Sens. 2003, 58, 19–30. [Google Scholar] [CrossRef]
  37. Xu, Q. High-fidelity component substitution pansharpening by the fitting of substitution data. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7380–7392. [Google Scholar]
  38. Xu, Q.; Zhang, Y.; Li, B.; Ding, L. Pansharpening Using Regression of Classified MS and Pan Images to Reduce Color Distortion. IEEE Geosci. Remote Sens. Lett. 2015, 12, 28–32. [Google Scholar]
  39. Chavez, P.S.; Sides, S.C.; Anderson, J.A. Comparison of three different methods to merge multiresolution and multispectral data: Landsat TM and SPOT panchromatic. Photogramm. Eng. Remote Sens. 1991, 57, 295–303. [Google Scholar]
  40. Chiang, J.-L. Knowledge-based principal component analysis for image fusion. Appl. Math. Inf. Sci. 2014, 8, 223–230. [Google Scholar] [CrossRef]
  41. Schowengerdt, R.A. Chapter 8—Image registration and fusion. In Remote Sensing, 3rd ed.; Schowengerdt, R.A., Ed.; Burlington: Burlington, NJ, USA, 2007; pp. 355–385. [Google Scholar]
  42. Pohl, C.; Genderen, J. Review article Multisensor image fusion in remote sensing: Concepts, methods and applications. Int. J. Remote Sens. 1998, 19, 823–854. [Google Scholar] [CrossRef] [Green Version]
  43. Mazaheri, S.; Sulaiman, P.S.; Wirza, R.; Dimon, M.Z.; Khalid, F.; Tayebi, R.M. Hybrid Pixel-Based Method for Cardiac Ultrasound Fusion Based on Integration of PCA and DWT. Math. Methods Appl. Med. Imag. 2015, 2015, 1–16. [Google Scholar] [CrossRef] [Green Version]
  44. Patel, H.; Upla, K. Survey on Image Fusion: Hand Designed to Deep Learning Algorithms. AJCT 2019, V, 1–9. [Google Scholar] [CrossRef]
  45. Xiao, M.; He, Z. Remote sensing image fusion based on Gaussian mixture model and multiresolution analysis. MIPPR 2013: Remote Sens. Image Process. Geogr. Inf. Syst. Other Appl. 2013, 8921, 173–180. [Google Scholar]
  46. Amolins, K.; Zhang, Y.; Dare, P. Wavelet based image fusion techniques—An introduction, review and comparison. ISPRS J. Photogramm. Remote Sens. 2007, 62, 249–263. [Google Scholar] [CrossRef]
  47. Zhang, Y.; Hong, G. An IHS and wavelet integrated approach to improve pan-sharpening visual quality of natural colour IKONOS and QuickBird images. Inf. Fus. 2015, 6, 225–234. [Google Scholar] [CrossRef]
  48. Choi, J.; Yeom, J.; Chang, A.; Byun, Y.; Kim, Y. Hybrid Pansharpening Algorithm for High Spatial Resolution Satellite Imagery to Improve Spatial Quality. IEEE Geosci. Remote Sens. Lett. 2013, 10, 490–494. [Google Scholar] [CrossRef]
  49. Metwalli, M.R.; Nasr, A.H.; Faragallah, O.S.; El-Rabaie, E.M.; Abbas, A.M.; Alshebeili, S.A.; Abd El-Samie, F.E. Efficient pan-sharpening of satellite images with the contourlet transform. Int. J. Remote Sens. 2014, 35, 1979–2002. [Google Scholar] [CrossRef]
  50. Khare, A.; Srivastava, R.; Singh, R. Edge Preserving Image Fusion Based on Contourlet Transform. Image Signal Process. 2012, 93–102. Available online: https://link.springer.com/chapter/10.1007/978-3-642-31254-0_11 (accessed on 15 January 2020).
  51. Choi, M.; Kim, R.Y.; Kim, M.-G. The curvelet transform for image fusion. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2004, 35. Available online: https://www.isprs.org/proceedings/XXXV/congress/yf/papers/931.pdf (accessed on 15 January 2020).
  52. Verma, M.; Kaushik, V.D.; Rao, C.V. Curvelet based image fusion. In Proceedings of the 2012 World Congress on Information and Communication Technologies, Trivandrum, India, 30 October–2 November 2012; pp. 959–964. [Google Scholar]
  53. Deshmukh, C.N.; Bajad, N.G. A comparative study of different image fusion techniques for tone-mapped images. Int. J. Sci. Eng. Res. 2016, 7, 474–478. [Google Scholar]
  54. Singh, S.; Rajput, R. A Comparative Study of Classification of Image Fusion Techniques. Int. J. Eng. Comput. Sci. 2014, 3, 7350–7353. [Google Scholar]
  55. Ivakhnenko, A.G.; Lapa, V.G. Cybernetic Predicting Devices; CCM Information Corporation: New York, NY, USA, 1965. [Google Scholar]
  56. Dechter, R. Learning while searching in constraint-satisfaction problems. In Proceedings of the Fifth National Conference on Artificial Intelligence (AAAI-86), Philadelphia, PA, USA, 11–15 August 1986; pp. 178–183. [Google Scholar]
  57. Liu, Y.; Chen, X.; Wang, Z.; Wang, Z.J.; Ward, R.K.; Wang, X. Deep learning for pixel-level image fusion: Recent advances and future prospects. Inf. Fus. 2018, 42, 158–173. [Google Scholar] [CrossRef]
  58. Piao, J.; Chen, Y.; Shin, H. A New Deep Learning Based Multi-Spectral Image Fusion Method. Entropy 2019, 21, 570. [Google Scholar] [CrossRef] [Green Version]
  59. He, W.; Yokoya, N. Multi-Temporal Sentinel-1 and -2 Data Fusion for Optical Image Simulation. ISPRS Int. J. Geo-Inf. 2018, 7, 389. [Google Scholar] [CrossRef] [Green Version]
  60. Liu, Y.; Chen, X.; Ward, R.; Wang, Z. Image Fusion With Convolutional Sparse Representation. IEEE Signal Process. Lett. 2016, 23, 1882–1886. [Google Scholar] [CrossRef]
  61. Zhai, J.; Dong, G.; Chen, F.; Xie, X.; Qi, C.; Lin, L. A Deep Learning Fusion Recognition Method Based On SAR Image Data. Procedia Comput. Sci. 2019, 147. [Google Scholar] [CrossRef]
  62. Chaudhuri, U.; Banerjee, B.; Bhattacharya, A.; Datcu, M. CMIR-NET: A Deep Learning Based Model For Cross-Modal Retrieval In Remote Sensing. Pattern Recognit. Lett. 2020, 131, 456–462. [Google Scholar] [CrossRef] [Green Version]
  63. Wessel, B.; Schmitt, A.; Wagner, L.; Roth, A. Generating Pseudo Quad-polarized SAR Images of Pursuit- monostatic TanDEM-X Data by Kennaugh Data Fusion. In Proceedings of the 12th European Conference on Synthetic Aperture Radar, Aachen, Germany, 4–7 June 2018. [Google Scholar]
  64. Schmitt, A. Änderungserkennung in Multitemporalen und Multipolarisierten Radaraufnahmen [Change Detection on Multi-Temporal and Multi-Polarized Radar Acquisitions]. Ph.D. Thesis, Karlsruhe Institute of Technology, Karlsruher Institut für Technologie, Karlsruhe, Germany, 2012. [Google Scholar]
  65. Colditz, R.; Wehrmann, T.; Bachmann, M.; Steinnocher, K.; Schmidt, M.; Strunz, G.; Dech, S. Influence of Image Fusion Approaches on Classification Accuracy—A Case Study. Int. J. Remote Sens. 2006, 27, 143–149. [Google Scholar] [CrossRef]
  66. Elkholy, M.; Hosny, M.M.; El-Habrouk, H.M. Studying the effect of lossy compression and image fusion on image classification. Alexandria Eng. J. 2019, 58, 143–149. [Google Scholar] [CrossRef]
  67. Pal, M. Factors Influencing the Accuracy of Remote Sensing Classifications: A Comparative Study. Ph.D. Thesis, University of Nottingham, Nottingham, UK, 2002. [Google Scholar]
  68. Moghaddam, H.H.; Torahi, A.A.; Abadi, P.Z.F. Using Discrete Wavelet Transform to increase the Accuracy of Hyper Spectral and High Resolution Images Fusion. J. Radar Opt. Remote Sens. 2019, 1, 22–30. [Google Scholar]
  69. Yulianti, Y.; Yulianti, E. Multi-Temporal Sentinel-2 Images for Classification Accuracy. J. Comput. Sci. 2019, 15, 258–268. [Google Scholar]
  70. Sukawattanavijit, C.; Srestasathiern, P. Object-based land cover classification based on fusion of multifrequency SAR data and THAICHOTE optical imagery. Remote Sens. Agric. Ecosyst. Hydrol. 2017, XIX, 13. [Google Scholar]
  71. Amarsaikhan, D.; Blotevogel, H.H.; van Genderen, J.L.; Ganzorig, M.; Gantuya, R.; Nergui, B. Fusing high-resolution SAR and optical imagery for improved urban land cover study and classification. Int. J. Image Data Fus. 2010, 1, 83–97. [Google Scholar] [CrossRef]
  72. Makarau, A.; Palubinskas, G.; Reinartz, P. Classification Accuracy Increase using Multisensor Data Fusion. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2011, XXXVIII, 181–186. [Google Scholar] [CrossRef] [Green Version]
  73. Jing, X.; Bao, Y. Analysis and Evaluation of IKONOS Image Fusion Algorithm Based on Land Cover Classification. Asian Agric. Res. 2015, 7, 52–56. [Google Scholar]
  74. Simone, G.; Farina, A.; Morabito, F.C.; Serpico, S.B.; Bruzzone, L. Image fusion techniques for remote sensing applications. Inf. Fus. 2002, 3, 3–15. [Google Scholar] [CrossRef] [Green Version]
  75. Luo, H. Classification Precision Analysis on Different Fusion Algorithm for ETM+ Remote Sensing Image. In Proceedings of the 2nd International Conference on Electronics, Network and Computer Engineering (ICENCE 2016), Yinchuan, China, 13–14 August 2016. [Google Scholar]
  76. Wenbo, W.; Jing, Y.; Tingjun, K. Study of Remote Sensing Image Fusion and Its Application in Image Classification. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, XXXVII, 1141–1146. [Google Scholar]
  77. Al-Wassai, F.A.; Kalyankar, N.V. Influences Combination Of Multi-Sensor Images On Classification Accuracy. Int. J. Adv. Res. Comput. Sci. 2013, 4, 10–20. [Google Scholar]
Figure 1. The orthogonal transform from reflectance (blue) to Kennaugh elements (red) in the two dimensional space of one measurement (green) according to Equation (2).
Figure 1. The orthogonal transform from reflectance (blue) to Kennaugh elements (red) in the two dimensional space of one measurement (green) according to Equation (2).
Remotesensing 12 00943 g001
Figure 2. Sentinel-2 time series of True Colour Images (left column) and Colour Infrared Images (right column) from July to October 2019 over the “Murnauer Moos” in southern Germany. For coordinate reference, scale, and north arrow see Figure 3a.
Figure 2. Sentinel-2 time series of True Colour Images (left column) and Colour Infrared Images (right column) from July to October 2019 over the “Murnauer Moos” in southern Germany. For coordinate reference, scale, and north arrow see Figure 3a.
Remotesensing 12 00943 g002
Figure 3. Multi-temporal combination of the Kennaugh elements to the four-by-four matrix K 4 × 4 calculated of the Sentinel-2 10 m -layers imaged in July, August, September, and October 2019 over the “Murnauer Moos”, see Figure 2. The depicted channels are indicated on the right side of each image. The coordinates are given in latitude and longitude on the WGS84 ellipsoid.
Figure 3. Multi-temporal combination of the Kennaugh elements to the four-by-four matrix K 4 × 4 calculated of the Sentinel-2 10 m -layers imaged in July, August, September, and October 2019 over the “Murnauer Moos”, see Figure 2. The depicted channels are indicated on the right side of each image. The coordinates are given in latitude and longitude on the WGS84 ellipsoid.
Remotesensing 12 00943 g003
Figure 4. Subset of ROI 1: the “Starnberg Lake” southern of Munich (coordinate reference WGS84).
Figure 4. Subset of ROI 1: the “Starnberg Lake” southern of Munich (coordinate reference WGS84).
Remotesensing 12 00943 g004
Figure 5. Subset of ROI 2: the “Ruhr Metropolitan Area” (coordinate reference WGS84).
Figure 5. Subset of ROI 2: the “Ruhr Metropolitan Area” (coordinate reference WGS84).
Remotesensing 12 00943 g005
Figure 6. Subset of ROI 3: the “Potsdam Lake Land” south-western of Berlin (coordinate reference WGS84).
Figure 6. Subset of ROI 3: the “Potsdam Lake Land” south-western of Berlin (coordinate reference WGS84).
Remotesensing 12 00943 g006
Figure 7. The distribution of the Kennaugh-like elements on the Octonion basis fusing Sentinel-1 and Sentinel-2 (ac) and on the Sedenion basis fusing ALOS-PALSAR-2 and Sentinel-2 (df) for class 2 “Forest” in linear, logarithmic, and normalized scaling per regions of interest.
Figure 7. The distribution of the Kennaugh-like elements on the Octonion basis fusing Sentinel-1 and Sentinel-2 (ac) and on the Sedenion basis fusing ALOS-PALSAR-2 and Sentinel-2 (df) for class 2 “Forest” in linear, logarithmic, and normalized scaling per regions of interest.
Remotesensing 12 00943 g007
Figure 8. The distribution of the Kennaugh-like elements on the Octonion basis fusing Sentinel-1 and Sentinel-2 (ac) and on the Sedenion basis fusing ALOS-PALSAR-2 and Sentinel-2 (df) for class 5 “Urban” in linear, logarithmic, and normalized scaling per ROI.
Figure 8. The distribution of the Kennaugh-like elements on the Octonion basis fusing Sentinel-1 and Sentinel-2 (ac) and on the Sedenion basis fusing ALOS-PALSAR-2 and Sentinel-2 (df) for class 5 “Urban” in linear, logarithmic, and normalized scaling per ROI.
Remotesensing 12 00943 g008
Figure 9. Total Accuracy and κ -coefficient of the contingency matrices achieved using a Maximum- Likelihood class discriminator in relation to the test site (“Starnberg Lake” solid, “Ruhr Metropolitan Area” dashed, and “Potsdam Lake Land” dot-dashed lines), the channel combination and scaling (linear reflectance, logarithmic Kennaugh-like elements, and normalized Kennaugh-like elements, see legend), and the radiometric sampling given by the number of bins.
Figure 9. Total Accuracy and κ -coefficient of the contingency matrices achieved using a Maximum- Likelihood class discriminator in relation to the test site (“Starnberg Lake” solid, “Ruhr Metropolitan Area” dashed, and “Potsdam Lake Land” dot-dashed lines), the channel combination and scaling (linear reflectance, logarithmic Kennaugh-like elements, and normalized Kennaugh-like elements, see legend), and the radiometric sampling given by the number of bins.
Remotesensing 12 00943 g009
Figure 10. Mean intra- (upper part) and inter-class (lower part) signature similarity on the Octonion basis using (a) linear, (b) logarithmic, and (c) normalized elements evaluated with individual (left side) and mean signatures (right side) for the “Potsdam Lake Land” test site.
Figure 10. Mean intra- (upper part) and inter-class (lower part) signature similarity on the Octonion basis using (a) linear, (b) logarithmic, and (c) normalized elements evaluated with individual (left side) and mean signatures (right side) for the “Potsdam Lake Land” test site.
Remotesensing 12 00943 g010
Figure 11. Mean intra- (upper part) and inter-class (lower part) signature similarity on the Sedenion basis using (a) linear, (b) logarithmic, and (c) normalized elements evaluated with individual (left side) and mean signatures (right side) for the “Starnberg Lake” test site.
Figure 11. Mean intra- (upper part) and inter-class (lower part) signature similarity on the Sedenion basis using (a) linear, (b) logarithmic, and (c) normalized elements evaluated with individual (left side) and mean signatures (right side) for the “Starnberg Lake” test site.
Remotesensing 12 00943 g011
Figure 12. Similarity gain of intra- vs. inter-class comparison for individual (left part) and mean signatures (right part) in three different scalings: normalized (solid), logarithmic (dashed), and linear (dotted) on the Octonion (left) and the Sedenion basis (right) evaluated for each class under study separately.
Figure 12. Similarity gain of intra- vs. inter-class comparison for individual (left part) and mean signatures (right part) in three different scalings: normalized (solid), logarithmic (dashed), and linear (dotted) on the Octonion (left) and the Sedenion basis (right) evaluated for each class under study separately.
Remotesensing 12 00943 g012
Table 1. Empirical central moments one to five normalized to linear deviations of the distributions adopted by the multi-spectral and multi-temporal reflectance images and Kennaugh-like elements fused on a two-dimensional Quaternion basis for the “Murnauer Moos” test site.
Table 1. Empirical central moments one to five normalized to linear deviations of the distributions adopted by the multi-spectral and multi-temporal reflectance images and Kennaugh-like elements fused on a two-dimensional Quaternion basis for the “Murnauer Moos” test site.
Reflectance Values
ChannelT1T2T3T4BlueGreenRedInfrared
Moment
10.1090.1020.0960.0710.0120.0200.0180.106
20.1370.1330.1270.1000.0190.0260.0270.133
30.1590.1610.1580.1290.0370.0380.0450.053
40.1960.2020.1990.1650.0810.0790.0830.172
50.2210.2340.2310.1950.1430.1380.1400.142
Kennaugh-Like Elements
Channel K 0 T K 1 T K 2 T K 3 T K 0 S K 1 S K 2 S K 3 S
Moment
10.2760.0270.0390.0180.1360.0970.0880.083
20.3060.0350.0500.0280.1760.1280.1140.109
30.1690.0130.0350.0240.2060.1480.1310.124
40.3540.0710.0860.0740.2600.1880.1650.158
50.3170.0720.1060.1100.3020.2110.1840.177
Linear Scale
10.5970.0780.1190.0470.1800.2490.2310.214
20.6200.0980.1540.0670.2240.3050.2890.263
30.0370.0340.1100.0450.2250.3120.2660.269
40.6490.1370.2200.1080.3080.3860.3770.336
50.3620.1040.2150.1000.3340.4080.3290.355
Normalized Scale
16.3610.6871.0570.4121.7902.8412.4452.210
26.7270.8671.3950.5892.4003.5103.0682.748
31.8390.3321.1400.4102.9593.7852.9862.936
47.3091.2432.1270.9844.0864.6124.0883.618
55.3901.0592.3340.9574.9374.9663.7803.894
logarithmic Scale
variabilityspectraltemporal
Table 2. The remote sensing data subject to the presented study.
Table 2. The remote sensing data subject to the presented study.
ROI & Coordinate ReferenceSensorProductDate
ROI 1ALOS-PALSAR-2Stripmap FP2017-04-05
Starnberg LakeSentinel-1IW (VV&VH)2017-09-01
UTM Zone 32NSentinel-2MSI-L2A2017-08-25
ROI 2ALOS-PALSAR-2Stripmap FP2018-08-12
Ruhr Metropolitan AreaSentinel-1IW (VV&VH)2018-08-12
UTM Zone 31NSentinel-2MSI-L2A2018-08-06
ROI 3ALOS-PALSAR-2Stripmap FP2017-04-23
Potsdam Lake LandSentinel-1IW (VV&VH)2018-06-04
UTM Zone 33NSentinel-2MSI-L2A2018-06-06
Table 3. Enhancement of the mean class distinction using the Octonion basis in different scaling variants (rows) evaluated and varying number of bins (columns) for the three test sites evaluated over all samples given in decibel [dB].
Table 3. Enhancement of the mean class distinction using the Octonion basis in different scaling variants (rows) evaluated and varying number of bins (columns) for the three test sites evaluated over all samples given in decibel [dB].
ROI 1“Starnberg Lake”
norm15.616.615.917.115.915.916.212.58.911.77.44.63.9
log15.014.715.914.013.68.712.55.75.34.33.82.33.9
lin26.726.223.618.321.69.712.29.91.83.61.31.60.0
ROI 2“Ruhr Metropolitan Area”
norm21.222.322.321.722.420.317.916.211.912.97.45.64.7
log20.020.119.017.316.312.612.97.57.06.75.53.94.7
lin31.032.031.532.027.923.026.423.53.48.02.52.50.1
ROI 3“Potsdam Lake Land”
norm21.522.523.926.324.924.121.917.413.115.39.16.35.4
log23.922.522.222.017.614.315.29.910.15.75.45.75.4
lin28.430.427.225.224.613.720.613.74.24.93.23.50.2
bins100715036251813965432
Table 4. Enhancement of the mean class distinction using the Sedenion basis in different scaling variants (rows) evaluated and varying number of bins (columns) for the three test sites evaluated over all samples given in decibel [dB].
Table 4. Enhancement of the mean class distinction using the Sedenion basis in different scaling variants (rows) evaluated and varying number of bins (columns) for the three test sites evaluated over all samples given in decibel [dB].
ROI 1“Starnberg Lake”
norm34.534.036.137.035.433.530.725.421.713.219.97.915.4
log36.433.731.527.222.822.211.58.516.46.516.11.315.4
lin32.133.130.927.228.120.421.418.26.18.02.15.70.7
ROI 2“Ruhr Metropolitan Area”
norm33.434.635.436.637.135.734.630.327.815.224.811.419.0
log35.834.233.831.027.128.613.813.020.37.620.12.619.0
lin33.835.334.834.232.327.830.126.17.316.11.67.60.3
ROI 3“Potsdam Lake Land”
norm34.635.333.532.531.029.327.123.320.414.219.09.013.8
log31.129.628.224.721.921.413.410.316.85.114.73.413.8
lin32.633.431.730.030.422.326.020.06.212.21.55.41.0
bins100715036251813965432

Share and Cite

MDPI and ACS Style

Schmitt, A.; Wendleder, A.; Kleynmans, R.; Hell, M.; Roth, A.; Hinz, S. Multi-Source and Multi-Temporal Image Fusion on Hypercomplex Bases. Remote Sens. 2020, 12, 943. https://doi.org/10.3390/rs12060943

AMA Style

Schmitt A, Wendleder A, Kleynmans R, Hell M, Roth A, Hinz S. Multi-Source and Multi-Temporal Image Fusion on Hypercomplex Bases. Remote Sensing. 2020; 12(6):943. https://doi.org/10.3390/rs12060943

Chicago/Turabian Style

Schmitt, Andreas, Anna Wendleder, Rüdiger Kleynmans, Maximilian Hell, Achim Roth, and Stefan Hinz. 2020. "Multi-Source and Multi-Temporal Image Fusion on Hypercomplex Bases" Remote Sensing 12, no. 6: 943. https://doi.org/10.3390/rs12060943

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop