Next Article in Journal
Dynamics of Water Quality in the Mirim–Patos–Mangueira Coastal Lagoon System with Sentinel-3 OLCI Data
Previous Article in Journal
Ensemble Machine Learning Approaches for Bathymetry Estimation in Multi-Spectral Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network for Moving Object Satellite Image Conversion

by
Andrew J. Lew
*,
Timothy Perkins
,
Ethan Brewer
,
Paul Corlies
and
Robert Sundberg
Spectral Sciences, Inc., 30 Fourth Avenue, Suite 2, Burlington, MA 01803, USA
*
Author to whom correspondence should be addressed.
Geomatics 2025, 5(3), 35; https://doi.org/10.3390/geomatics5030035
Submission received: 4 June 2025 / Revised: 16 July 2025 / Accepted: 21 July 2025 / Published: 23 July 2025

Abstract

Integrating diverse image datasets acquired from different satellites is challenging. Converting images from one sensor to another, like from WorldView-3 (WV) to SuperDove (SD), involves both changing image channel wavelengths and per-band intensity scales because different sensors can acquire imagery of the same scene at different wavelengths and intensities. A parametrized convolutional network approach has shown promise converting across sensor domains, but it introduces distortion artefacts when objects are in motion. The cause of spectral distortion is due to temporal delays between sequential multispectral band acquisitions. This can result in spuriously blurred images of moving objects in the converted imagery, and consequently misaligned moving object locations across image bands. To resolve this, we propose an enhanced model, the Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network (PIGESBCCN), which better accounts for known spatial, spectral, and temporal correlations between bands via band reordering and branched model architecture.

1. Introduction

Satellite imagery can provide crucial benefits in tracking global economic and ecological changes. However, integrating diverse datasets from the myriads of observation satellites available poses significant challenges. Issues arise from differences in sensing technologies, the absence of standardized calibration methods prior to satellite launches, and changes in instruments over time. Although improved calibration techniques could enhance alignment of future data acquisition [1], a large amount of historical data remains mutually incompatible. For instance, the U.S. Geological Survey’s Global Fiducials Library combines Landsat images from 1972 with Sentinel-2 images from 2015 [2], highlighting the need for methods to effectively bridge different image types.
Converting satellite imagery from the domain of one sensor, such as WorldView-3 (WV), to another, such as SuperDove (SD), is non-trivial. As a first step, spectral resampling is necessary to translate the wavelength response of the WV sensor filters into the desired wavelengths of the SD image bands. This is because WV and SD sensors acquire imagery at different wavelengths. However, the resampled image remains quantitatively and qualitatively divergent from an image truly acquired in SD space, so additional radiometric conversions are required. Building off previous windowed image regression tasks [3], we have demonstrated initial success with a parametrized convolutional neural network (PCNN) approach that leverages a Bayesian hyperparameter optimization loop to identify a performant model in this complex and often opaque conversion problem [4].
Though demonstrated to be qualitatively and quantitatively performant on converting stationary images, we identify spurious conversion artefacts with the PCNN approach when objects in the field of view are in motion. In these instances, the time delay between band acquisitions means that the moving object should appear localized in slightly different spatial locations across the bands, and a rainbow-like visual effect should be observed when viewing multiple bands at once (such as in a traditional RGB view) [5]. The issue arises because the previous PCNN model treats all image bands together in the prediction of a converted spectra, meaning that the effect of the moving object in each band spuriously bleeds over to every other band. This results in an expanded blurred blob for moving objects in the converted image, instead of spatiotemporally localized object signatures separated by band. Preserving these image features is of practical importance with object tracking applications to accurately represent where and at what time a moving object of interest appears.
While experimental data for the evaluation of specific moving objects can be prohibitive to acquire, radiometrically accurate simulations of 3D objects and environments have been studied for decades. In particular, the Quick Image Display (QUID) model continually developed by Spectral Sciences, Inc. since the early 1990s [6] incorporates six basic signature components including thermal surface emissions, scattered solar radiation, scattered earthshine, scattered skyshine, plume emission from hot molecular or particular species, and ambient atmospheric transmission/emission. QUID computes radiance images at animation rates by factoring thermal emission and reflectance calculations into angular and wavelength dependent terms. This factorization allows for the precomputation of wavelength-dependent terms which do not change with target-observer angles, enabling fast object rendering and scene simulation capabilities. As a result, QUID has found broad application as a general tool for moving object simulation, from ground-based automobiles [7] to aerial missiles [8]. Thus, simulation can be a key tool for evaluating performance of image conversion methods, particularly in specialized scenarios where experimental data is inaccessible.
To address this issue of correctly treating moving objects, here we introduce an expanded Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network (PIGESBCCN) to handle known spatial, spectral, and temporal correlations between bands and evaluate against QUID-simulated moving object scenes. Methods for data preparation, model optimization and evaluation, and for simulating imagery of moving target signatures labeled with ground truth data of object size and motion are discussed in Section 2, followed by results in Section 3 and conclusions in Section 4.

2. Materials and Methods

2.1. Dataset Curation

Satellite imagery of the same location in Turkey was acquired via both WV and SD sensors. WV images were acquired at spectral bands of 425, 480, 545, 605, 660, 725, 832, 950 nm. SD images were acquired at 443, 490, 531, 565, 610, 665, 705, 865 nm. WV images were band resampled to SD space using the open-source Spectral Python library, commonly used for processing hyperspectral image data [9]. After band resampling, labeled pairs of data were formed by taking matched pixel locations from the WV and SD images, and expanding a window around the WV location to obtain a region for convolutional input. Symmetric padding was used for windows at the periphery of the image. Of these WV/SD pairs, 70% were used as training data, 20% as validation data, and 10% reserved as testing data for a total of 2,303,417 training data pairs, 658,119 validation data pairs, and 329,060 test data pairs.

2.2. Hyperparameter Tuning for Model Optimization

The neural network was implemented in Python 3.8.19 using the PyTorch 2.2.2 library [10] and parametrized for application of automated hyperparameter tuning workflows. Here, we utilize the Adaptive Experimentation Platform (Ax) [11] for Bayesian optimization of hyperparameters. Specifically, for the surrogate function we use the commonly implemented Gaussian process, and for the acquisition function we use expected improvement as shown in Equations (1) and (2):
u ( x ) = m a x ( 0 , f f x )
a E I x = E   [ u ( x )   |   x , D ]
where f’ is the minimal value of f observed so far, u(x) is the utility of choosing f(x), D are the set of previous observations, and E is the expected value function. The search space of hyperparameters treated here is defined as below:
  • Gaussian blur size: a range of integers from 3 to 10, inclusive;
  • Window size: a range of integers from 3 to 20, inclusive;
  • Kernel size: a choice of integers between 3 and 5;
  • Pooling mode: a choice of modes between average and max;
  • Learning rate: a range of floats from 0.0001 to 0.01, inclusive;
  • Batch size: a choice of integers between 256, 512, 1024, and 2048.
As a result of this process, we identify an optimized PIGESBCCN model with hyperparameters of blur size 6, window size 14, kernel size 5, average pooling mode, learning rate 0.00076, and batch size 256, for a total model size of 15,752 parameters.
While the training loss is minimized via direct parameter training and validation loss is minimized via hyperparameter tuning, the test set comprises data that the model has never seen. Importantly, our optimized model performs well on the test set, with (1) test error distributions in family with both training and validation, and (2) parity plots with r-squared values > 0.96 across every band, as shown in the following Figure 1.

2.3. Spectral Angle Deviation Calculation

Spectral angles quantify the similarity between spectra, with smaller angles reflecting greater similarity. This is determined by measuring the angle between vectorized spectra in N-dimensional space [12], where the images in this study consist of N = 8 bands. To assess the performance of the ML conversion process, we compute the WV-to-SD deviation for both pre- and post-ML converted WV. The spectral angle α is then calculated using the formula provided in Equation (3):
α = cos 1 ( i = 1 8 w i s i i = 1 8 w i 2 i = 1 8 s i 2 )
for each co-registered pixel location in a pair of WV and SD images with WV spectra w , and SD spectra s . We can then plot heatmaps of these spectral angles to highlight locations of high deviance.

2.4. QUID Target Simulations

The QUID scene simulation method employs 3D wireframe models attributed with spectral bidirectional reflectance distribution functions (BRDFs) to characterize the reflective/emissive properties of the object surfaces and incorporate atmospheric simulations from MODTRAN 6 [13]. The environmental illumination from the sun and other sources includes the solar and diffuse fluxes necessary to illuminate the ground, a cloud field, and a target viewed through any line-of-sight through the atmosphere. The data also defines the path radiance and transmittance to model the opacity and brightness of the atmosphere between a sensor and a target. The current capabilities can compute accurate intensity images of targets in wavelengths from 0.4 µm to 15 µm. Our simulation approach separates in-band spectral quantities from angular factors to provide scene simulation rates of hundreds of hertz on a desktop workstation.
Using this software, we simulated a black sport utility vehicle (SUV) at sea level, moving toward a heading of 123.3° south of east (for a southwest direction) at 13 m/s (29 mph). The sun was placed at a solar azimuth angle of 164.0° and a solar elevation angle of 0°. The black SUV was simulated across 8 spectral bands, Coastal Blue (443 nm), Blue (490 nm), Green 1 (531 nm), Green 2 (565 nm), Yellow (610 nm), Red (665 nm), Red Edge (705 nm), and NIR (865 nm). Importantly, however, these bands were temporally acquired in the order of Blue, Red, Green 1, Green 2, Yellow, Red Edge, NIR, Coastal Blue, with timing matching SD imagery. As a result, the output QUID-simulated scene exhibits band localized SUV locations consistent with a true SD acquisition.

3. Results

WV sensors acquire their 8 spectral bands not in wavelength order. Rather, the spectral indices are temporally acquired in the order of 1, 8, 7, 5, 4, 3, 6, then 2. Thus, if one were to conduct convolutions over spectrally adjacent bands (i.e., bands 1 and 2), this would lead to a large target blurring artefact, as the temporal gap between band 2 (last temporally acquired) and band 1 (first temporally acquired) would result in greatly different target locations. Instead, we leverage the known temporal order of acquisition to first rearrange the bands, such that target motion is smooth between bands and convolutions act on temporally adjacent bands. Then, we architect the model with separate convolutional branches for each output band as an additional precaution against large cross-band target position blurring.
The PIGESBCCN model architecture, as shown in Figure 2, first separates out each band of the input image with band-specific Gaussian blurs derived from analysis of identical contrast targets in both the WV (input) and SD (output) domains. Specifically, we incorporate specific Gaussian blurs scaled by factors of 4.2576, 4.2679, 4.2670, 4.2498, 4.2841, 4.4390, 4.2026, and 4.3625 for the 8 image channels, respectively. We incorporate that physical knowledge to inform the model, instead of just hoping the ML model parameters learn the correct relative blurring. We then rearrange the order of the input scene bands by temporal acquisition. This rearrangement informs the model of the correct physical time progression, instead of just hoping the ML model parameters learn the correct band reordering. Once reordered, we chunk the image into 8 temporally adjacent 3-band sections. The first and last chunks at the temporal periphery repeat their earliest and latest bands, respectively, to yield 3-band chunks the same shape as temporally internal chunks. Then, the physical knowledge that correct output has band-localized object positions further informs our design of the model architecture, as separate convolutional branches per output image channel, instead of just hoping the ML model parameters learn correct band separation. Thus, after sectioning, these temporal chunks are then rearranged back into the original spectral ordering. Then, each of the 8 3-band chunks are fed through separate parameterized convolutional network paths to obtain a prediction for each band, and ultimately, a converted spectra for each pixel of the input image. In these ways, our knowledge of the physical system informs our methodology, solidifies known relationships between the data, and ameliorates spurious blur, band reordering, and band mixing errors that may otherwise manifest with a non-physics-informed basic convolutional approach.
Two images capturing the same spatial location—a scene in Turkey—were obtained using the WV and SD sensors, respectively. The spectra in the SD image are used to label their respective WV regions, constituting a paired dataset to facilitate learning the relationship between the WV and SD spectra. Due to the differing intensity scales of the WV and SD images (0–255 for WV and 0–24,545 for SD), we first normalize the pixel values of each image to a range of 0 to 1 by dividing by the maximum value of the image. After normalization, we split the pixels from these images into training, validation, and test datasets using a 70-20-10 split. Further details are provided in Section 2.
The training set is employed for direct parameter training. Then, as in previous work [14], we employ an automated hyperparameter tuning process using Bayesian optimization. The validation set guides the hyperparameter tuning loop, ultimately resulting in a final model that minimizes both training and validation loss. The paired WV and SD scenes used for training and the loss curves for the optimized model are shown in Figure 3. Details on hyperparameter tuning and optimized model performance are provided in Section 2.
To quantify the effectiveness of ML spectral conversion, we present the spectral angle deviations from actual SD spectra in Figure 4. Details of the calculation are in Section 2. Panel a compares spectral angle deviations from SD to the WV image after simple band resampling, while panel b compares from SD to the WV image converted with the trained ML model. In this visualization, brighter pixels represent greater deviations. The predominantly dark image in panel b indicates that the ML converted image is significantly closer to the actual SD data compared to simple band resampling. A horizontal trace across the field of view indicates that the average spectral angle deviation along that trace is 5.12° for the WV image after band resampling, which decreases significantly to an average of just 1.42° following conversion with the trained ML model.
Now, we utilize QUID to simulate a moving object in the input WV scene. Here, we specifically simulate a black SUV, as a representative generic target that can be found on an urban highway. Note that due to the time delay between band acquisitions, the moving object appears as a “rainbow” in a visible color image as it traverses in space. As the SUV in question is black, these colors then appear inverted when on a bright grey road background—masking the respective color band from the background. The QUID-simulated object scene is then passed through our PIGESBCCN model and converted to SD space to evaluate how well the ML conversion preserves band-separated object locations, as shown in the following Figure 5. Details of the target simulation are provided in Section 2.
To evaluate the ML model performance on converting a moving object to SD space, we use a simple Gaussian blur of the WV QUID images, both with and without a moving target, as a comparative baseline. This blur serves to simulate the difference in spatial resolution between the WV and SD sensors while maintaining correctly band-separated object locations. Specifically, using band-dependent full width half maximums of 4.258, 4.268, 4.267, 4.250, 4.284, 4.439, 4.203, and 4.363, for spectral bands of 0.443 nm, 0.490 nm, 0.531 nm, 0.565 nm, 0.610 nm, 0.665 nm, 0.705 nm, and 0.865 nm, respectively. We then pass the same WV QUID images through the PIGESBCCN ML conversion model. Finally, we subtract out the blurred background from the blurred background plus target image to isolate the effect on the target, as shown in Figure 6.
Zooming in on the target in a particular band allows us to qualitatively compare the difference between the ML conversion and baseline Gaussian blurring. Here, an example focusing on band 4 is shown in the third column of Figure 6. The ML converted target qualitatively appears with a similar location and size as the Gaussian blurred target. Note that the color of the ML converted target is not the same as the baseline blurred target. This is to be expected, as the baseline blurred target remains in WV color space while the ML converted target is in simulated SD color space.
We then conduct background subtraction for each of the 8 bands and use a value threshold of 5% to mask the target from the background. Then, comparing the baseline Gaussian blurred target and the ML converted target allows us to visualize the deviation between the ML method and baseline procedure for each band, as shown in Figure 7. First, this band-by-band visualization confirms proper target localization with slightly different spatial locations per band. Second, we see that these target locations for each band qualitatively align between the ML method and baseline blurring.
To quantitatively investigate the possibility of systematic errors associated with ML conversion, we compare target locations between the ML converted and baseline Gaussian approaches in Figure 8. Here, we see that the ML conversion yields target locations with good agreement to the baseline approach. More traditional methods of image conversion, without the band separation scheme outlined in this work, yield converted images where the presence of moving target in each band blurs over all output bands. In those cases, target reconstruction results in a single average effective position instead of varying per band. While this single effective position may be close to the positions of temporally internal bands (such as bands 4 or 5), other bands (such as 1 or 2) will contribute to large errors. Specifically, this single effective position yields Root Mean Squared Error (RMSE) of 1.8 and 2.7 pixels in the X and Y coordinates, respectively. In contrast, the PIGESBCCN converted target with band localized positions yields RMSE of only 0.49 and 0.41 pixels for the X and Y coordinates, respectively. This subpixel location accuracy indicates that the ML conversion successfully preserves target location without introducing systematic spatial translation errors. Furthermore, the averaged target size for the PIGESBCCN-treated imagery is close to the Gaussian blur baseline—only 1.6 pixels larger for the ML converted imagery (77 pixels) than the baseline (75.4 pixels).

4. Discussion and Conclusions

ML models can successfully learn complex non-linear relationships to convert one data domain to another. Application to satellite imagery can allow greater usage of data from disparate sources for combined studies.
Correct treatment of moving targets within satellite imagery was identified as a key mode for improvement over previous PCNN models, requiring a more sophisticated approach. Leveraging known physical relationships of the problem domain to put specific structure and guiderails on model architecture allows for corrected performance. Specifically, band-dependent blurring, temporal reordering and chunking, and separated band-specific prediction branches allow for improved treatment of spatial, temporal, and spectral characteristics.
As a result, the ML conversion approach detailed in this work operates quantitatively not only on stationary backgrounds but also for moving targets. The PIGESBCCN model enables accurate spectral conversion between WV and SD sensors, while maintaining moving target sizes and target locations across bands. Preservation of band localized moving targets is critical for moving target detection algorithms, a major application of satellite imagery data.
Future work may involve expanding into other image sensors and spectral bands, increasing computational efficiency for real-time conversion operations, and integrating our model architecture within live, periodic model training workflows for continually improved performance as more satellite imagery is acquired over time. Treatment of moving object blurring might not be important for general uses in, for example, agricultural monitoring or segmentation applications over large time windows. However, correct treatment of moving objects is important for select applications in, for example, small target detection, and the treatment of such experimental data may be addressed in future work.

Author Contributions

Conceptualization, A.J.L. and R.S.; methodology, A.J.L., E.B., P.C. and T.P.; software, A.J.L. and T.P.; validation, A.J.L., T.P. and P.C.; formal analysis, A.J.L.; investigation, A.J.L.; resources, R.S.; data curation, A.J.L. and R.S.; writing—original draft preparation, A.J.L.; writing—review and editing, A.J.L., T.P., E.B., P.C. and R.S.; visualization, A.J.L.; supervision, R.S.; project administration, R.S. All authors have read and agreed to the published version of the manuscript.

Funding

Support for this effort was funded under Air Force Research Laboratory Contract FA9453-21-C-0026. The views expressed are those of the author and do not necessarily reflect the official policy or position of the Department of the Air Force, the Department of Defense, or the U.S. government. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors are employees of Spectral Sciences, Inc.

Abbreviations

The following abbreviations are used in this manuscript:
WVWorldView-3
SDSuperDove
MLMachine Learning
PIGESBCCNPhysics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network
PCNNParameterized Convolutional Neural Network
QUIDQUick Image Display
SUVSport Utility Vehicle

References

  1. Chander, G.; Hewison, T.J.; Fox, N.; Wu, X.; Xiong, X.; Blackwell, W.J. Overview of intercalibration of satellite instruments. IEEE Trans. Geosci. Remote Sens. 2013, 51, 1056–1080. [Google Scholar] [CrossRef]
  2. Fisher, G.B.; Slonecker, E.T.; Dilles, S.J.; Molnia, F.B.; Angeli, K.M. Using Global Fiducials Library High-Resolution Imagery, Commercial Satellite Imagery, Landsat and Sentinel Satellite Imagery, and Aerial Photography to Monitor Change at East Timbalier Island, Louisiana, 1953–2021; Core Science Systems and the National Civil Applications Center: Reston, VA, USA, 2023. [Google Scholar]
  3. Lew, A.J.; Stifler, C.A.; Cantamessa, A.; Tits, A.; Ruffoni, D.; Gilbert, P.U.P.A.; Buehler, M.J. Deep learning virtual indenter maps nanoscale hardness rapidly and non-destructively, revealing mechanism and enhancing bioinspired design. Matter 2023, 6, 1975–1991. [Google Scholar] [CrossRef]
  4. Lew, A.J.; Brewer, E.; Perkins, T.; Corlies, P.; Grassi, J.; Marks, J.; Friedman, A.; Jeong, L.; Sundberg, R. Enhancing Cross-Sensor Data Integration through Domain Conversion of Satellite Imagery with Hyperparameter Optimized Machine Learning. In Proceedings of the Algorithms, Technologies, and Applications for Multispectral and Hyperspectral Imaging XXXI, Orlando, FL, USA, 13–17 April 2025; SPIE: Bellingham, WA, USA, 2025. [Google Scholar]
  5. Keto, E.; Andres Watters, W. Detection of Moving Objects in Earth Observation Satellite Images: Verification. arXiv 2024, arXiv:2406.13697. [Google Scholar] [CrossRef]
  6. Sundberg, R.L.; Gruninger, J.; Nosek, M.; Burks, J.; Fontaine, E. Quick image display (QUID) model for rapid real-time target imagery and spectral signatures. In Proceedings of the Technologies for Synthetic Environments: Hardware-in-the-Loop Testing II, Orlando, FL, USA, 21–25 April 1997; SPIE: Bellingham, WA, USA, 1997; Volume 3084, pp. 272–281. [Google Scholar]
  7. Sundberg, R.; Adler-Golden, S.; Perkins, T.; Vongsy, K. Detection of spectrally varying BRDF materials in hyperspectral reflectance imagery. In Proceedings of the 7th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Tokyo, Japan, 2–5 June 2015; pp. 1–4. [Google Scholar]
  8. Perkins, T.; Sundberg, R.; Cordell, J.; Tun, Z.; Owen, M. Real-time target motion animation for missile warning system testing. In Proceedings of the Technologies for Synthetic Environments: Hardware-in-the-Loop Testing XI, Orlando, FL, USA, 17–21 April 2006; SPIE: Bellingham, WA, USA, 2006; Volume 6208, pp. 1–12. [Google Scholar]
  9. Tiwari, V.; Kumar, V.; Pandey, K.; Ranade, R.; Agarwal, S. Simulation of the hyperspectral data from multispectral data using Python programming language. In Proceedings of the Multispectral, Hyperspectral, and Ultraspectral Remote Sensing Technology, Techniques and Applications VI, New Delhi, India, 4–7 April 2016; SPIE: Bellingham, WA, USA, 2016; Volume 9880, pp. 203–211. [Google Scholar]
  10. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N. PyTorch: An Imperative Style, High-Performance Deep Learning Library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
  11. Meta Platforms, Inc. Adaptive Experimentation Platform. 2024. Available online: https://github.com/facebook/Ax (accessed on 8 April 2024).
  12. Kruse, F.A.; Lefkoff, A.B.; Boardoman, J.W. The Spectral Image Processing System (SIPS)—Interactive Visualization and Analysis of Imaging Spectrometer Data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
  13. Berk, A.; Conforti, P.; Kennett, R.; Perkins, T.; Hawes, F.; van den Bosch, J. MODTRAN6: A major upgrade of the MODTRAN radiative transfer code. In SPIE Defense and Security; International Society for Optics and Photonics: Bellingham, WA, USA, 2014; p. 90880H. [Google Scholar] [CrossRef]
  14. Lew, A.J. A Brief Survey of ML Methods Predicting Molecular Solubility: Towards Lighter Models via Attention and Hyperparameter Optimization. Preprints 2024, 2024090849. [Google Scholar] [CrossRef]
Figure 1. Error histograms and parity plots for ML converted WV vs. real SD spectral intensities (ah) per band (1–8, respectively), illustrating both (1) agreement between training, validation, and test sets and (2) accurate test predictions with r-squared values > 0.96.
Figure 1. Error histograms and parity plots for ML converted WV vs. real SD spectral intensities (ah) per band (1–8, respectively), illustrating both (1) agreement between training, validation, and test sets and (2) accurate test predictions with r-squared values > 0.96.
Geomatics 05 00035 g001
Figure 2. PIGESBCCN incorporates physics-informed Gaussian blurring, band rearrangement, temporally adjacent band chunking, and separated band prediction branches to better encode known spatial, temporal, and spectral relationships into the model architecture. Red arrows indicate the position of a moving object in scene, and colored triangles in Temporal Chunking indicate clusters of bands routed to separate convolutional branches (Satellite Imagery Credit: ©2024 Maxar/NextView License).
Figure 2. PIGESBCCN incorporates physics-informed Gaussian blurring, band rearrangement, temporally adjacent band chunking, and separated band prediction branches to better encode known spatial, temporal, and spectral relationships into the model architecture. Red arrows indicate the position of a moving object in scene, and colored triangles in Temporal Chunking indicate clusters of bands routed to separate convolutional branches (Satellite Imagery Credit: ©2024 Maxar/NextView License).
Geomatics 05 00035 g002
Figure 3. (a) Spatially matched WorldView-3 (©2024 Maxar/NextView License) and SuperDove (©2024 Planet Labs PBC, USG Plus) imagery of a scene in Turkey used to train PIGESBCCN. (b) Training and validation loss curves of hyperparameter optimized PIGESBCCN model.
Figure 3. (a) Spatially matched WorldView-3 (©2024 Maxar/NextView License) and SuperDove (©2024 Planet Labs PBC, USG Plus) imagery of a scene in Turkey used to train PIGESBCCN. (b) Training and validation loss curves of hyperparameter optimized PIGESBCCN model.
Geomatics 05 00035 g003
Figure 4. Spectral angle deviation maps for (a) WV band resampled image vs. SD baseline and the (b) ML converted image vs. SD baseline. (c) The average spectral angle difference across a sample horizontal profile (red horizontal line in panels a and b) decreases from 5.12° to 1.42° after ML conversion.
Figure 4. Spectral angle deviation maps for (a) WV band resampled image vs. SD baseline and the (b) ML converted image vs. SD baseline. (c) The average spectral angle difference across a sample horizontal profile (red horizontal line in panels a and b) decreases from 5.12° to 1.42° after ML conversion.
Geomatics 05 00035 g004
Figure 5. QUID-simulated black SUV at a roadway intersection with (a) object mask showing the location of the SUV in the RGB bands. This manifests as (b) inverted colors (green to magenta, red to cyan, blue to yellow) when the black SUV is simulated on a bright grey road. This scene is then converted from WV to (c) SD imagery via the PIGESBCCN model, where the rainbow effect is faint—but crucially still present.
Figure 5. QUID-simulated black SUV at a roadway intersection with (a) object mask showing the location of the SUV in the RGB bands. This manifests as (b) inverted colors (green to magenta, red to cyan, blue to yellow) when the black SUV is simulated on a bright grey road. This scene is then converted from WV to (c) SD imagery via the PIGESBCCN model, where the rainbow effect is faint—but crucially still present.
Geomatics 05 00035 g005
Figure 6. Taking QUID image pairs with and without a black SUV allows us to isolate and compare (a) the original target SUV, (b) the ML conversion model’s effect on the target, and (c) a baseline Gaussian blur.
Figure 6. Taking QUID image pairs with and without a black SUV allows us to isolate and compare (a) the original target SUV, (b) the ML conversion model’s effect on the target, and (c) a baseline Gaussian blur.
Geomatics 05 00035 g006
Figure 7. Per band comparison of (a) original targets, (b) target ML conversion, and (c) baseline Gaussian blurring.
Figure 7. Per band comparison of (a) original targets, (b) target ML conversion, and (c) baseline Gaussian blurring.
Geomatics 05 00035 g007
Figure 8. Quantitative agreement in (a) x coordinate target locations and (b) y coordinate target locations between ML conversion and baseline blurring.
Figure 8. Quantitative agreement in (a) x coordinate target locations and (b) y coordinate target locations between ML conversion and baseline blurring.
Geomatics 05 00035 g008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lew, A.J.; Perkins, T.; Brewer, E.; Corlies, P.; Sundberg, R. Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network for Moving Object Satellite Image Conversion. Geomatics 2025, 5, 35. https://doi.org/10.3390/geomatics5030035

AMA Style

Lew AJ, Perkins T, Brewer E, Corlies P, Sundberg R. Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network for Moving Object Satellite Image Conversion. Geomatics. 2025; 5(3):35. https://doi.org/10.3390/geomatics5030035

Chicago/Turabian Style

Lew, Andrew J., Timothy Perkins, Ethan Brewer, Paul Corlies, and Robert Sundberg. 2025. "Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network for Moving Object Satellite Image Conversion" Geomatics 5, no. 3: 35. https://doi.org/10.3390/geomatics5030035

APA Style

Lew, A. J., Perkins, T., Brewer, E., Corlies, P., & Sundberg, R. (2025). Physics-Informed Gaussian-Enforced Separated-Band Convolutional Conversion Network for Moving Object Satellite Image Conversion. Geomatics, 5(3), 35. https://doi.org/10.3390/geomatics5030035

Article Metrics

Back to TopTop