1. Introduction
Atomic fluorescence spectrometry (AFS) and inductively coupled plasma optical emission spectrometry (ICP-OES) are fundamental analytical techniques widely employed for elemental quantification in environmental monitoring, food safety, geological exploration, clinical biomonitoring, and related fields [
1,
2]. AFS enables femtogram-level detection of trace toxicological elements via high-brightness atomic fluorescence that is generated in the ultraviolet (UV) region [
3]. In contrast, ICP-OES enables simultaneous multi-line detection by exploiting broad-spectrum atomic emission that spans 165–900 nm and covers the UV, visible, and near-infrared spectral ranges [
4]. In contemporary instrumentation, UV imaging detection systems are predominantly based on a charge-coupled device (CCD) and complementary metal–oxide–semiconductor (CMOS) technologies. However, substantial dark current noise inherently limits their capacity for weak UV signal acquisition [
5,
6]. Although intensified CCDs (ICCDs) can achieve high-performance detection of weak radiation spectra, they require deep cryogenic cooling to suppress circuit dark noise and to operate at high readout frame rates. The associated costs and complexity therefore restrict their widespread adoption within the aforementioned fields [
7]. Conversely, the single photon-counting imaging detector (SPC imaging detector) can achieve high-sensitivity imaging in the UV band without the need for deep cooling, offering advantages such as low power consumption, compactness, and system simplicity [
8]. The wedge-and-strip anode (WSA) SPC imaging detector, which was developed by a team at the Changchun Institute of Optics, Fine Mechanics and Physics (CIOMP), Chinese Academy of Sciences, has been applied to n the EUV camera aboard the Chang’e-7 lunar probe as well as in the wide-angle auroral imager on the Feng-Yun-3D meteorological satellite [
9,
10,
11]. Despite these merits, further enhancement of the dynamic range and spatial resolution remains a priority. To this end, the digital micromirror device (DMD) can be adapted to operate in the UV band through UV-enhancement processing, thereby providing a new technical pathway that enables advanced UV imaging detection [
12]. As a reflective spatial light modulator, the DMD offers several intrinsic advantages, including a small micromirror unit size, a high switching speed, and the absence of macroscopic moving parts [
13,
14]. By coupling a UV-enhanced DMD with a space-qualified SPC imaging detector, a UV single-photon imaging system has been developed. This system enables highly sensitive SPC imaging detector of UV spectral signals with low dark count rates, small effective pixel sizes, and high spatial light modulation speeds, without imposing additional cryogenic cooling requirements. However, the lack of one-to-one spatial correspondence between DMD pixels and the detector’s sensitive elements, together with inherent nonlinear distortion introduced by the SPC imaging detector, results in pronounced geometric distortion in the acquired spectral images. In spectral regions with dense line distributions, even pixel-level drift can severely degrade quantitative accuracy. Currently, a systematic calibration and distortion correction model tailored to this specific UV imaging chain remains unavailable.
Geometric distortion in digital imaging systems significantly compromises image fidelity by introducing complex pixel displacements and structural deformations [
15,
16]. These distortions typically arise from multiple sources during image acquisition, including optical lenses, image detectors, and other optical path components [
17]. While existing research predominantly addresses lens-induced distortion, conventional correction techniques generally require multi-view image sets to achieve high-precision geometric calibration. However, the logistical constraints of acquiring multi-view data across diverse application scenarios severely limit the practical utility of these approaches [
18,
19]. Furthermore, the performance of feature-based distortion correction methods, such as those relying on checkerboard corner detection, is inherently constrained by feature localization accuracy. Consequently, these approaches often fail to robustly handle scenes characterized by complex, non-standard distortions [
20,
21]. Recently, deep learning architectures have demonstrated substantial progress in distortion correction. For example, the RDTR framework leverages a unified geometric model combined with model-aware pre-training on images of arbitrary resolution to effectively correct radial distortion [
22]. Similarly, methods such as DR-GAN adopt end-to-end generative frameworks integrated with self-supervised strategies, which directly learn structural mappings through low-to-high perceptual losses, thereby enabling single-stage, real-time radial distortion correction without explicit calibration parameters [
23,
24]. Nevertheless, the efficacy of these learning-based methods in complex distortion scenarios remains limited, as it depends strongly on the consistency between predefined distortion priors and the actual distortion characteristics. Constrained by these theoretical primitives, current data-driven approaches remain difficult to adapt for the unique, idiosyncratic distortions of practical UV imaging systems.
The intrinsic nonlinear distortion of the array-based single-photon counter is typically corrected using polynomial mapping models. This strategy avoids explicit modeling of the physical distortion mechanisms and directly fits a multivariate polynomial to represent the spatial transformation from the detector output coordinate system to an ideal reference coordinate system. Owing to its compact formulation, computational robustness, and ease of calibration, this strategy has become a widely adopted approach for geometric distortion correction to date [
25,
26].
At the system level, Feng Wei et al. investigated pixel alignment between a DMD and a CCD-based imaging system at the pixel scale [
27,
28]. In their approach, pixel-level alignment between DMD micromirrors and CCD pixels is first achieved via Moiré fringe feedback. Subsequently, the direct linear transformation (DLT) algorithm is employed to compute a homography matrix, thereby establishing a global coordinate mapping from the DMD mask to the CCD in a single step. This procedure enables pixel-wise light intensity modulation and geometric distortion correction. However, the DLT method is applicable only to scenarios involving a small field of view and pure rotation, and it is unable to correct the severe distortions and non-planar imaging conditions encountered in UV optical systems.
This paper focuses on a UV single-photon imaging system that integrates a DMD and a SPC imaging detector. The overall system architecture is first described. Distortion correction is then investigated, addressing both the intrinsic distortion of the SPC imaging detector and the mapping mismatch between the DMD and the SPC imaging detector. To this end, a cascaded method combining polynomial fitting with multilayer perceptron (MLP) regression is introduced. A high-precision spatial mapping model is constructed and experimentally validated, supporting pixel-level image reconstruction and distortion compensation. Consequently, the quantitative accuracy and stability of the UV imaging system are substantially improved, providing a reliable pixel-level basis for subsequent elemental quantification. This advance could further improve the performance of high-end UV spectrometers.
2. Systems and Methods
This section systematically describes the constituent modules of the UV single-photon imaging system and analyzes the underlying optical principles. To address the limitations of the current UV imaging detector, particularly the limited dynamic range, a UV single-photon imaging system is developed to meet the specified requirements. This system integrates the advantages of the SPC imaging detector (namely high sensitivity and low noise in the UV band) with the features of the DMD, including small pixel size and a high spatial light modulation rate. As illustrated in
Figure 1, the system primarily comprises three components: a SPC imaging detector, a DMD, and a coupling optics module. The coupling optics module includes a convex mirror, a concave mirror, and an optical chamber. The DMD is positioned at the front end of the optical of the UV single-photon imaging system, functioning as a high-speed binary optical shutter. When the DMD micromirrors are in the “on” state, incident light from the object passes through the coupling optical system and forms an image on the SPC imaging detector. Conversely, in the “off” state, the light path is blocked, preventing light from reaching the SPC imaging detector. This mechanism enables programmable two-dimensional gating, supports real-time modulation with user-defined coded masks, and facilitates UV imaging.
Figure 2 illustrates the hardware prototype of the UV single-photon imaging system used in the experiment. The DMD used is a DLP9500 (manufactured by Texas Instruments in Dallas, TX, USA), with a resolution of 1920 × 1080 pixels and a micromirror pitch of 10.8 µm. The host computer performs real-time programming of the DMD driver circuit through the dedicated control software, enabling microsecond-scale micromirror state refresh. A self-developed WSA SPC imaging detector is used, with an equivalent pixel size of approximately 60 µm. Powered by a high-voltage supply (HV Supply), the detector generates output charge pulses, which are first pre-amplified and pulse-shaped by the front-end electronics (FEE). The pulses are subsequently sampled by a high-speed analog-to-digital acquisition circuit (HS-ADAC) and transmitted to image control software for event-level reconstruction. Additionally, the DMD operates via time-division flipping with a micromirror pitch of 10.8 µm, this enables time-division image acquisition at the detector and thereby enhances resolution.
The incident spectral signal passes through the optical system to the SPC imaging detector, where photon events are detected by partitioning the charge across three electrodes. The detector’s inherent distortion is the primary source of system-level distortion. Although the DMD enables pixel-level spatial modulation of incoming photons, misalignment between the DMD and the detector can introduce additional geometric distortion. These distortions are secondary to the intrinsic distortion of the SPC imaging detector, which predominantly affects overall image quality. Accurate pixel-level modulation requires a precise correspondence between DMD micromirrors and detector pixels. This alignment is essential for controlling the photon flux incident on each pixel. The DMD projects the modulated signal onto the detector, where events are localized by three-electrode charge partitioning. When the electron cloud is excessively large or small, edge regions can exhibit S-shaped and sinusoidal modulation distortions. These distortions are corrected using a series of dot-array masks loaded sequentially onto the DMD. After propagation through the optical system, the corresponding dot-array images are recorded by the detector, enabling acquisition of end-to-end distortion samples for accurate calibration and correction.
To address the errors described above, a hybrid correction framework, termed “polynomial fitting combined with MLP regression,” is proposed and evaluated. This framework consists of four main steps, as shown in
Figure 3: (i) acquisition of the distorted image, (ii) feature point matching, (iii) initial correction using polynomial fitting, and (iv) fine pixel-level correction with an MLP network.
First, a regularly spaced dot-array mask is generated over the DMD imaging plane, with the pattern centered on a reference pixel. The mask is loaded onto the DMD, and the micromirrors are flipped accordingly. Next, the UV light source is activated. The modulated beam propagates through the optical path and forms a distorted dot-array image on the SPC imaging detector, thereby completing the acquisition of the distorted image. Then, based on the gray-scale distribution and centroid deviation of the distorted image, an appropriate feature point matching method is selected. The feature point strategy is optimized according to the observed gray-scale distribution and centroid deviation pattern of the dots in the distorted image. Preliminary correction is then performed using polynomial fitting, followed by fine pixel-level correction with an MLP network.
During the correction process, the image is input as a two-dimensional matrix with grayscale values. The main task is to establish a high-precision mapping between distorted and reference coordinates, which is used to update the feature points. The remaining pixels are then interpolated under this mapping to generate the final corrected image. The following sections describe feature-point extraction and the construction of the polynomial and MLP models.
2.1. Feature Point Extraction
A dot-array mask is generated on the DMD imaging plane, centered on a reference pixel, with a 29-pixel step between adjacent dots. After reflection by the DMD and propagation through the coupling optical system, the light forms a distorted image on the SPC imaging detector. array image. Before extracting feature points (dot centroids) from the distorted image, preprocessing is applied to the dot-array image. The distorted image is represented by Equation (1):
where
denotes the image domain, and
represents the global maximum grayscale value. To reduce background noise interference, threshold segmentation is applied using a threshold
, empirically set to 20% of the maximum grayscale value. Pixels with grayscale values higher than the threshold
are identified as valid feature points, while all others are considered background and excluded from subsequent calculations. The resulting binary image is represented by Equation (2):
Connected component analysis is performed on the binary image
, yielding N independent regions denoted as
, as shown in Equation (3):
The regions are filtered by area, and only components with more than 4 pixels are retained as valid regions. For each valid region
, the zero-order and first-order moments are calculated to obtain its centroid coordinates, as shown in Equation (4):
The centroids of the dots in the distorted image are calculated using the described algorithm and adopted as the feature point coordinates.
Furthermore, based on the imaging principle, distortion is minimal near the image center and is therefore treated as negligible. Nine control points in the central region are used to generate a reference image by inverse projection under an ideal geometric relationship. The dot-array mask is translated across the DMD plane in 1-pixel increments to acquire images sequentially. In total, 30 pairs of distorted and reference images are acquired, and feature-point matching is performed for each pair.
2.2. Polynomial Model
The system images control points with known spatial positions. The mapping relationship between the distorted image coordinates and reference coordinates is described by a set of higher-order polynomial equations, as shown in Equation (5):
where
and
represent the pixel coordinates in the distorted and reference images, respectively;
and
denote the correction coefficients; and N is the order of the polynomial. More severe distortion generally requires a higher polynomial order
, increasing the number of coefficients
and
. For an Nth-order polynomial, the required number of correction coefficients is
. Based on practical correction performance, the polynomial order was set to 4. Five sets of matched feature-point coordinates were selected, and the coefficients
and
were estimated by least squares. The remaining 25 distorted images were then mapped into the reference coordinate system using the estimated polynomial coefficients. Bilinear interpolation was used for grayscale resampling, yielding coarsely corrected images in geometry. These images provide the initial mapping for the subsequent refinement stage.
2.3. MLP Model
The preliminarily corrected images obtained through polynomial fitting are subjected to a second round of feature-point extraction. Correspondences between feature points in the corrected images and those in the reference images are used to train an MLP model. The MLP model is adopted to capture nonlinear residual distortions that are poorly modeled by polynomial fitting. Polynomial models effectively capture coarse distortions and are often used as an initial calibration step; however, their ability to correct residual distortions is limited. In contrast, the MLP model can refine the remaining distortions beyond the polynomial stage. As shown in
Figure 4, the network consists of four cascaded fully connected blocks. Each block, except the output block, includes a linear layer followed by a nonlinear activation function, forming a deep feedforward architecture. Network weights are learned iteratively until the loss converges. The learned mapping fits the training samples and shows spatial generalization. Specifically, given an input coordinate in the imaging plane, the trained model predicts the corresponding coordinate in the reference space, enabling high-precision mapping across the field of view.