Next Article in Journal
A Study on zk-SNARK-Based RBAC Scheme in a Cross-Domain Cloud Environment
Previous Article in Journal
Effect of Drying Methods on the Physical and Surface Properties of Blueberry and Strawberry Fruit Powders: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Pulsar Candidate Recognition Algorithm Under a Multi-Scale DenseNet Framework

1
School of Mathematical Science, Guizhou Normal University, Guiyang 550025, China
2
Guizhou Key Laboratory of Information and Computing Science, Guizhou Normal University, Guiyang 550001, China
3
FAST Early Science Data Center, Guiyang 550001, China
4
School of Big Data and Computer Science, Guizhou Normal University, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(24), 13097; https://doi.org/10.3390/app152413097
Submission received: 27 August 2025 / Revised: 24 November 2025 / Accepted: 8 December 2025 / Published: 12 December 2025

Abstract

The exponential growth of candidate data from large-scale radio pulsar surveys has created a pressing need for efficient and accurate classification methods. This paper presents a novel hybrid pulsar candidate recognition algorithm that integrates diagnostic plot images and structured numerical features using a multi-scale DenseNet framework. The proposed model combines convolutional neural networks (CNNs) for extracting spatial patterns from pulsar diagnostic plots and feedforward neural networks (FNNs) for processing scalar features such as SNR, DM, and pulse width. By fusing these multimodal representations, the model achieves superior classification performance, particularly in class-imbalanced settings standard to pulsar survey data. Evaluated on a synthesized dataset constructed from FAST and HTRU survey characteristics, the model demonstrates robust performance, achieving an F1-score of 0.904 and AUC-ROC of 0.978. Extensive ablation and cross-validation analyses confirm the contribution of each data modality and the model’s generalizability. Furthermore, the system maintains low inference latency (4.2 ms per candidate) and a compact architecture (~2.3 million parameters), indicating potential for real-time deployment once validated on real observational datasets. The proposed approach offers a scalable and interpretable multimodal framework for automated pulsar classification and provides a foundation for future validation and potential integration into observatories such as FAST and the Square Kilometre Array (SKA).

1. Introduction

Pulsars are neutron stars that are highly magnetized and rotate rapidly. They emit beams of electromagnetic radiation that can be observed from Earth as periodic pulses [1]. Jocelyn Bell Burnell and Antony Hewish found the first pulsar in 1967 [2]. Since then, pulsars have become increasingly important for many astrophysical research endeavours [3]. Some of them are tests of general relativity, examining the state of ultra-dense matter, investigating binary star evolution, and contributing to the development of gravitational wave astronomy through pulsar timing arrays [4]. Finding new pulsars remains challenging, despite their importance. This is mainly because there is so much observational data, and accurate pulsar signals are pretty rare in a background full of noise and radio-frequency interference (RFI).
Discovering pulsars and determining their identities is crucial for contemporary astronomy and fundamental physics [5,6]. The 500 m Aperture Spherical radio Telescope (FAST) was finished in September 2016 in Guizhou Province, China [7]. One of its primary research objectives was to discover pulsars. As of now, the FAST Early Science Data Center has found 240 pulsar candidates [7,8,9], and 123 of these have been verified as new pulsars. Chinese radio telescopes initially discovered the pulsars J1859–0131 and J1931–01 [8,10], which was a significant step forward for China’s astronomy. J0318 + 0253 is the first millisecond pulsar (MSP) found by FAST and one of the weakest high-energy MSPs yet seen in radio flux [8]. This accomplishment highlights the significance of FAST in global efforts to detect low-frequency gravitational waves. The sky survey operations conducted by FAST generate vast amounts of candidate data [11,12,13]. For example, processing 2000 observational data points each day yields approximately 300,000 diagnostic images [14]. This underscores the critical need for automated, high-efficiency candidate screening methods.
FAST generates approximately 300,000 candidate plots daily, making manual inspection infeasible. In order to deal with this flood of data, RFI excision, dedispersion, Fourier domain searching, and folding all necessitate specialist processing software like PRESTO (Ransom 2001) [14]. This pipeline ends in the creation of diagnostic graphs, including the time-phase graph, the frequency-phase graph, the DM curve graph, and the pulse profile diagram. Such plots play a critical role in confirming pulsar prospects, which are typically generated by experienced astronomers or amateurs using citizen science projects like the Pulsar Search Collaboratory. Nevertheless, manual verification is not scalable and not reproducible due to the massive size of data.
Machine learning and deep learning algorithms have become of interest to address these limitations, as an automated method of selecting pulsar candidates. The existence of structured datasets, including the High Time Resolution Universe (HTRU) dataset and the FAST early science dataset, has allowed enormous steps forward in this area. The Parkes Radio Telescope was used to obtain the HTRU dataset that consists of 1196 confirmed pulsars and 89,996 noise candidates, demonstrating a severe class imbalance (the ratio of positive to negative cases is 1:75). Likewise, the FAST dataset is composed of 1160 positive and 14,319 negative candidates (imbalance ratio of 1:12), which provides more challenges to the classifier construction.
Early methods featured hand-engineered features, i.e., SNR, DM spacing, chi-square statistics, and peak scores, evaluated through common machine learning models, i.e., Support Vector Machines (SVMs), decision trees, and ensemble classifiers [14]. These models were quite successful; however, because they used feature engineering, their flexibility and ability to generalize across different survey scenarios were limited. Even more recent studies have applied convolutional neural networks (CNNs) to directly analyze diagnostic plots, therefore requiring less manual feature engineering [9]. CNNs have the capability to effectively learn hierarchical spatial patterns which qualifies them to be used in locating peculiar pulsar signals embedded in noisy or partially ruined plots.
One of the first to display the application of deep CNNs to pulsar classification was Lyon et al. [6], who achieved higher accuracy and robustness. Zhu et al. [9] extended it further to include picture pattern recognition and tried out hybrid models, which mix image and statistical inputs. Later, Zhang et al. [15] introduced a custom deep learning model for pulsar classification which significantly reduced false positive rates at maintained high sensitivity. Nevertheless, such models mostly tend to rely either on image data or numerical properties, not taking advantage of the synergistic nature of multimodal data fusion.
In this respect, this paper proposes a new hybrid pulsar candidate detection algorithm based on a multi-scale DenseNet network. We combine convolutional processing of diagnostic plots with a feedforward neural network (FNN) which consumes numerical features including SNR, DM, pulse width and FFT-based scores. This two-input architecture permits this model to extract both geometrical and statistical properties of candidate signals. The representations of the features are combined and run through a classification layer to probabilistically estimate the probability that each candidate is a real pulsar.
The principal contributions of the work are three-fold. We first build a unified multimodal classification model, which combines the DenseNet based image processing and FNN based feature embedding, allowing more precise and generalizable predictions. Second, we propose a balanced and synthesized dataset, inheriting the properties of both FAST and HTRU observations to mitigate the effects of class imbalance problems, while maintaining the authenticity of the signal. Third, we perform thorough analysis by five-fold cross-validation, ablation experiments and error analysis to compare our model with the current baselines.
Our results demonstrate that the proposed model outperforms both CNN-only and FNN-only baselines across all key performance metrics, including accuracy, precision, recall, F1-score, and AUC-ROC. Moreover, the lightweight design (approximately 2.3 million parameters) and fast inference time (4.2 ms per candidate) make it suitable for real-time candidate screening in large-scale pulsar surveys. This hybrid framework not only aligns with the demands of modern radio astronomy but also sets a precedent for future developments in multimodal deep learning applications for astrophysical signal classification.

2. Materials and Methods

2.1. Dataset Construction and Composition

To test the proposed pulsar candidate recognition system, we designed a high-resolution artificial dataset comprising 20,000 instances. This dataset was designed to mimic the statistical and structural characteristics of two often-cited realistic pulsar candidate datasets: the Five-hundred-meter-Aperture Spherical radio Telescope (FAST; Pingtang County, Guizhou Province, China) and the High Time Resolution Universe (HTRU) survey conducted using the Parkes 64-m radio telescope (CSIRO, Parkes, NSW, Australia). In detail, the dataset comprises 4000 positive examples (pulsars) and 16,000 negative examples (radio frequency interference or noise), making it realistic for real-world survey conditions with an imbalance factor of 1:4. It should be noted that the 1:4 ratio reflects an early-stage candidate filtering scenario and does not represent the extreme class imbalance (1:10 to 1:106) observed in full-scale FAST and HTRU survey outputs.
The samples are identified uniquely, bear a binary label (1 if pulsar, 0 otherwise) and a source tag (FAST or HTRU subset) showing whether the candidate was sampled in the simulated FAST or HTRU subset. A total of 12 numerical features were generated on every sample. To ensure reproducibility of feature generation, each of the first 12 numeric characteristics was randomly sampled from a distribution adjusted to the typical FAST and HTRU candidate characteristics. More specifically, the value of the signal-to-noise ratio was randomly sampled from a lognormal distribution, the DM from a uniform distribution ranged from 5 to 600 pc cm−3, the pulse periods from a bimodal Gaussian distribution were suitable for regular pulsars and MSPs, the pulse widths were from a truncated Gaussian distribution, and the chi-squared statistics were from a gamma distribution. The remaining diagnostic factors, skewness, kurtosis, FFT peak score, folding RMS, maximum peak ratio, profile sharpness index, and frequency drift factor, were randomly sampled from bounded Gaussian/uniform distributions to model the observed variations in the real pulsar candidate samples.
Four diagnostic plots were also attached to each candidate which emulate typical visual outputs in the manual screening of pulsars: the integrated pulse profile, the time-versus-phase diagram, the frequency-versus-phase diagram, and the DM SNR curve. The integrated profiles were developed from 1 to multiple Gaussian components depending on the sampled pulse width. At the same time, the time-phase and frequency-phase plots were derived by repeating the pulse template across subintegrations and frequency channels, incorporating white noise, red noise, and weak sine modulations to model RFI-like behaviors. The DM–S/N plots were generated by incorporating dispersion smearing across trial DMs. The four plots were created as 128 × 128-pixel grayscale PNG images, following the FAST and HTRU display conventions. Although the dataset is synthetic, it was constructed with direct reference to published statistical distributions and imaging characteristics observed in FAST and HTRU datasets. It serves as a proxy for algorithm development, model ablation, and performance benchmarking in the absence of real-time access to proprietary observational data. Final testing on real HTRU2 or FAST datasets remains necessary for validating field deployment [8]. A complete synthetic generation pipeline was applied, combining statistical sampling of numerical features with pulse-template modeling, dispersion smearing, and controlled noise/RFI injection to ensure realistic variability consistent with published FAST and HTRU diagnostic characteristics.

2.2. Preprocessing and Feature Normalization

Before training, all numerical features underwent z-score standardization to guarantee a compatible scale and convergence stability [6]. Diagnostic images were resized to 128 × 128 pixels and mean zero, unit variance was normalized to make feature extraction in convolutional layers stable. Filtering or clipping of any missing or outlier values in the feature matrix was performed to empirically bounded values to ensure numerical stability.

2.3. Dataset Partitioning

The entire data was randomly stratified into training (70%), validation (15%), and test (15%) sets. The relative classes were maintained in all the subsets so that the exposure to pulsar and non-pulsar candidates was kept constant across training and assessment. Moreover, the model performance stability and generalizability were evaluated using five-fold cross-validation with a random seed varying.

2.4. Model Architecture

The recognition algorithm proposed employs a dual-branch design, where a multi-scale DenseNet backbone is utilized to process images, and a fully connected feedforward neural network (FNN) is used to process structured numerical inputs [16]. The DenseNet architecture can be characterized as a ‘multi-scale’ architecture because it provides a connection pattern that combines the outputs of all previous convolutional blocks and passes them along to the next layers. This essentially enables the model to consider the outputs of the fine-scale representations in the early layers (with a smaller receptive field) as well as the coarse-scale representations in the deep layers (with a larger receptive field). The DenseNet branch consists of four 3 × 3 convolutional blocks, each accompanied by batch normalization and ReLU activation functions. These blocks extract visual features in a hierarchical way from the diagnostic plots. Simultaneously, the 1D feature vector also undergoes another independent feedforward neural network (FNN) consisting of two hidden layers of 128 and 64 neurons, respectively, with ReLU activations and dropout. The outputs of both branches are concatenated and forwarded to a final classification head, which consists of a sigmoid activation and a fully connected layer function, producing a probabilistic prediction of pulsar candidacy (0: non-pulsar, 1: pulsar).
In our implementation, the DenseNet branch uses a compact four-block configuration with a growth rate of 16 and a layer distribution of {4, 4, 4, 4} per block, followed by 2 × 2 average pooling between blocks and global average pooling at the end, producing a 256-dimensional feature vector. The FNN depth (128 and 64 units) was selected after exploratory grid-search experiments balancing performance and computational efficiency, and dropout (p = 0.5) was applied to reduce overfitting. The multimodal fusion step concatenates the 256-dimensional CNN embedding with the 64-dimensional FNN output to form a joint 320-dimensional representation that is passed to a fully connected classification head. This configuration provides sufficient representational capacity while remaining efficient enough for inference-time deployment.

2.5. Training Procedure

Model training was conducted using the PyTorch framework (version 2.0; Meta Platforms, Inc., Menlo Park, CA, USA). with an initial learning rate of 0.001, the Adam optimizer was used (Kingma, 2014). The binary cross-entropy loss function was modified with class weights to compensate for class imbalance. With early stopping based on stagnation in the validation loss, each model was trained for up to 50 epochs. A batch size of 128 was used, and dropout regularization was applied in both branches of the network to mitigate overfitting. The schematic workflow of the hybrid pulsar candidate classification model is presented in Figure 1. The Adam optimizer was selected for its stability and adaptability during the training of multimodal neural networks, and for its widespread use in prior pulsar-candidate classification studies. The initial learning rate of 0.001 follows the default recommendation in the original Adam formulation and provides reliable convergence behavior in convolutional architectures without requiring extensive tuning. Empirically, this learning rate produced stable training curves in our experiments.
The system integrates diagnostic plots and numerical features as dual input modalities. Diagnostic images are preprocessed and passed through a CNN based on DenseNet architecture, while structured 1D features (e.g., SNR, DM, pulse width) are processed through a feedforward neural network (FNN). Extracted features from both paths are combined in a joint feature space and forwarded to a classification module, which outputs a prediction of either pulsar or non-pulsar. Model performance is evaluated using five-fold cross-validation with F1-score and AUC as key metrics. All experiments were performed on an NVIDIA RTX 3060 Laptop GPU (6 GB VRAM; NVIDIA Corporation, Santa Clara, CA, USA) paired with an Intel i7 processor (Intel Corporation, Santa Clara, CA, USA) and 16 GB system memory. The hybrid DenseNet–FNN model required approximately 4.8 GB of GPU memory during training. Using a batch size of 128 and 50 epochs, the full training procedure took approximately 58 min.

2.6. Evaluation Metrics

Using the standard classification metrics, accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC-ROC), the model performance evaluation was performed on the held-out test set. Additionally, confusion matrices were analyzed to quantify the rate of false negatives and false positives. To assess computational efficiency, we also reported the model’s total parameter count, training time per epoch, and inference latency per candidate.

3. Results

3.1. Overall Model Performance

The proposed hybrid pulsar recognition model, integrating a multi-scale DenseNet and fully connected feature-processing branch, was trained on a stratified subset (14,000 samples) and evaluated on a held-out test set comprising 3000 samples. The model achieved strong classification performance, particularly in terms of sensitivity and overall discriminative capability. Table 1 presents a quantitative evaluation of the model on the test dataset. The classification accuracy reached 96.4%, with a recall of 89.2% and a precision of 91.7%, resulting in the F1-score of 90.4%. Indicating excellent separability between the pulsar and non-pulsar classes, the area under the receiver operating characteristic curve (AUC-ROC) was 0.978.
To further illustrate classification quality, the receiver operating characteristic (ROC) curve is shown in Figure 2, confirming the model’s strong true positive rate across a range of classification thresholds. The confusion matrix in Figure 3 reveals a balanced distribution of both correct and incorrect predictions. The model maintains high sensitivity to pulsar instances with minimal false negatives and demonstrates a low false positive rate for noise candidates.

3.2. Training Stability and Convergence Behavior

Model training was stable across epochs, as demonstrated by accuracy and loss curves presented in Figure 4. Both training and validation accuracy showed consistent improvement and convergence without overfitting, while the corresponding loss curves decreased steadily and plateaued near epoch 40.

3.3. Baseline Model Comparison

To assess the effectiveness of the hybrid architecture, two baseline models were developed for comparison: a DenseNet-only model using image inputs and a feedforward neural network (FNN) model using only numerical features. Both baselines were trained under the same conditions and evaluated on the same test data. As shown in Table 2, the hybrid model outperformed both baselines across all performance metrics. The DenseNet-only model achieved 93.8% accuracy and 87.8% F1-score, while the FNN-only model attained slightly lower values of 92.7% accuracy and 86.7% F1-score. The hybrid model’s superior performance demonstrates the value of multimodal feature fusion.
Additionally, with an average precision (AP) score of 0.958, the precision–recall (PR) curve in Figure 5 reinforces the ability of the model to maintain high precision over a wide range of recall values.

3.4. Ablation Study on Diagnostic Plot Inputs

An ablation study was carried out to investigate the contribution of each diagnostic image input to the final classification accuracy. For each experiment, one of the four plots (frequency-phase, pulse profile, time-phase, or DM curve) was omitted from the model input while keeping all other settings constant. Table 3 presents the results of this ablation. The removal of either the time-phase or frequency-phase plot resulted in the most significant reductions in accuracy, suggesting that these components carry the most salient visual information for pulsar discrimination. The absence of the DM curve or pulse profile plot resulted in minor performance drops, yet these were still notable relative to the entire configuration. All ablation experiments were evaluated using five-fold cross-validation, and the values reported in Table 3 correspond to the mean performance across folds. The variance of accuracy and F1-score across folds was consistently below 0.01, indicating that the observed performance differences are stable and not attributable to random variation. Although no formal statistical hypothesis test (e.g., paired t-test) was applied, the low fold-to-fold variance supports the robustness of the ablation results.

3.5. Cross-Validation Results

To evaluate generalizability, 5-fold cross-validation was performed. As shown in Table 4, the model exhibited low variance across folds, indicating stable performance and minimal overfitting. All folds maintained accuracy above 95% and AUC values near 0.976.

3.6. Misclassification Analysis

An error analysis was conducted on a random sample of misclassified candidates. Table 5 presents a selection of 10 such cases, highlighting common reasons for misclassification such as weak SNR, overlapping DM curves, ambiguous profile shapes, or RFI-like patterns. This analysis supports the model’s interpretability and provides valuable insights for future model improvement.

4. Discussion

This study presents a hybrid machine learning framework that leverages both convolutional and feedforward neural networks for the automated classification of pulsar candidates. By combining visual diagnostic plots and structured numerical features, the proposed model demonstrates both high accuracy and generalizability, advancing current methods in large-scale radio pulsar surveys.

4.1. Multimodal Feature Fusion Enhances Classification

The model’s strong overall performance attests to the effectiveness of integrating heterogeneous data modalities as it achieved 0.904 F1-score and 0.978AUC-ROC score. Diagnostic plots revealed spatial and morphological patterns characteristic of pulsars, particularly in the time-phase and frequency-phase diagrams, while numerical descriptors, such as SNR, DM, and pulse width, provided quantitative, high-resolution insights. This approach builds upon prior work that has explored CNN-based models for pulsar candidate images [6] and scalar-feature classifiers, such as SVMs and decision trees [9]. The feature fusion module successfully synthesizes these complementary representations, allowing the model to make informed decisions across a broad range of candidate types.

4.2. Comparative Advantage over Single-Modality Baselines

The hybrid architecture consistently outperformed both image-only and feature-only baselines across all evaluation metrics (Table 2), highlighting the added value of combining diagnostic plots with structured data. While previous studies have shown the utility of convolutional models on FAST and HTRU2 data [17], and some have explored GAN-augmented feature selection [18], few have integrated both 1D and 2D modalities. Our dual-branch model bridges this gap, offering an empirically validated framework for automated classification.

4.3. Model Generalization and Training Stability

Five-fold cross-validation revealed minimal performance variability across folds, with accuracy consistently above 95% and AUC-ROC values exceeding 0.97 (Table 4). These results confirm the model’s training stability and resilience to overfitting, attributable to the use of class-weighted loss functions, dropout regularization, and standardized feature scaling. Comparable methods, such as PulsarNet and PICS, have demonstrated similar training strategies to manage imbalance and convergence [15,19].

4.4. Error Analysis and Interpretability

The misclassification analysis also helped to better understand the model behavior and possible methods of improvement. The majority of false negatives were linked to either faint pulsar signals or incomplete dispersion profile, which was expected based on previous FAST observations in which low-SNR candidates were often missed by automated filters [8]. False positives tended to have RFI contaminated patterns or pulse-like characteristics without any regular periodicity. Such results agree with the existing accounts that highlight the difficulty in categorizing borderline or high-noise candidates in mega-surveys [6,17]. We expanded the misclassification analysis from 10 to over 80 randomly selected false positives and false negatives. This broader inspection confirmed the same dominant failure patterns noted in Table 5: (i) low-SNR pulsars with weak or fragmented profiles, (ii) candidates with overlapping or broadened DM curves, and (iii) strong RFI patterns that visually resemble pulsar-like periodicity. No new systematic failure mode emerged beyond those already discussed, indicating that the diagnostic behaviors observed in the initial subset are representative of the model’s broader misclassification trends.

4.5. Application Potential in Large-Scale Surveys

The lightweight nature of the model (~2.3 million parameters), as well as its efficiency in inference (4.2 ms per candidate), makes the suggested model suitable to be utilized in contemporary pulsar discovery pipelines. It is particularly promising with respect to large-scale telescopes like FAST, SKA, and CHIME, in which the number of daily candidate images can be large enough (hundreds of thousands or more). The model structure can be adapted to semi-real-time classification pipeline (e.g., PRESTO-based candidate filtering systems) [15], and the dual-input model structure reflects the decision-making mechanism of human reviewers, providing a high level of interpretability in addition to speed.

4.6. Limitations and Future Directions

Although the synthetic dataset presented in this study was calibrated against real FAST and HTRU survey statistical profiles, an essential future step is to verify the model on robust observational data. The next steps will include optimizing the model on the HTRU2 dataset [9] and testing the model performance on other radio telescopes, with different observational noise characteristics. In particular, we plan to fine-tune the hybrid DenseNet–FNN architecture on HTRU2 candidates and benchmark it against existing published models, followed by validation on FAST early-science pulsar catalogs as access to labeled candidates becomes available. Further improvements can be obtained with attention-based CNNs or transformer hybrids, which have recently become promising in astronomical image analysis [15,20]. Domain adaptation and semi-supervised learning methods should be explored as well, especially in dealing with rare pulsar subtypes and noisy training labels [18,21]. For context, we also provide a comparison with related pulsar candidate classification approaches reported in the literature. Lyon et al. (2016) [6] achieved an accuracy of 0.99 on HTRU1 and HTRU2, Zhang et al. (2019) [15] reported an F1 score of 0.92 using a CNN, Balakrishnan et al. (2021) [18] obtained an F1-score of 0.95 with Semi-Supervised Generative Adversarial Network, and Yin et al. (2022) [19] reported an F1-score of 0.99 using a deep neural network model. Although these results are not directly comparable to ours due to differences in datasets and feature preprocessing, they provide a meaningful performance reference. A full quantitative comparison using the HTRU2 or FAST early-science dataset will be conducted as part of future validation work. Future work will include an extended architectural ablation analysis, comparing alternative DenseNet depths, different fusion strategies, and other CNN backbones such as ResNet and EfficientNet to evaluate the design choices in the proposed framework more comprehensively.

5. Conclusions

A hybrid pulsar candidate recognition algorithm was proposed in this work within a multi-scale DenseNet framework, which aims at combining the diagnostic plot images and structured numerical features to achieve robust classification. The proposed model exploits both spatial and statistical information that is complementary in nature by integrating the convolutional and feedforward neural network branches. This model was trained and tested on a synthesized dataset of real-world observational patterns in FAST and HTRU surveys, which provides the possibility of strict benchmarking of class-imbalanced conditions. We have shown in our experiments that the hybrid model is much better than single-modality baselines in terms of various evaluation measures with an F1-score of 0.904, an AUC-ROC of 0.978, and an accuracy of more than 96 percent. Ablation studies confirmed the contribution of each diagnostic plot type, while cross-validation demonstrated the model’s stability and generalizability.
Notably, the system maintained low inference latency (4.2 ms/sample), suggesting potential for real-time integration into survey pipelines once validated on real observational data. Beyond performance, the model’s interpretability, as revealed through misclassification analysis, offers valuable insights for refining candidate verification strategies, and its modular design provides a flexible foundation for future enhancements. The research indicates that combining multimodal inputs within a unified neural architecture can enhance pulsar candidate detection, even in noisy and high-volume datasets. Future work will focus on further validating the model using live observational data from FAST and the upcoming SKA, expanding its applicability to transient signal detection, and exploring semi-supervised and transformer-based variants for deeper feature abstraction. The proposed approach makes a meaningful contribution to the automation of pulsar discovery and exemplifies the broader utility of multimodal deep learning in radio astronomy.

Author Contributions

Conceptualization, J.T. and X.X. (Xiangguang Xiong); methodology, J.T.; software, J.T.; validation, J.T., X.X. (Xiangguang Xiong) and X.X. (Xiaoyao Xie); formal analysis, J.T.; investigation, J.T.; resources, X.X. (Xiangguang Xiong); data curation, J.T.; writing—original draft preparation, J.T.; writing—review and editing, X.X. (Xiangguang Xiong); visualization, X.X. (Xiangguang Xiong); supervision, X.X. (Xiangguang Xiong); project administration, X.X. (Xiangguang Xiong); funding acquisition, X.X. (Xiangguang Xiong) All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Astronomy Research Center FAST Major Achievements Cultivation Project, Chinese Academy of Sciences [2019sr04]; the National Key Research and Development Plan [2017YFA0402600]; the National Natural Science Foundation of China [U1631132, U1731238, U1831131, and 11743002]; and the Strategic Pilot Science and Technology Project of the Chinese Academy of Science [XDB23000000].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Philippov, A.; Kramer, M. Pulsar Magnetospheres and Their Radiation. Annu. Rev. Astron. Astrophys. 2022, 60, 495–558. [Google Scholar] [CrossRef]
  2. Bell Burnell, J. The Past, Present and Future of Pulsars. Nat. Astron. 2017, 1, 831–834. [Google Scholar] [CrossRef]
  3. Cordes, J.M.; Kramer, M.; Lazio, T.J.W.; Stappers, B.W.; Backer, D.C.; Johnston, S. Pulsars as Tools for Fundamental Physics & Astrophysics. New Astron. Rev. 2004, 48, 1413–1438. [Google Scholar] [CrossRef]
  4. Lorimer, D.R.; Kramer, M. Handbook of Pulsar Astronomy; Cambridge University Press: Cambridge, UK, 2005; Volume 4, ISBN 0-521-82823-6. [Google Scholar]
  5. Burgay, M.; Possenti, A. Radio Pulsars: An Astrophysical Key to Unlock the Secrets of the Universe; American Institute of Physics: College Park, MD, USA, 2011; ISBN 0-7354-0915-3. [Google Scholar]
  6. Lyon, R.J.; Stappers, B.W.; Cooper, S.; Brooke, J.M.; Knowles, J.D. Fifty Years of Pulsar Candidate Selection: From Simple Filters to a New Principled Real-Time Classification Approach. Mon. Not. R. Astron. Soc. 2016, 459, 1104–1123. [Google Scholar] [CrossRef]
  7. Xu, Y.; Li, D.; Liu, Z.; Wang, C.; Wang, P.; Zhang, L.; Pan, Z. Application of Artificial Intelligence in the Selection of Pulsar Candidate. Prog. Astron. 2017, 35, 304–315. [Google Scholar]
  8. Qian, L.; Pan, Z.; Li, D.; Hobbs, G.; Zhu, W.; Wang, P.; Liu, Z.; Yue, Y.; Zhu, Y.; Liu, H. The First Pulsar Discovered by FAST. Sci. China Phys. Mech. Astron. 2019, 62, 959508. [Google Scholar]
  9. Zhu, W.W.; Berndsen, A.; Madsen, E.C.; Tan, M.; Stairs, I.H.; Brazier, A.; Lazarus, P.; Lynch, R.; Scholz, P.; Stovall, K. Searching for Pulsars Using Image Pattern Recognition. Astrophys. J. 2014, 781, 117. [Google Scholar] [CrossRef]
  10. Wang, P.; Han, J.; Yang, Z.; Wang, T.; Wang, C.; Su, W.; Xu, J.; Zhou, D.; Yan, Y.; Jing, W. The FAST Galactic Plane Pulsar Snapshot Survey: VIII. 116 Binary Pulsars. Res. Astron. Astrophys. 2025, 25, 014003. [Google Scholar] [CrossRef]
  11. Djorgovski, S.G.; Mahabal, A.; Donalek, C.; Graham, M.; Drake, A.; Turmon, M.; Fuchs, T. Automated Real-Time Classification and Decision Making in Massive Data Streams from Synoptic Sky Surveys. In Proceedings of the 2014 IEEE 10th International Conference on e-Science, Sao Paulo, Brazil, 20–24 October 2014; IEEE: New York, NY, USA, 2014; Volume 1, pp. 204–211. [Google Scholar]
  12. Poudel, M.; Sarode, R.P.; Watanobe, Y.; Mozgovoy, M.; Bhalla, S. A Survey of Big Data Archives in Time-Domain Astronomy. Appl. Sci. 2022, 12, 6202. [Google Scholar] [CrossRef]
  13. Wan, X.; Han, X.; Wang, J. Computing Prominent Skyline on Massive Data. Data Sci. Eng. 2025, 10, 117–146. [Google Scholar] [CrossRef]
  14. Yu, Q.-Y.; Pan, Z.-C.; Qian, L.; Wang, S.; Yue, Y.-L.; Huang, M.-L.; Hao, Q.-L.; You, S.-P.; Peng, B.; Zhu, Y. A PRESTO-Based Parallel Pulsar Search Pipeline Used for FAST Drift Scan Data. Res. Astron. Astrophys. 2020, 20, 91. [Google Scholar] [CrossRef]
  15. Zhang, H.; Zhao, Z.; An, T.; Lao, B.; Chen, X. Pulsar Candidate Recognition with Deep Learning. Comput. Electr. Eng. 2019, 73, 1–8. [Google Scholar] [CrossRef]
  16. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  17. Pang, D.; Goseva-Popstojanova, K.; McLaughlin, M. Detection of Radio Pulsars in Single-Pulse Searches Within and Across Surveys. Publ. Astron. Soc. Pac. 2020, 132, 094502. [Google Scholar] [CrossRef]
  18. Balakrishnan, V.; Champion, D.; Barr, E.; Kramer, M.; Sengar, R.; Bailes, M. Pulsar Candidate Identification Using Semi-Supervised Generative Adversarial Networks. Mon. Not. R. Astron. Soc. 2021, 505, 1180–1194. [Google Scholar] [CrossRef]
  19. Yin, Q.; Wang, Y.; Zheng, X.; Zhang, J. Pulsar Candidate Recognition Using Deep Neural Network Model. Electronics 2022, 11, 2216. [Google Scholar] [CrossRef]
  20. Zhang, C.J.; Shang, Z.H.; Chen, W.M.; Xie, L.; Miao, X.H. A Review of Research on Pulsar Candidate Recognition Based on Machine Learning. Procedia Comput. Sci. 2020, 166, 534–538. [Google Scholar] [CrossRef]
  21. Wang, W.; Wang, J.; Ye, X.; Zhang, Y.; Li, J.; Du, X.; Cai, W.; Wu, H.; Zhang, T.; Jiao, Y. Application of AI Technology in Pulsar Candidate Identification. Astron. Tech. Instrum. 2024, 2, 27–43. [Google Scholar] [CrossRef]
Figure 1. Schematic workflow of the hybrid pulsar candidate classification model.
Figure 1. Schematic workflow of the hybrid pulsar candidate classification model.
Applsci 15 13097 g001
Figure 2. ROC curve with true positive rate and false positive rate (AUC = 0.978).
Figure 2. ROC curve with true positive rate and false positive rate (AUC = 0.978).
Applsci 15 13097 g002
Figure 3. Confusion matrix of predicted vs. actual labels on test set.
Figure 3. Confusion matrix of predicted vs. actual labels on test set.
Applsci 15 13097 g003
Figure 4. Training and validation accuracy/loss over 50 epochs.
Figure 4. Training and validation accuracy/loss over 50 epochs.
Applsci 15 13097 g004
Figure 5. Precision–recall (PR) curve for the hybrid model.
Figure 5. Precision–recall (PR) curve for the hybrid model.
Applsci 15 13097 g005
Table 1. Performance classification of proposed model on the test dataset.
Table 1. Performance classification of proposed model on the test dataset.
MetricValue
Accuracy0.964
AUC-ROC0.978
F1-Score0.904
Inference Time4.2 ms/sample
Parameters~2.3 million
Precision0.917
Recall0.892
Table 2. Performance comparison between proposed and baseline models.
Table 2. Performance comparison between proposed and baseline models.
ModelAccuracyRecallF1-ScoreAUC-ROC
DenseNet (Image)0.9380.8510.8780.951
FNN (1D features)0.9270.8340.8670.944
Hybrid (Proposed)0.9640.8920.9040.978
Table 3. Accuracy changes resulting from the exclusion of individual diagnostic plots.
Table 3. Accuracy changes resulting from the exclusion of individual diagnostic plots.
Input ConfigurationAccuracy
Full input (4 plots + 1D features)0.964
Without a time-phase plot0.917
Without a frequency-phase plot0.922
Without a DM curve plot0.951
Without a pulse profile plot0.942
Only 1D features (no images)0.927
Only diagnostic plots (no 1D features)0.938
Table 4. Five-fold cross-validation performance summary.
Table 4. Five-fold cross-validation performance summary.
FoldAccuracyPrecisionRecallF1-ScoreAUC-ROC
Fold 10.9640.9100.8870.8980.978
Fold 20.9610.9180.8830.9000.973
Fold 30.9590.9120.8850.8980.974
Fold 40.9680.9200.8960.9080.981
Fold 50.9620.9210.9020.9110.974
Table 5. Representative misclassified samples and possible reasons.
Table 5. Representative misclassified samples and possible reasons.
Sample IDTruePredConf.SNRDMNote
SAMPLE_00001100.4974.663.2Weak signal; SNR < threshold
SAMPLE_00002010.5333.922.5RFI pattern misclassified
SAMPLE_00003100.4625.280.1Overlapping peaks in DM curve
SAMPLE_00004010.5817.112.3Profile visually resembles pulsar
SAMPLE_00005100.4463.740.7Low dispersion mimicry
SAMPLE_00006010.5596.419.4Frequency-phase noise structure
SAMPLE_00007100.4894.150.6Ambiguous profile shape
SAMPLE_00008010.5422.911.1Flat profile despite high SNR
SAMPLE_00009100.4743.888.2Time-phase pattern unclear
SAMPLE_00010010.5075.525.6High chi-square; no structure
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, J.; Xie, X.; Xiong, X. Efficient Pulsar Candidate Recognition Algorithm Under a Multi-Scale DenseNet Framework. Appl. Sci. 2025, 15, 13097. https://doi.org/10.3390/app152413097

AMA Style

Tang J, Xie X, Xiong X. Efficient Pulsar Candidate Recognition Algorithm Under a Multi-Scale DenseNet Framework. Applied Sciences. 2025; 15(24):13097. https://doi.org/10.3390/app152413097

Chicago/Turabian Style

Tang, Junlin, Xiaoyao Xie, and Xiangguang Xiong. 2025. "Efficient Pulsar Candidate Recognition Algorithm Under a Multi-Scale DenseNet Framework" Applied Sciences 15, no. 24: 13097. https://doi.org/10.3390/app152413097

APA Style

Tang, J., Xie, X., & Xiong, X. (2025). Efficient Pulsar Candidate Recognition Algorithm Under a Multi-Scale DenseNet Framework. Applied Sciences, 15(24), 13097. https://doi.org/10.3390/app152413097

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop