Next Article in Journal
TMP-M2Align: A Topology-Aware Multiobjective Approach to the Multiple Sequence Alignment of Transmembrane Proteins
Previous Article in Journal
Texture-Adaptive Fabric Defect Detection via Dynamic Subspace Feature Extraction and Luminance Reconstruction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

OIKAN: A Hybrid AI Framework Combining Symbolic Inference and Deep Learning for Interpretable Information Retrieval Models

by
Didar Yedilkhan
,
Arman Zhalgasbayev
*,
Sabina Saleshova
and
Nursultan Khaimuldin
*
“Smart City” Research Center, Astana IT University, 010000 Astana, Kazakhstan
*
Authors to whom correspondence should be addressed.
Algorithms 2025, 18(10), 639; https://doi.org/10.3390/a18100639
Submission received: 29 July 2025 / Revised: 3 October 2025 / Accepted: 4 October 2025 / Published: 10 October 2025
(This article belongs to the Section Algorithms for Multidisciplinary Applications)

Abstract

The rapid expansion of AI applications in various domains demands models that balance predictive power with human interpretability, a requirement that has catalyzed the development of hybrid algorithms combining high accuracy with human-readable outputs. This study introduces a novel neuro-symbolic framework, OIKAN (Optimized Interpretable Kolmogorov–Arnold Network), designed to integrate the representational power of feedforward neural networks with the transparency of symbolic regression. The framework employs Gaussian noise-based data augmentation and a two-phase sparse symbolic regression pipeline using ElasticNet, producing analytical expressions suitable for both classification and regression problems. Evaluated on 60 classification and 58 regression datasets from the Penn Machine Learning Benchmarks (PMLB), OIKAN Classifier achieved a median accuracy of 0.886, with perfect performance on linearly separable datasets, while OIKAN Regressor reached a median R2 score of 0.705, peaking at 0.992. In comparative experiments with ElasticNet, DecisionTree, and XGBoost baselines, OIKAN showed competitive accuracy while maintaining substantially higher interpretability, highlighting its distinct contribution to the field of explainable AI. OIKAN demonstrated computational efficiency, with fast training and low inference time and memory usage, highlighting its suitability for real-time and embedded applications. However, the results revealed that performance declined more noticeably on high-dimensional or noisy datasets, particularly those lacking compact symbolic structures, emphasizing the need for adaptive regularization, expanded function libraries, and refined augmentation strategies to enhance robustness and scalability. These results underscore OIKAN’s ability to deliver transparent, mathematically tractable models without sacrificing performance, paving the way for explainable AI in scientific discovery and industrial engineering.

1. Introduction

Neuro-symbolic methods occupy a distinctive place in the history of AI, marking a key milestone in the evolution of machine intelligence. Early efforts relied on symbolic approaches, hand-crafted rules, and logic-based systems that excelled at knowledge manipulation and reasoning but proved inadequate for learning from raw data [1]. The subsequent rise in sub-symbolic techniques, notably artificial neural networks and statistical learning enabled automatic feature extraction and pattern recognition; however, these models often functioned as opaque “black boxes” offering little insight into learned relationships [2]. The convergence of these paradigms into neuro-symbolic AI represents a crucial advance toward more general and robust intelligence, addressing longstanding challenges such as commonsense reasoning and structured knowledge representation. Although symbolic regression yields human-readable expressions that reveal underlying data-generating laws, it incurs prohibitive computational costs and limited scalability.
Modern science, in particular physics, biology, and chemistry, increasingly relies on data-driven modeling. However, this requires models that reveal the underlying structure of phenomena, not just the accuracy of prediction [3,4]. In such cases, interpretability is not just a requirement for confidence but plays an important role in discovery [5]. Although traditional methods such as symbolic regression produce human-readable results, they are computationally expensive and cannot scale [6,7]. Conversely, Deep neural networks (DNNs) are renowned for their ability to generalize across high-dimensional spaces; however, the internal representations they learn often reside in complex, opaque latent structures [8,9]. By integrating neural learning with symbolic formalisms, neuro-symbolic AI seeks to bridge this interpretability gap without sacrificing performance, thereby enabling both high-level reasoning and transparent, verifiable insights into data.
Traditional deep architectures like Multilayer Perceptron lack semantic transparency. While MLPs can approximate any continuous function according to the universal approximation theorem [10], their learned parameters, dense weight matrices, and non-linear activations are difficult to decode into meaningful rules. Kolmogorov–Arnold Networks (KANs) were proposed as a more interpretable alternative, inspired by the Kolmogorov–Arnold Representation Theorem (KART), which states that any multivariate continuous function can be expressed as a sum of univariate functions [11]. KANs replace weight matrices with learnable univariate transformations, allowing visualization and partial interpretability [12]. However, they suffer from high memory and time complexity and often fail to outperform tuned MLPs in large-scale settings [13].
This paper introduces a novel neuro-symbolic learning framework inspired by KART, OIKAN (Optimized Interpretable Kolmogorov–Arnold Network), designed to address the limitations of both MLPs and symbolic regression models. Building upon the foundations of earlier KAN architectures, OIKAN incorporates lightweight differentiable symbolic basis components, applies regularization techniques to enhance simplicity and reduce overfitting, and adopts an interface similar to scikit-learn to facilitate model introspection and symbolic extraction. The framework has been applied to a variety of scientific problems, including the discovery of symbolic equations in fluid dynamics, interpretable modeling of gene expression, and robust simulations in chemical reaction networks. The results confirm its ability to deliver competitive accuracy while producing analytically tractable and scientifically meaningful equations, positioning it as a promising tool for advancing interpretable artificial intelligence in scientific research. In the context of a smart city, OIKAN offers a robust foundation for building interpretable models capable of revealing complex relationships in heterogeneous urban datasets. Its ability to generate symbolic representations from sensor data can facilitate transparent decision-making in real-time environmental monitoring, traffic optimization, and air quality forecasting. By embedding the OIKAN framework into the analytical pipeline, the project can enhance explainability, reproducibility, and scientific insight-key factors for sustainable urban development and evidence-based policymaking.

2. Related Studies

2.1. Trustworthy and Interpretable AI

The need for trustworthy and interpretable AI has become crucial in modern information systems. While deep learning models deliver high predictive power, their “black-box” nature limits validation, risk assessment, and bias detection in sensitive areas such as healthcare, law, and industrial automation. To address these issues, hybrid neuro-symbolic approaches are being explored, combining the generalization of deep models with the transparent reasoning of symbolic systems [14].
Research on trustworthy AI identifies six key reliability dimensions: safety and robustness, non-discrimination, explainability, privacy, accountability, and environmental well-being. Among these, explainability and accountability are vital for information retrieval tasks, where reproducibility and traceability are essential. Although post hoc explainability methods like SHAP, LIME, and Integrated Gradients provide insights, they often lack stability, causal validity, and scalability. This highlights the value of interpretable-by-design models, where reasoning is embedded within the architecture [15].
The rapid growth of large language models (LLMs) in retrieval and recommendation systems has amplified transparency concerns. Despite alignment techniques such as supervised fine-tuning and reinforcement learning with human feedback, LLMs frequently produce hallucinations, biases, and inconsistencies. Taxonomies of LLM trustworthiness outline numerous sub-dimensions, including reasoning fidelity and misuse resistance, yet the lack of granular interpretability mechanisms limits their auditability and adoption in evidence-critical domains [16].

2.2. Kolmogorov–Arnold Networks (KANs) and Variants

In high-stakes fields, interpretable AI models are as important as accuracy, ensuring stakeholders can verify and trust outputs. Comparative studies show that while MLPs outperform KANs in traditional tasks like vision, NLP, and audio, KANs excel in symbolic formula discovery due to their B-spline activations [17]. Other research indicates that modified KANs, especially those based on orthogonal polynomials, can rival MLP-based methods in scientific tasks such as differential equation solving, though they remain sensitive to initialization and architecture choices [18].
Various KAN architectures have been proposed recently. The original PyKAN implemented KAN layers with cubic B-spline activations. Follow-ups sought to improve efficiency or flexibility. Efficient-KAN provided engineering optimizations (vectorized implementations) to speed up training [19]. Fast-KAN replaces B-splines with Gaussian radial basis functions (RBFs), yielding ~3.3× faster inference (albeit still more costly than an MLP). Fourier-KAN (KAF) uses a Fourier-series-based activation instead of a single spline, combining several periodic components; it achieved strong results on vision, NLP, audio, and even PDE tasks [20]. Other variants include wavelet-based Wav-KAN, rKAN (rational functions), and fKAN (fractional B-splines). The core idea remains as the need to replace each MLP weight by a small function so that the network can both learn features (via its layer structure) and optimize univariate functions with high precision.

2.3. KANs in Scientific and Physics-Informed Learning

A comprehensive survey by [21] reviews KAN theory and practice. They note that unlike MLPs, KANs do not fix activations but optimize them, providing enhanced interpretability and generalization. For example, the research [22] shows small KAN models reaching or surpassing MLPs in curve fitting. Authors of [23] apply KANs to time series and hyperspectral imaging, finding that KANs capture complex dependencies effectively.
The authors of [24] introduce KINN (Kolmogorov–Arnold-Informed Neural Network) for physics-informed learning. They substitute MLPs in PINNs with KANs. In a variety of solid mechanics PDEs, KINN significantly outperforms MLP-PINNs in accuracy and convergence speed. They highlight that KANs require fewer parameters and provide interpretability due to their decomposition of inputs into scalar functions. In essence, KAN-based networks like KINN demonstrate that KART-inspired architectures can yield both efficient learning and physically meaningful structure.
However, there is debate about KAN’s novelty. Some note that a KAN can be reparametrized as a conventional MLP with a special structure (so representationally they are equivalent). Nonetheless, in practice, KANs act as neural architectures with built-in function priors. OIKAN leverages this by using KART to decompose its network: it builds hierarchical univariate transformations that mirror the theorem’s form [25]. This allows OIKAN to expose each intermediate univariate function as a symbolic sub-model, greatly aiding interpretability.
In the context of life sciences and remote sensing, authors of [26,27] show how neuro-symbolic methods can augment data efficiency, enable knowledge injection, and enhance explainability. These applications underscore the growing relevance of neuro-symbolic frameworks in real-world scientific workflows, where understanding the model’s rationale is as vital as accuracy.

2.4. Symbolic Regression Approaches

Symbolic regression (SR) aims to learn analytic expressions (e.g., algebraic equations) that fit data, without pre-specifying a model form. Classic SR methods include Genetic Programming (GP), Bayesian symbolic regression, and tools like Eureqa. GP-based SR randomly evolves expression trees of operators and functions. For example, Schmidt and Lipson’s Eureqa system [28] automatically rediscovered physical laws (Hamiltonians, Lagrangians) from time-series data, demonstrating the power of GP in uncovering interpretable scientific formulas. Bayesian SR methods extend GP by encoding priors or scoring complexity, yielding more concise models. As one review explains, GP “has been traditionally and commonly used” for SR, but “suffers from limitations” such as overly complex expressions and difficulty incorporating domain knowledge [29]. Bayesian SR can address some of these issues by imposing prior distributions over functions to promote interpretability. A systematic comparative analysis of prominent symbolic regression approaches was conducted, focusing on key criteria such as typical use cases, computational speed, model interpretability, accuracy on complex datasets, and the availability of software tools. Table 1 summarizes the findings, highlighting the trade-offs that researchers and practitioners must consider when selecting an SR method for real-world applications.
Closely related are sparse linear models with basic functions. Methods like LASSO and ElasticNet [30] fit linear combinations of features while enforcing sparsity via L1 (and L2) penalties. These yield interpretable coefficients (many zeros) and can incorporate non-linear basis expansions (e.g., splines) for smooth curves. Penalized linear models remain widely used as interpretable baselines: [14] compared their SR approach (FEAT) against sparse logistic regression (essentially a LASSO on expanded features) and found FEAT attained equal or higher AUC with far simpler models.
Table 1. Comparative analysis of symbolic regression methods.
Table 1. Comparative analysis of symbolic regression methods.
MethodUse CaseSpeedInterpretability Accuracy (Complex Data)Tools/
Libraries
Genetic Programming (GP)Non-linear symbolic modelingLow (slow, resource-heavy) HighHighOperon, gplearn, PySR [31]
Least SquaresHybrid symbolic fittingHighMedium (model-dependent)MediumOperon, PySINDy [32]
Sparse Regression (LASSO, STLSQ)Sparse interpretable modelsMedium (with regularization)HighMedium (needs basis functions)Lasso, PySINDy [33]
Bayesian SRUncertainty-aware modelingLow (compute-intensive)High (probabilistic)MediumCustom implementations [29]
NN-Based (AI Feynman)Symbolic distillation from neural networksLow (slow training) Medium (postprocessed)High (low-noise data) AI Feynman [34]

2.5. Neuro-Symbolic Frameworks

Researchers have shown near-perfect performance on complex visual reasoning tasks by combining deep vision models with symbolic execution. For example, ref. [35] proposed a neuro-symbolic VQA system that parses an image into a scene graph and a question into a symbolic program, executing the program to answer with 99.8% accuracy on CLEVR, while providing interpretable reasoning. Similarly, in NLP, neural language models can generate logical assertions or knowledge graphs for symbolic reasoning tasks such as question answering or theorem proving, a pattern reflected in recent neuro-symbolic NLP work integrating neural text understanding with ontologies and logic-based reasoners [36].
Hybrid neuro-symbolic models leverage the strengths of both approaches. MLP-KAN [37], for instance, combines an MLP “representation learner” with a KAN “function learner,” achieving strong results in computer vision, NLP, and formula regression by dynamically selecting between neural and symbolic branches. More broadly, neuro-symbolic reasoning integrates deep learning with symbolic logic to achieve both robust learning and interpretable reasoning [38]. Approaches include differentiable logic layers, symbolic memory components, and program induction frameworks. DreamCoder [39], for example, learns domain-specific programs by combining neural pattern recognition with symbolic program synthesis, where symbolic modules serve as interpretable reasoning layers.

2.6. Positioning of OIKAN

Within OIKAN, knowledge learned from data is represented by sparse symbolic expressions generated during the second pipeline phase. Once the KAN-augmented MLP has internalized an internal representation of the training set, the symbolic regression module distills that representation into an explicit formula. In this hybrid mechanism, the neural component provides mappings from raw data to higher-level feature estimates, while the symbolic component supplies the explicit structure for final decision-making. The resulting AI system delivers deterministic, verifiable inference alongside transparent decision explanations derived from known symbolic relations. Notably, this design aligns with the view that robust AI requires both pattern learning and symbolic reasoning [40]. OIKAN uses neural computation to learn from examples, but entrusts the symbolic layer to formalize that learned knowledge and handle it in a principled, interpretable manner. By bridging these aspects, neuro-symbolic systems are considered a promising route toward AI that learns from experience while it can reason and explain in human-like terms [41,42].
From an ethical standpoint, providing a symbolic explanation for a prediction fosters trust: stakeholders gain the ability to trace how specific inputs mathematically produced a given output, simplifying the identification of biases or errors in the decision process. This practice aligns with emerging guidelines for trustworthy AI, which prioritize explainability and transparency as fundamental principles. Unlike opaque deep learning models, the hybrid neuro-symbolic methodology delivers a level of algorithmic transparency capable of reducing the risk that hidden discriminatory patterns remain undetected [36].
To assess the positioning of OIKAN within the broader landscape of symbolic regression and neuro-symbolic machine learning tools, a detailed comparative analysis is provided in Table 2. This analysis categorizes each tool by its core algorithmic paradigm. The comparison highlights both the diversity and the limitations of existing approaches, underscoring the need for frameworks like OIKAN that balance accuracy, transparency, and practical deployment.

2.7. Research Gaps

In summary, the literature underscores that interpretability is essential in scientific and high-stakes domains [43] because black-box models carry risks that simple, explainable models avoid. Despite these advances, important gaps remain. There is a lack of integrated, open-source frameworks that seamlessly combine neural and symbolic components for scientific data. Most symbolic or KAN toolboxes are research prototypes with limited engineering support. Moreover, many comparative studies show that model performance and efficiency vary widely by application [44], indicating the need for flexible systems that can adapt to diverse datasets.
OIKAN is motivated by these gaps: it aims to provide a unified, open neuro-symbolic framework that scales to real data and produces symbolic formulas. By integrating KAN layers with conventional architectures, OIKAN seeks to harness neural pattern learning while maintaining interpretability and symbolic output. In essence, OIKAN addresses the call for AI systems that are both accurate and explainable, fulfilling the need for transparent, reproducible discovery tools in science and beyond.
Table 2. Cross-Method Comparison of the Proposed Framework and Existing Learning Paradigms.
Table 2. Cross-Method Comparison of the Proposed Framework and Existing Learning Paradigms.
SourceToolAlgorithm
Type
Time & Memory EfficiencyTabular Data
Focus
Multi-
Obj.
Neural
Integration
Interpretability
[45]PySRGenetic
Programming
Medium (efficient Julia backend, moderate memory)Partial (can integrate
with ML pipelines)
YesYesHigh
[46]OperonGenetic
Programming
High (fast C++ implementation, low memory)NoYesYesHigh
[47]GPLearnGenetic
Programming
Low (Python-based, slower)NoNoNoHigh
[48]EQL/
DeepSymReg
Neural NetworkMedium (GPU-
accelerated,
moderate memory)
YesNoNoMedium
[49]PyKANMathematical NetMediumPartial (mathematical nets)NoNoMedium
[50]DSRReinforcement LearningLow
(computationally intensive, slower convergence)
YesNoNoMedium
[51]QLatticeGraph-basedMediumNoNoNoHigh
[52]SR-TransformerTransformer-basedMedium to High (depends on model size, GPU heavy)YesNoNoLow
[53]EureqaEvolutionaryMediumNoYesYesHigh
ProposedOIKANNeuro-Symbolic High (lightweight MLPs)Yes (MLPs)Adapt.Adapt.High

3. Materials and Methods

3.1. High-Level Overview and Main Contributions

This study explores a diverse set of interpretable AI models and techniques, including classical algorithms, neural network architectures such as multilayer perceptrons (MLPs) and Kolmogorov–Arnold Networks (KANs), symbolic regression, and hybrid neural-symbolic approaches. The primary objective is to develop a computationally efficient model capable of producing transparent mathematical expressions that characterize the relationship between input features and target variables for both classification and regression tasks using tabular data.
The core model developed, OIKAN v0.0.3, introduces a unified neuro-symbolic framework that integrates lightweight neural architectures with differentiable symbolic basis functions such as sin(x), log(x), and exp(x). Unlike conventional neural networks that operate as black boxes, OIKAN generates compact and interpretable symbolic formulas while retaining the predictive power of modern machine learning models. This dual capability directly addresses one of the central challenges in AI: balancing predictive accuracy, interpretability, and computational efficiency.
A key novelty of OIKAN lies in its flexible architecture and configurable design. It supports both classification and regression tasks within the same unified pipeline, incorporates symbolic regression based on ElasticNet regularization for controlled formula complexity, and offers adaptive parameters (evaluate_nn and augmentation_factor) to balance automation with experimental control. By bridging the gap between symbolic transparency and neural expressiveness, OIKAN contributes to advancing trustworthy, explainable, and domain-agnostic AI.

3.2. Mathematical Foundation of the OIKAN

The OIKAN framework was developed with a focus on model interpretability, robustness, and computational efficiency, combining deep learning and symbolic regression in a unified architecture. The following subsections describe the theoretical foundations and implementation components that constitute the OIKAN pipeline.

3.2.1. Kolmogorov–Arnold Representation Theorem (KART)

The design of OIKAN is rooted in the Kolmogorov–Arnold Representation Theorem (KART), which guarantees that any continuous multivariate function f: [0, 1]n → R can be decomposed into a series of univariate continuous functions. Specifically, it can be represented as:
f x 1 , , x n = q = 0 2 n Φ q p = 1 n ψ p , q x p ,
where Φ q and ψ p , q are continuous univariate functions. This decomposition enables the transformation of complex mappings into tractable hierarchical compositions [54], guiding the symbolic regression stage in OIKAN for enhanced interpretability.

3.2.2. Neural Network Function Approximation (TabularNet)

To extract non-linear patterns from tabular data, OIKAN employs a feedforward neural network module referred to as “TabularNet”. This network is designed to approximate a function f: Rn → Rm, where n denotes the number of input features and m represents the output dimension (i.e., m = 1 for regression tasks, or m equal to the number of classes in classification problems) [55].
The general structure of the neural network can be described as:
h x = W L σ L 1 W L 1 σ L 2 σ 1 W 1 x + b 1 + b L 1 + b L ,
where
  • W l and b l are the weight matrix and bias vector for the layer l .
  • σ l is the activation function used at layer l .
  • L is the number of layers, with sizes defined by input_size, hidden_sizes, and output_size.
Training is performed using the Adam optimizer. The loss function depends on the task:
  • For regression, Mean Squared Error (MSE):
L = 1 N i = 1 N y i y i ^ 2 ,
  • For classification, Cross-Entropy Loss:
L = 1 N i = 1 N c = 1 C y i , c log y i , c ^ ,
The trained model generates augmented target values for symbolic regression, capturing non-linear patterns in the data.

3.2.3. Gaussian Noise-Based Data Augmentation

To improve generalization and enhance robustness, OIKAN applies Gaussian noise to the input data during training [56]. Given an input dataset X R N × n , where N is the number of data samples and n is the dimensionality of each sample, the augmented dataset is generated as:
X aug = X + ϵ k k = 1 , , K ,
ϵ k N 0 , σ 2 ,
where σ is the standard deviation specified by the sigma parameter. This step improves the robustness of both the neural and symbolic models to small perturbations in input features.

3.2.4. Two-Stage Sparse Symbolic Regression

OIKAN incorporates a two-stage sparse symbolic regression mechanism using ElasticNet regularization, combining L1 (LASSO) and L2 (Ridge) regularization to produce sparse, interpretable models [57]. The target function is approximated as a linear combination of basis functions:
y ^ = i = 1 M β i ϕ i x ,
where
  • ϕ i x are basis functions (e.g., polynomials, logarithms, exponentials, sines).
  • β i are coefficients estimated by ElasticNet.
  • M is the number of basis functions.
Stage 1: Coarse Model with Polynomial Features
In the initial stage, a coarse approximation of the target function is constructed by expanding the input space with polynomial basis functions up to the second degree. The resulting feature set is defined as:
Φ coarse x = 1 , x i , x i 2 , x i x j i , j = 1 , , n , i j ,
This representation captures linear, quadratic, and pairwise interaction terms among the input variables. ElasticNet regression is then applied to fit the model by solving the following optimization problem:
m i n β = 1 N i = 1 N y i j = 1 M coarse β j ϕ j x i 2 + α λ | | β | | 1 + 1 λ 2 | | β | | 2 2 ,
where
  • ( β R M coarse )  is the vector of model coefficients.
  • ϕ j x i is the j-th feature (or basis function) evaluated at input x i .
  • N is the number of training samples.
  • α is the regularization strength (alpha parameter, α   >   0 ).
  • M coarse is the number of polynomial features.
  • λ 0,1 : mixing parameter (default value: 0.5):
    λ = 1 : Lasso (L1) penalty only.
    λ = 0 : Ridge (L2) penalty only.
    0   <   λ <   1 : ElasticNet (mixture of both).
  • | β | 1 : L1 norm, promoting sparsity.
  • | β | 2 : L2 norm, promoting small coefficient magnitudes.
The use of ElasticNet promotes sparsity through the L1 term and prevents overfitting via the L2 term, making it effective in handling multicollinearity and selecting the most relevant features. This approach facilitates the generation of simplified, interpretable expressions, even in the presence of redundant or highly correlated features.
To identify the most informative variables, feature importance scores k are computed by aggregating the absolute contributions of each variable across all basis functions in which it appears:
I k = j : k features ϕ j β j ,
The top k features with the highest importance scores are then selected for the next stage.
Stage 2: Refined Model with Non-linear Transformations
In the second stage, the model is enriched by incorporating non-linear transformations of the previously selected top-k features. This refined feature set includes:
Φ additional x = { x i 3 , log 1 + x i , exp clip x i , 10,10 , sin x i i top - k ,
The final input space is defined as the union of polynomial and non-linear basis functions:
m i n β = 1 N i = 1 N y i j = 1 M refined β j ϕ j x i 2 + α λ | | β | | 1 + 1 λ 2 | | β | | 2 2 ,
ElasticNet is applied once again to this enriched set using the same objective function, now with Mrefined basis functions. The resulting symbolic model retains only non-zero coefficients and the corresponding basis functions, yielding a concise and interpretable mathematical expression.
For multi-class classification tasks, this symbolic modeling procedure is repeated independently for each class.

3.2.5. Symbolic Expression Generation

The final model comprises symbolic expressions that approximate the target function:
f x = j S β j ϕ j x ,
where S = {   j   : β j >   10 6 } denotes the set of indices corresponding to non-negligible coefficients. In classification scenarios, OIKAN constructs one symbolic expression per class. The predicted class probabilities are then obtained via a softmax function applied to the symbolic logits:
P y = c x = exp f c x c = 1 C e x p ( f c ( x ) ) ,
The evaluation of OIKAN models was carried out using a set of standard metrics for both classification and regression tasks. These metrics provide a comprehensive overview of model performance in terms of accuracy, error, and computational efficiency. A summary of the metrics, including their formulas, types, and descriptions, is presented in Table 3.
To ensure experimental reproducibility and consistency across tasks, a standardized sequence of steps was followed—encompassing data loading, model initialization, training, symbolic expression extraction, and performance evaluation. Algorithm 1 outlines the complete symbolic modeling pipeline using either the OIKANClassifier or OIKANRegressor.
Algorithm 1 Unified Symbolic Modeling Using OIKANClassifier or OIKANRegressor
Require: Labeled dataset (X,y); OIKANClassifier hyperparameters/OIKANRegressor hyperparameters; test set proportion t∈(0,1)
Ensure: Trained symbolic model; extracted symbolic formulas; evaluation metrics (accuracy or R2); feature importance metrics
1: Load the dataset (e.g., Iris for classification, California Housing for regression)
2: Split data into training and testing subsets using proportion t
3: Initialize the OIKANClassifier or OIKANRegressor with the specified hyperparameters
4: Train the model on Xtrain,ytrain, including neural network evaluation and symbolic regression
5: Predict class labels or target values for the test set using model.predict(X_test)
6: For classification: compute and print accuracy using accuracy_score(y_test, y_pred; For regression: evaluate performance using the R2 score
7: Generate and print a classification report using classification_report(y_test, y_pred)
8: Retrieve and print symbolic formulas for each class using model.get_formula()
9: Retrieve and print feature importance values using model.feature_importances()
10: Save the trained model to disk using model.save(path)
11: Reload the model using model.load(path)
12: Verify the consistency of symbolic formulas in the original format from the loaded model
13: Retrieve and print symbolic formulas in SymPy format using get_formula(type = ‘sympy’)
14: Retrieve and print symbolic formulas in LaTeX format using get_formula(type = ‘latex’)
15: return Trained model, symbolic formula representations, accuracy, and feature importances

3.3. Architecture and Base Class Parameters for the OIKAN

Model training begins by importing the OIKAN classes and initializing them with appropriate hyperparameters, such as hidden layer sizes, activation functions (e.g., ReLU, tanh), augmentation factors, and batch size. The user invokes the fit(X, y) method to initiate training on tabular datasets, followed by prediction using the predict(X) method. Symbolic formula extraction is performed using the “get_formula()” function, which can return outputs in SymPy, LaTeX, or a full mathematical representation. Models can be saved and loaded via the “save()” and “load()” functions, respectively, using JSON format.
Table 4 provides an overview of the main workflow stages implemented in the OIKAN framework. The system begins with preprocessing and data augmentation, proceeds through a two-stage symbolic regression process, and concludes with model compilation into a human-readable formula.
The framework integrates neural representation learning with symbolic regression techniques, allowing the generation of interpretable mathematical models from data. The system was implemented in Python 3.13.0 and made available as an installable package via PyPI (pip install -qU oikan). The framework consists of core components, including the OIKANRegressor and OIKANClassifier, accessible via the oikan.model module, while utility functions reside in oikan.utils (Figure 1).
Internally, the OIKAN abstract base class orchestrates the workflow. For regression and classification tasks, it delegates the learning process to specialized subclasses (OIKANRegressor or OIKANClassifier). The learning pipeline includes a feedforward neural network component (TabularNet), which learns intermediate representations. Data augmentation is achieved via Gaussian noise injection, controlled by the augmentation_factor parameter. The overall class structure and component dependencies are illustrated in Figure 2.
Once the neural model captures the data structure, the symbolic regression phase begins. This stage transforms the features using a predefined set of basic functions, including powers (xn), logarithmic (log(x)), sinusoidal (sin(x)), and exponential (exp(x)) transformations. The symbolic regression utilizes a two-stage ElasticNet model. Initially, it fits the transformed data and then prunes irrelevant or low-importance terms based on thresholding and a Top K selection approach. The final symbolic expression is extracted and returned as an interpretable mathematical formula. The architecture that connects these components is shown in Figure 3.
The framework ensures reproducibility through modular architecture and compatibility with multiple programming languages (C++, C, Python, Go, Rust, JavaScript) via model compilation tools. Code, data augmentation routines, and formula extraction methods are openly accessible through the OIKAN repository. Users must cite the library and include configuration details to replicate the symbolic learning results.
Dataflow within the framework begins with raw input data fed into the neural component. The feedforward network undergoes training with augmented data, followed by transformation using basis functions. Symbolic regression and term pruning are then conducted, culminating in mathematical formula extraction and optional compilation. This modular workflow from input to interpretable output is illustrated in Figure 4.
To ensure adaptability and optimization for a wide range of tasks, the OIKAN framework was designed with a high degree of parameter flexibility. This design enables users to fine-tune models for specific applications, thereby enhancing performance and interpretability. Table 5 summarizes the key parameters available in the OIKAN base class, applicable to both classification and regression tasks.
Among the implemented parameters, “evaluate_nn” stands out as a particularly useful feature. When enabled, it allows for a preliminary evaluation of the neural network (TabularNet) before executing full training. This can significantly reduce computational cost and save time by identifying underperforming configurations early in the process.
The “top_k” parameter is another critical element, determining the number of top-ranked features selected for non-linear transformation during the symbolic regression phase. This parameter plays a direct role in shaping both model accuracy and computational efficiency.
Additionally, the “verbose” flag improves usability by integrating with the “tqdm” library to provide real-time progress updates during training. This enhances user experience by offering transparent feedback throughout both neural and symbolic learning stages.
The system requirements for installing and running the OIKAN framework are summarized in Table 6. These specifications ensure compatibility across platforms and provide guidelines for memory, storage, and optional GPU usage to support scalable training performance.

3.4. Datasets and Experimental Setup

To benchmark the OIKAN framework, datasets from the Penn Machine Learning Benchmarks (PMLB) [58] were employed, covering 60 classification and 56 regression tasks. These datasets varied widely in size, dimensionality, and complexity. Smaller datasets (e.g., 1096_FacultySalaries, parity5) included approximately 1000 augmented instances, while larger datasets (e.g., 344_mv, 215_2dplanes, 564_fried) contained up to 97,842 rows after augmentation. Gaussian noise-based data augmentation was applied with factors ranging from 2 to 240, with the total number of rows capped at 100,000 to ensure computational feasibility.
For evaluation, classification tasks were assessed using accuracy, precision, and F1-score, while regression tasks were measured by Root Mean Square Error (RMSE), Mean Absolute Percentage Error (MAPE), and R2 score. To complement predictive performance, we also recorded computational efficiency metrics, including training time, inference time, and memory usage.
To provide a comparative baseline, two reference models were selected: DecisionTree, representing classical interpretable methods, and XGBoost, representing high-performing black-box models. This setup allowed a direct analysis of the trade-off between interpretability and predictive accuracy across standardized PMLB tasks.
The symbolic regression component of OIKAN employed ElasticNet regularization, with default parameters l1_ratio = 0.5, alpha = 1, and random_state = 42 to ensure reproducibility.
In addition to the PMLBs, two widely used datasets were included for validation: the Iris dataset for classification (150 samples, 4 features, 3 species) and the California Housing dataset for regression (20,640 samples, 8 features, target = median house value). Both datasets were partitioned into 80% training and 20% testing subsets. The models were configured with two hidden layers of 32 neurons, ReLU activation, a data augmentation factor of 10, and batch training. Furthermore, the evaluate_nn parameter was utilized to pre-assess the neural component before the symbolic regression phase.
All experiments were conducted in a GPU-accelerated environment (Tesla P100, CUDA) on the Kaggle platform.

4. Results

4.1. Performance of OIKAN Classifier on Benchmark Datasets

To evaluate the classification performance and computational efficiency of the proposed framework, OIKANClassifier was tested on over 50 datasets from the PMLB benchmark suite. The results provide insight into the model’s ability to generalize across diverse classification tasks of varying size and complexity.
Table 7 presents the classification performance metrics, including accuracy, precision, and F1-score, for each dataset. The model achieved perfect accuracy (1.000) on six well-known datasets such as “iris”, “mushroom”, “analcatdata_creditscore”, “corral”, and “prnn_crabs”, demonstrating strong performance on symbolically or linearly separable problems.
A summary of the model’s aggregated classification performance includes:
  • Median accuracy: 0.886
  • Average accuracy on top 10 datasets: 0.987
  • Datasets with accuracy ≥ 0.95: 17 out of 50+ (34%)
  • Datasets with accuracy ≤ 0.50: 10 out of 50+ (20%)
In addition to classification performance, runtime and memory usage were analyzed to assess the model’s suitability for deployment. Table 8 summarizes the time and memory complexity metrics across the same set of PMLB datasets. Median training time was 4.9 s, and median prediction time was 0.006 s. Memory usage remained low during inference (median: 0.072 MB), with training memory consumption reaching 35.4 MB on average. Inference times were consistently below 0.05 s across all datasets, supporting real-time and embedded system applicability.
Training duration ranged from under one second on smaller datasets (e.g., confidence, analcatdata_fraud) to approximately 1.7 h on high-dimensional datasets such as mfeat_factors. Memory usage scaled with dataset size and dimensionality, peaking at approximately 24 GB for clean2.

4.2. Performance of OIKAN Regressor Across Symbolic and Noisy Tasks

The performance of OIKAN Regressor was assessed using over 50 regression datasets from the PMLB benchmark collection, with all experiments conducted on a GPU P100. Results are detailed in Table 9 and Table 10.
The model demonstrated strong performance on datasets characterized by symbolic structure or smooth analytical dependencies (Table 9). In particular, high R2 scores exceeding 0.9 were observed on datasets such as 2dplanes, cloud, and mv, accompanied by correspondingly low RMSE values. The median R2 score across all datasets was approximately 0.705, with 20% of the datasets (10 out of 50+) achieving R2 values of 0.85 or higher.
The top five performing datasets in terms of R2 were:
  • 344_mv: R2 = 0.992, RMSE = 0.950
  • 523_analcatdata_neavote: R2 = 0.938, RMSE = 0.885
  • 215_2dplanes: R2 = 0.937, RMSE = 1.098
  • 210_cloud: R2 = 0.920, RMSE = 0.353
  • 624_fri_c0_100_5: R2 = 0.875, RMSE = 0.327
However, performance significantly declined on noisy or non-symbolic datasets such as parity-like, pollution, and cpu_act. For example:
  • 227_cpu_small: RMSE = 189.2, R2 undefined (possibly due to numerical instability)
  • 218_house_8L: RMSE = 46,944.5, R2 = 0.215
  • 573_cpu_act: RMSE = 164.1, R2 undefined
Overall, 36% of the datasets had R2 scores of 0.50 or lower, reflecting difficulty in modeling data with high noise or poorly defined functional relationships.
The model’s error metrics (RMSE and MAPE) generally followed the trend of the R2 score. However, MAPE values were affected in cases involving large target magnitudes or near-zero predictions, indicating sensitivity to scale that may require normalization or alternative loss functions.
In terms of computational efficiency (Table 10), the OIKANRegressor maintained low training and inference costs:
  • Median training time: ~3.5 s
  • Median training memory: ~14 MB
  • Median prediction time: <0.01 s
  • Median prediction memory: <0.05 MB
Even for high-dimensional datasets such as 590_fri_c0_1000_50 (50 features), training completed in 7.39 s using 357 MB of memory. On a larger dataset like 201_pol (48 features, ~100 K rows), training required ~4 min and ~3.9 GB of memory.
The quantitative evaluation of interpretability was conducted by measuring the symbolic compactness of the extracted formulas. On average, classification formulas reached 9852 characters in length (ranging from 500 to 110,665), while regression formulas were notably shorter, with an average of 1043 characters (ranging from 235 to 5206). The relatively large sizes observed in classification tasks are explained by the need to construct a separate regression formula for each class, and in both regression and classification settings, datasets with many features and complex interaction patterns tend to generate substantially longer symbolic expressions. These lengths are also strongly influenced by the choice of hyperparameters: under default settings, formula sizes can vary considerably. In particular, if the top_k parameter, which controls the number of retained terms during function minimization, is left unconstrained, the resulting expressions may become unnecessarily long and less interpretable. Careful tuning of this parameter for each dataset can substantially reduce expression length and improve readability. This observation highlights an important direction for future work, developing adaptive mechanisms for hyperparameter selection that can balance predictive accuracy with symbolic compactness, thereby enhancing both the efficiency and interpretability of the framework.

4.3. Comparative Analysis with Baseline Models

To further analyze the variability of OIKAN’s classification outcomes, the benchmark results were categorized into performance tiers. As shown in Table 11, OIKAN achieved perfect accuracy on six datasets. This distribution illustrates both the model’s strengths on symbolically or linearly separable tasks and its limitations on noisier or higher-dimensional problems.
The comparative evaluation across the PMLBs highlights clear differences in performance between OIKAN and established baseline models. All results for OIKAN, ElasticNet, XGBoost, and DecisionTree are documented in the Supplementary Materials, including detailed CSV files for classification and regression benchmarks, as well as a comprehensive analytics notebook. In the classification setting, OIKAN achieved a median accuracy of 0.886, with corresponding precision and F1 scores of 0.91 and 0.89, respectively (Table 12). While the model reached perfect scores on several datasets and demonstrated competitive results in the top-10 benchmarks, for instance, an accuracy of 0.975 with equally high precision and F1 values, its overall performance distribution was broader, with median values trailing behind the baselines. ElasticNet and DecisionTree both achieved accuracies of around 0.94, with strong precision and F1 values, while XGBoost consistently outperformed all models, approaching 0.97 across metrics.
In regression tasks, OIKAN reached a median R2 of 0.705, a mean absolute percentage error of 3.33, and a root mean squared error of 22.77 (Table 13). Although it attained R2 values above 0.8 on multiple datasets and peaked at 0.82, its variability across datasets was pronounced, particularly in error measures where extreme outliers appeared. ElasticNet provided slightly stronger regression performance with an R2 of 0.64 and lower error rates, whereas DecisionTree occupied a middle ground with an R2 of 0.80. XGBoost again proved the most effective, delivering an R2 of 0.92 alongside consistently low error values.

4.4. Interpretability Assessment

Beyond predictive performance, experiments conducted on the PMLB benchmark datasets revealed that the complexity of OIKAN grows proportionally with the dimensionality of the data. As the number of features increases, the model faces increasing difficulty in extracting compact symbolic formulas. This challenge is further amplified by the architecture’s dual ElasticNet application and the use of data augmentation strategies. Augmentation increases the number of samples, while polynomial and interaction terms expand the feature space, causing the computational complexity to grow rapidly with the number of features. Although additional symbolic functions, such as trigonometric or logarithmic bases, are also included, their contribution to the overall complexity is negligible compared to the quadratic growth caused by feature expansion. To address system limitations and ensure stable execution on the Kaggle platform, datasets exceeding 600 MB were excluded from the benchmarking process; this threshold was experimentally determined to prevent CPU and RAM overload. The experiments further demonstrated that the number of features had a stronger impact on model complexity, and consequently on runtime and memory usage, than the number of samples, due to the quadratic asymptotic behavior. Additionally, the final architecture exhibited poor performance on several small datasets, which, combined with the suboptimal performance of state-of-the-art models on the same datasets, suggests that noise and a lack of clear internal structure limited the ability to detect predictive patterns.
According to the results of benchmark scores on PMLB datasets, classification models showed good performance, while regression models have difficulties generalizing and predicting unseen values. To demonstrate OIKAN’s performance on a real case, an example of credit risk assessment in the financial industry was taken, where the classification of a client’s credit ability gives permission or refusal to a client to obtain a loan. In this case, it is crucial to ensure a high level of interpretability and explainability of the model to justify the acceptance or rejection of a client’s loan application. To evaluate the performance and interpretability of OIKAN under various conditions, three experiments were conducted using different sets of parameters. The quantitative comparison of OIKAN and XGBoost on the credit score dataset is presented in Table 14 below.
In terms of credit rating classification (multi-classification: “good,” “poor,” “standard”), OIKAN performed better without data augmentation than with augmentation (Figure 5). This is because the accuracy of the neural network’s predictions was low, and instead of generalizing the training data, TabularNet reduced the quality of the data by increasing the number of misclassified results, which worsened the performance of the final model. Therefore, it is worth noting that data augmentation does not always help and can even worsen results. The performance of the OIKAN model (0.64 accuracy & 0.63 f1-score) without data augmentation showed satisfactory results, comparatively worse than one of the best classification models, XGBoost (0.75 accuracy & 0.75 f1-score). At the same time, OIKAN showed higher interpretability, providing formulas (in various formats: “standard”, “sympy”, “LaTeX”) that can be used to obtain current values, as well as analyzing formulas to provide data on the significance of features on the result. The same formulas cannot be obtained by XGBoost, because it requires a large tree ensemble that very challenging to reproduce as a formula. In terms of training time, OIKAN without data augmentation shows the best time of 2.8 s which was smaller than the same experiment with XGBoost.
Taken together, these findings suggest that OIKAN is capable of producing highly accurate or even perfect results on select datasets but exhibits greater variability and generally lower median performance than the baselines. This reflects the trade-off between predictive accuracy and interpretability: while ElasticNet, DecisionTree, and XGBoost achieve stronger and more stable results, OIKAN uniquely provides symbolic formulas that make its predictions transparent and explainable. This interpretability positions OIKAN as a valuable tool for research contexts where analytical insight and transparency are prioritized alongside accuracy. Additional details of the benchmarking setup and the updated results are provided in Appendix A.

5. Discussion and Conclusions

OIKAN is a novel neuro-symbolic framework that seamlessly integrates deep learning with symbolic inference to produce interpretable models for both classification and regression: a KAN-augmented MLP first learns robust feature representations (with optional Gaussian noise augmentation), then a two-stage sparse symbolic regression, ElasticNet selection, followed by non-linear basis fitting, distills them into concise, human-readable formulas. Fully Scikit-learn-compatible and exportable to JSON, it balances neural-network accuracy with symbolic clarity for real-time and embedded use.
The framework was rigorously evaluated across format enable seamless integration and cross-platform deployment, supporting implementation in languages such as Python, C++, and Rust.a comprehensive set of 60 classification and 58 regression tasks from the Penn Machine Learning Benchmarks (PMLB). The classification component, OIKANClassifier, demonstrated excellent generalization capabilities, achieving a median accuracy of 0.886 and perfect performance (accuracy = 1.000) on symbolically or linearly separable datasets such as iris, mushroom, and corral. In regression tasks, the OIKANRegressor achieved a median R2 score of 0.705, with peak performance on datasets featuring analytical structure, such as 344_mv (R2 = 0.992) and 2dplanes (R2 = 0.937). These results highlight OIKAN’s ability to extract symbolic representations of data with high fidelity, making it especially suitable for scientific discovery tasks.
OIKAN also proved to be computationally efficient and scalable. The median training time for classification tasks was approximately 4.9 s, and 3.5 s for regression. Prediction times were consistently low (median: <0.01 s), and memory usage during inference remained minimal (median: <0.072 MB), demonstrating the framework’s suitability for real-time and embedded applications.
OIKAN demonstrates unique advantages in interpretability and transparency. Unlike black-box approaches such as boosted trees or neural ensembles, OIKAN generates sparse mathematical expressions in closed form, enabling direct validation, testing, and transfer between programming environments. This feature makes OIKAN particularly suitable for domains where explainability and reproducibility are critical. In addition, its lightweight training and inference footprints make it well-suited for deployment in real-time systems and embedded devices, while its capacity to produce human-readable formulas facilitates debugging, regulatory compliance, and the generation of scientific insights.
Although the DecisionTree model also offers interpretability and reusability, it produces tree-based conditional structures rather than explicit mathematical formulas. This key difference positions OIKAN as a framework that prioritizes understanding and symbolic clarity, even at the cost of reduced predictive performance, particularly in datasets lacking a compact symbolic structure. These findings reaffirm the well-established trade-off between accuracy and interpretability, suggesting that OIKAN should be viewed not as a competitor to high-accuracy black-box models but as a complementary approach for tasks where transparency and symbolic reasoning are paramount.
Building on these results, OIKAN also exhibited notable limitations on datasets characterized by high dimensionality, noise, or non-symbolic relationships (e.g., mfeat_factors, parity5, cpu_act), where performance dropped significantly. These observations highlight the importance of enhancing the model’s robustness, for example, through adaptive regularization, an expanded library of base functions, and refined augmentation strategies to better handle noisy or unstructured data.
Performance can degrade on very large datasets with more than 50 K samples due to heavier basis selection and expression fitting; in high-dimensional spaces exceeding 100 features, sparsity tends to decrease and expressions grow longer, hindering readability. On predominantly linear relationships, ElasticNet often matches or exceeds accuracy with simpler linear forms. Sensitivity to feature and label noise can inflate the learned basis and reduce fidelity, and severe class imbalance can diminish classification quality without dedicated mitigation.
Taken together, these findings emphasize both the strengths and limitations of the current OIKAN implementation. On one hand, the framework provides interpretable symbolic models and flexible configuration parameters, such as evaluate_nn and augmentation_factor, which support targeted experimentation and reproducibility. On the other hand, its sensitivity to feature dimensionality and data noise points to the need for greater scalability and stability.
Looking ahead, future enhancements will focus on addressing these limitations through architectural optimization, integration of ensemble and attention-based symbolic components, and deeper exploration of symbolic feature selection methods. Furthermore, improving the framework’s scalability and robustness on large, high-dimensional datasets will be critical for broader applicability. Moreover, future research directions include evaluating extended symbolic basis sets and implementing more advanced hyperparameter optimization strategies to further enhance both performance and interpretability across diverse datasets. Overall, OIKAN represents a significant advancement in interpretable machine learning. By merging symbolic reasoning with neural representations, it delivers a practical, flexible, and interpretable solution for tasks requiring transparency and analytical insight, particularly in fields such as scientific modeling, healthcare diagnostics, and engineering design.

Supplementary Materials

The supplementary materials for this article are available online: https://github.com/silvermete0r/oikan (accessed on 1 October 2025) (contains the full implementation of the OIKAN framework, including model definition files, training scripts, utilities, and documentation); https://pepy.tech/projects/oikan (accessed on 1 October 2025) (provides download statistics, package version history, and links to installation instructions for the OIKAN package); https://www.kaggle.com/code/armanzhalgasbayev/oikan-v0-0-3-get-started-template-notebook (accessed on 1 October 2025) (demonstrates hands-on usage of the OIKAN v0.0.3 package: loading data, training models, evaluating performance, and visualizing results).

Author Contributions

Conceptualization, D.Y., N.K. and S.S.; methodology, A.Z. and S.S.; software, A.Z.; validation, D.Y., N.K., A.Z. and S.S.; formal analysis, A.Z. and S.S.; investigation, D.Y., N.K., A.Z. and S.S.; resources, N.K. and A.Z.; data curation, A.Z.; writing—original draft preparation, A.Z. and S.S.; writing—review and editing, D.Y., N.K., A.Z. and S.S.; visualization, A.Z. and S.S.; supervision, S.S.; project administration, D.Y. and N.K.; funding acquisition, D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been funded by the Committee of Science of the Ministry of Science and Higher Education of the Republic of Kazakhstan (Grant No. BR24992852 “Intelligent models and methods of Smart City digital ecosystem for sustainable development and the citizens’ quality of life improvement”).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The following links provide implementation details, documentation, baseline models, and benchmark comparisons for the OIKAN framework.
Table A1. Useful links related to OIKAN implementation and benchmarking.
Table A1. Useful links related to OIKAN implementation and benchmarking.
CategoryDescriptionLink
Source CodeOfficial GitHub repository with full implementation and exampleshttps://github.com/silvermete0r/oikan (accessed on 1 September 2025)
DocumentationOfficial library documentation page with usage and tutorialshttps://silvermete0r.github.io/oikan/ (accessed on 1 September 2025)
Baseline Model
(Interpretable)
DecisionTree model
from scikit-learn
https://scikit-learn.org/stable/modules/tree.html (accessed on 1 September 2025)
Baseline Model (Black-box)XGBoost documentationhttps://xgboost.readthedocs.io (accessed on 1 September 2025)
Benchmark ResultsUpdated PMLB benchmarking notebook comparing OIKAN, XGBoost, and DecisionTreehttps://www.kaggle.com/code/armanzhalgasbayev/oikan-v0-0-3-auto-benchmarking-sr/ (accessed on 1 September 2025)
Credit Score
Classification
Credit score classification using a public credit-score dataset.https://www.kaggle.com/code/armanzhalgasbayev/oikan-ml-credit-score-classification (accessed on 1 October 2025)

References

  1. Kautz, H. The third AI summer: AAAI Robert S. Engelmore memorial lecture. AI Mag. 2022, 43, 105–125. [Google Scholar]
  2. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
  3. Cranmer, M.; Sanchez Gonzalez, A.; Battaglia, P.; Xu, R.; Cranmer, K.; Spergel, D.; Ho, S. Discovering symbolic models from deep learning with inductive biases. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; Curran Associates Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 17429–17442. [Google Scholar]
  4. Iten, R.; Metger, T.; Wilming, H.; Del Rio, L.; Renner, R. Discovering physical concepts with neural networks. Phys. Rev. Lett. 2020, 124, 010508. [Google Scholar] [CrossRef]
  5. Makke, N.; Chawla, S. Interpretable scientific discovery with symbolic regression: A review. Artif. Intell. Rev. 2024, 57, 2. [Google Scholar] [CrossRef]
  6. La Cava, W.; Singh, T.R.; Taggart, J.; Suri, S.; Moore, J.H. Learning concise representations for regression by evolving networks of trees. arXiv 2018, arXiv:1807.00981. [Google Scholar]
  7. Udrescu, S.M.; Tan, A.; Feng, J.; Neto, O.; Wu, T.; Tegmark, M. AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; Curran Associates Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 4860–4871. [Google Scholar]
  8. Johnston, W.J.; Fusi, S. Abstract representations emerge naturally in neural networks trained to perform multiple tasks. Nat. Commun. 2023, 14, 1040. [Google Scholar] [CrossRef]
  9. Elmoznino, E.; Bonner, M.F. High-performing neural network models of visual cortex benefit from high latent dimensionality. PLoS Comput. Biol. 2024, 20, e1011792. [Google Scholar] [CrossRef] [PubMed]
  10. Hornik, K. Approximation capabilities of multilayer feedforward networks. Neural Netw. 1991, 4, 251–257. [Google Scholar] [CrossRef]
  11. Kolmogorov, A.N. On the representations of continuous functions of many variables by superposition of continuous functions of one variable and addition. Dokl. Akad. Nauk USSR 1957, 114, 953–956. [Google Scholar]
  12. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Tegmark, M. Kan: Kolmogorov-Arnold networks. arXiv 2024, arXiv:2404.19756. [Google Scholar]
  13. Cherednichenko, O.; Poptsova, M. Kolmogorov–Arnold networks for genomic tasks. Brief. Bioinform. 2025, 26, bbaf129. [Google Scholar] [CrossRef]
  14. La Cava, W.G.; Lee, P.C.; Ajmal, I.; Ding, X.; Solanki, P.; Cohen, J.B.; Herman, D.S. A flexible symbolic regression method for constructing interpretable clinical prediction models. NPJ Digit. Med. 2023, 6, 107. [Google Scholar] [CrossRef]
  15. Liu, H.; Wang, Y.; Fan, W.; Liu, X.; Li, Y.; Jain, S.; Liu, Y.; Jain, A.; Tang, J. Trustworthy AI: A computational perspective. ACM Trans. Intell. Syst. Technol. 2022, 14, 1–59. [Google Scholar] [CrossRef]
  16. Liu, Y.; Yao, Y.; Ton, J.-F.; Zhang, X.; Guo, R.; Cheng, H.; Klochkov, Y.; Taufiq, M.F.; Li, H. Trustworthy LLMs: A survey and guideline for evaluating large language models’ alignment. arXiv 2023, arXiv:2308.05374. [Google Scholar]
  17. Yu, R.; Yu, W.; Wang, X. Kan or MLP: A fairer comparison. arXiv 2024, arXiv:2407.16674. [Google Scholar] [CrossRef]
  18. Shukla, K.; Toscano, J.D.; Wang, Z.; Zou, Z.; Karniadakis, G.E. A comprehensive and fair comparison between MLP and KAN representations for differential equations and operator networks. Comput. Methods Appl. Mech. Eng. 2024, 431, 117290. [Google Scholar] [CrossRef]
  19. Chen, Z.; Zhang, X. LSS-SKAN: Efficient Kolmogorov-Arnold Networks based on single-parameterized function. arXiv 2024, arXiv:2410.14951. [Google Scholar]
  20. Mintisan. Awesome-KAN: A Curated List of Resources Related to Kolmogorov–Arnold Networks (KANs). Available online: https://github.com/mintisan/awesome-kan (accessed on 19 June 2025).
  21. Ji, T.; Hou, Y.; Zhang, D. A comprehensive survey on Kolmogorov Arnold Networks (KAN). arXiv 2024, arXiv:2407.11075. [Google Scholar] [CrossRef]
  22. Liu, J. Kolmogorov-Arnold networks for symbolic regression and time series prediction. J. Mach. Learn. Res. 2024, 25, 95–110. [Google Scholar]
  23. Xu, L. Time-Kolmogorov-Arnold Networks and multi-task Kolmogorov-Arnold Networks for time series prediction. J. Time Ser. Anal. 2024, 45, 200–220. [Google Scholar]
  24. Wang, Y.; Sun, J.; Bai, J.; Anitescu, C.; Eshaghi, M.S.; Zhuang, X.; Liu, Y. Kolmogorov Arnold Informed Neural Network: A physics-informed deep learning framework for solving forward and inverse problems based on Kolmogorov Arnold Networks. arXiv 2024, arXiv:2406.11045. [Google Scholar] [CrossRef]
  25. Zhalgasbayev, A.; Khaimuldin, N. Optimized Interpretable Kolmogorov-Arnold Networks (OIKAN) for Neuro-Symbolic Machine Learning. In Proceedings of the 2nd International Students Conference “Digital Generation-2025”; Astana IT University: Astana, Kazakhstan, 2025; pp. 638–646. ISBN 978-601-7911-72-0. [Google Scholar]
  26. Harmon, I.; Weinstein, B.; Bohlman, S.; White, E.; Wang, D.Z. A neuro-symbolic framework for tree crown delineation and tree species classification. Remote Sens. 2024, 16, 4365. [Google Scholar] [CrossRef]
  27. Hoehndorf, R.; Pesquita, C.; Zhapa-Camacho, F. Neuro-symbolic AI in life sciences. In Handbook on Neurosymbolic AI and Knowledge Graphs; IOS Press: Amsterdam, The Netherlands, 2025; pp. 924–951. [Google Scholar]
  28. Schmidt, M.; Lipson, H. Distilling free-form natural laws from experimental data. Science 2009, 324, 81–85. [Google Scholar] [CrossRef] [PubMed]
  29. Jin, Y.; Fu, W.; Kang, J.; Guo, J.; Guo, J. Bayesian symbolic regression. arXiv 2019, arXiv:1910.08892. [Google Scholar]
  30. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar] [CrossRef]
  31. Sette, S.; Boullart, L. Genetic programming: Principles and applications. Eng. Appl. Artif. Intell. 2001, 14, 727–736. [Google Scholar] [CrossRef]
  32. Björck, Å. Least squares methods. In Handbook of Numerical Analysis; Elsevier: Amsterdam, The Netherlands, 1990; Volume 1, pp. 465–652. [Google Scholar]
  33. Bertsimas, D.; Pauphilet, J.; Van Parys, B. Sparse regression. Stat. Sci. 2020, 35, 555–578. [Google Scholar]
  34. Udrescu, S.M.; Tegmark, M. AI Feynman: A physics-inspired method for symbolic regression. Sci. Adv. 2020, 6, eaay2631. [Google Scholar] [CrossRef]
  35. Yi, K.; Wu, J.; Gan, C.; Torralba, A.; Kohli, P.; Tenenbaum, J. Neural-symbolic VQA: Disentangling reasoning from vision and language understanding. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018; Volume 31, pp. 1039–1050. [Google Scholar]
  36. Vsevolodovna, R.I.M.; Monti, M. Enhancing Large Language Models through Neuro-Symbolic Integration and Ontological Reasoning. arXiv 2025, arXiv:2504.07640. [Google Scholar] [CrossRef]
  37. He, Y.; Xie, Y.; Yuan, Z.; Sun, L. MLP-KAN: Unifying Deep Representation and Function Learning. arXiv 2024, arXiv:2410.03027. [Google Scholar] [CrossRef]
  38. Garcez, A.D.A.; Gori, M.; Lamb, L.C.; Serafini, L.; Spranger, M.; Tran, S.N. Neural-symbolic computing: An effective methodology for principled integration of machine learning and reasoning. arXiv 2019, arXiv:1905.06088. [Google Scholar] [CrossRef]
  39. Ellis, K.; Wong, L.; Nye, M.; Sable-Meyer, M.; Cary, L.; Anaya Pozo, L.; Tenenbaum, J.B. DreamCoder: Growing generalizable, interpretable knowledge with wake–sleep Bayesian program learning. Philos. Trans. R. Soc. A 2023, 381, 20220050. [Google Scholar] [CrossRef] [PubMed]
  40. Acharya, K.; Raza, W.; Dourado, C.; Velasquez, A.; Song, H.H. Neurosymbolic reinforcement learning and planning: A survey. IEEE Trans. Artif. Intell. 2023, 5, 1939–1953. [Google Scholar] [CrossRef]
  41. Colelough, B.C.; Regli, W. Neuro-symbolic AI in 2024: A systematic review. arXiv 2025, arXiv:2501.05435. [Google Scholar] [CrossRef]
  42. Nawaz, U.; Anees-ur-Rahaman, M.; Saeed, Z. A review of neuro-symbolic AI integrating reasoning and learning for advanced cognitive systems. Intell. Syst. Appl. 2025, 26, 200541. [Google Scholar] [CrossRef]
  43. Ennab, M.; Mcheick, H. Designing an interpretability-based model to explain the artificial intelligence algorithms in healthcare. Diagnostics 2022, 12, 1557. [Google Scholar] [CrossRef]
  44. de Franca, F.O.; Virgolin, M.; Kommenda, M.; Majumder, M.S.; Cranmer, M.; Espada, G.; La Cava, W.G. Interpretable Symbolic Regression for Data Science: Analysis of the 2022 Competition. arXiv 2023, arXiv:2304.01117. [Google Scholar] [CrossRef]
  45. Cranmer, M. Interpretable machine learning for science with PySR and SymbolicRegression. arXiv 2023, arXiv:2305.01582. [Google Scholar]
  46. La Cava, W.; Burlacu, B.; Virgolin, M.; Kommenda, M.; Orzechowski, P.; de França, F.O.; Moore, J.H. Contemporary symbolic regression methods and their relative performance. Adv. Neural Inf. Process. Syst. 2021, 2021, 1. [Google Scholar] [PubMed]
  47. gplearn: Genetic Programming in Python. Available online: https://github.com/trevorstephens/gplearn (accessed on 19 June 2025).
  48. Kim, S.; Lu, P.Y.; Mukherjee, S.; Gilbert, M.; Jing, L.; Čeperić, V.; Soljačić, M. Integration of neural network-based symbolic regression in deep learning for scientific discovery. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4166–4177. [Google Scholar] [CrossRef] [PubMed]
  49. pykan: Kolmogorov–Arnold Networks in Python. Available online: https://github.com/KindXiaoming/pykan (accessed on 19 June 2025).
  50. Kulkarni, T.D.; Saeedi, A.; Gautam, S.; Gershman, S.J. Deep successor reinforcement learning. arXiv 2016, arXiv:1606.02396. [Google Scholar] [CrossRef]
  51. Broløs, K.R.; Machado, M.V.; Cave, C.; Kasak, J.; Stentoft-Hansen, V.; Batanero, V.G.; Wilstrup, C. An approach to symbolic regression using Feyn. arXiv 2021, arXiv:2104.05417. [Google Scholar] [CrossRef]
  52. SRTransformer: Symbolic Regression with Transformers. Available online: https://github.com/yinghanlong/SRtransformer (accessed on 19 June 2025).
  53. Eureqa (DataRobot Documentation). Available online: https://docs.datarobot.com/en/docs/modeling/analyze-models/describe/eureqa.html (accessed on 19 June 2025).
  54. Schmidt-Hieber, J. The Kolmogorov–Arnold representation theorem revisited. Neural Netw. 2021, 137, 119–126. [Google Scholar] [CrossRef]
  55. Bebis, G.; Georgiopoulos, M. Feed-forward neural networks. IEEE Potentials 1994, 13, 27–31. [Google Scholar] [CrossRef]
  56. McHutchon, A.; Rasmussen, C. Gaussian process training with input noise. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Granada, Spain, 12–15 December 2011; Curran Associates Inc.: Red Hook, NY, USA, 2011; Volume 24. [Google Scholar]
  57. Tay, J.K.; Narasimhan, B.; Hastie, T. Elastic net regularization paths for all generalized linear models. J. Stat. Softw. 2023, 106, 1–31. [Google Scholar] [CrossRef] [PubMed]
  58. Olson, R.S.; La Cava, W.; Orzechowski, P.; Urbanowicz, R.J.; Moore, J.H. PMLB: A large benchmark suite for machine learning evaluation and comparison. BioData Min. 2017, 10, 1–13. [Google Scholar] [CrossRef]
Figure 1. Installation and Usage Workflow for OIKAN Framework.
Figure 1. Installation and Usage Workflow for OIKAN Framework.
Algorithms 18 00639 g001
Figure 2. OIKAN Class Structure and Component Relationships.
Figure 2. OIKAN Class Structure and Component Relationships.
Algorithms 18 00639 g002
Figure 3. High-level Architecture Diagram of the OIKAN.
Figure 3. High-level Architecture Diagram of the OIKAN.
Algorithms 18 00639 g003
Figure 4. End-to-End Diagram of the OIKAN.
Figure 4. End-to-End Diagram of the OIKAN.
Algorithms 18 00639 g004
Figure 5. OIKAN-Derived Feature Importance on Credit-Rating (Good/Poor/Standard).
Figure 5. OIKAN-Derived Feature Importance on Credit-Rating (Good/Poor/Standard).
Algorithms 18 00639 g005
Table 4. Main workflow stages in the OIKAN framework.
Table 4. Main workflow stages in the OIKAN framework.
Step Description Purpose
Data Preprocessing and Augmentation A feedforward NN (TabularNet) analyzes input data to capture initial patterns. Gaussian noise is added to the data, creating multiple perturbed versions to augment the dataset. The NN then generates predictions or logits for the augmented data.Enhances model robustness by introducing controlled variability, allowing the model to learn generalized patterns. The NN provides a preliminary mapping of data relationships, which is refined in later steps.
Two-Stage
Symbolic Regression
Stage 1 (Coarse Model): Polynomial features (degree 2) are generated from the augmented data. An ElasticNet model, combining L1 (LASSO) and L2 (Ridge) regularization, is fitted to compute feature importances based on coefficients. The top-k features are selected.
Stage 2 (Refined Model): Non-linear basis functions are added for the top-k features. A second ElasticNet model is fitted to the combined polynomial and non-linear features.
Constructs an interpretable model by first identifying the most relevant features (Stage 1) and then enhancing the model with non-linear transformations (Stage 2). ElasticNet ensures sparsity and robustness, balancing model complexity and accuracy.
Model CompilationThe coefficients and basic functions from the refined model are used to construct a symbolic mathematical expression, representing the relationship between features and the target. The expression is simplified using SymPy and supports formats like original, SymPy, or LATEX.Produces a human-readable, interpretable formula that encapsulates the learned relationships, enabling efficient predictions and easy inspection of the model’s logic.
Model Export
and Use
The symbolic model, including basis functions, coefficients, and feature metadata, is saved in JSON format. The model can be loaded and used in multiple languages (C, C++, JavaScript, Rust, Go, Python). Facilitates model portability and deployment across diverse platforms, ensuring accessibility and usability for different applications while maintaining interpretability and efficiency.
Table 5. Definitions of the OIKAN base class parameters.
Table 5. Definitions of the OIKAN base class parameters.
ParameterStageDescriptionValues
hidden_sizesNN TrainingList of hidden layer sizes for the MLP. List of integers, default = [64, 64]
activationNN TrainingActivation function for the neural network layers.Options: ‘relu’, ‘tanh’, ‘leaky_relu’, ‘elu’, ‘swish’, ‘gelu’, default=‘relu’
epochsNN TrainingNumber of epochs for training the neural network.Integer, default = 100
lrNN TrainingLearning rate for the optimizer. Float, default = 0.001
batch_sizeNN TrainingBatch size for mini-batch training. Integer, default = 32
evaluate_nnNN TrainingWhether to evaluate the NN before full training. Boolean, default = False
augmentation_factorData AugmentationNumber of augmented samples generated per original sample.Integer, default = 10
sigmaData AugmentationStandard deviation of Gaussian noise added during data augmentation.Float, default = 0.1
alphaSymbolic RegressionL1 regularization strength for Lasso in symbolic regression.Float, default = 0.1
top_kSymbolic RegressionNumber of top features to select in hierarchical symbolic regression.Integer, default = 5
verboseGeneralWhether to display training progress for neural network and symbolic regression.Boolean, default = False
random_stateGeneralRandom seed for reproducibility.Integer, default = None
Table 6. Minimum system requirements for working with OIKAN.
Table 6. Minimum system requirements for working with OIKAN.
RequirementDetails
PythonVersion 3.7 or higher
Operating SystemPlatform independent (Windows/macOS/Linux)
MemoryRecommended minimum 4 GB RAM
Disk Space~100 MB for installation (including dependencies)
GPUOptional (for faster training)
Dependenciestorch, numpy, scikit-learn, sympy, tqdm
Table 3. Evaluation Metrics for Comparative Assessment of OIKAN Performance in Classification and Regression Tasks.
Table 3. Evaluation Metrics for Comparative Assessment of OIKAN Performance in Classification and Regression Tasks.
MetricsFormulaTypeDescription
R2 Score (Coefficient of Determination) R 2 = 1 i = 1 n y i y i ^ 2 i = 1 n y i y ¯ 2 RegressionProportion of variance explained by the model. Higher (0 to 1) is better.
Root Mean Squared Error (RMSE) RMSE = 1 n i = 1 n y i y i ^ 2 RegressionMeasures average prediction error magnitude in target units. Lower is better.
Mean Absolute Percentage Error (MAPE) MAPE = 100 % n i = 1 n y i y i ^ y i RegressionAverage percentage error of predictions. Lower is better.
Accuracy7
Accuracy = Number   of   Correct   Predictions Total   Number   of   Predictions = T P + T N T P + T N + F P + F N
ClassificationProportion of correct predictions. Higher is better, but less reliable for imbalanced data.
Precision (Weighted) Precision weighted = k = 1 K n k n T P k T P k + F P k ClassificationWeighted proportion of positive predictions that are correct. Higher is better.
F1-Score (Weighted) F 1 weighted = k = 1 K n k n 2 Precision k Recall k Precision k + Recall k ClassificationWeighted harmonic mean of precision and recall. Higher is better for balanced performance.
Table 7. Evaluation Results of the OIKAN Classifier on PMLB Benchmark Datasets.
Table 7. Evaluation Results of the OIKAN Classifier on PMLB Benchmark Datasets.
Dataset#Feat. *RowsAccuracyPrecisionF1
iris412001.0001.0001.000
analcatdata_creditscore610401.0001.0001.000
corral612801.0001.0001.000
prnn_crabs716001.0001.0001.000
mushroom2264,9901.0001.0001.000
analcatdata_authorship7067201.0001.0001.000
clean216852,7801.0001.0001.000
dermatology3429200.9860.9880.986
optdigits6444,9600.9850.9850.985
new_thyroid517200.9770.9780.976
dis2930,1700.9750.9610.968
mfeat_karhunen6416,0000.9730.9730.972
mfeat_pixel24016,0000.9680.9680.967
hypothyroid2525,3000.9610.9570.959
monk3644300.9550.9560.955
balance_scale450000.9520.9550.944
clean116838000.9480.9480.948
kr_vs_kp3625,5600.9470.9470.947
page_blocks1043,7800.9440.9450.937
ionosphere3428000.9300.9320.929
analcatdata_lawsuit421100.9250.9140.916
coil20008578,5700.9090.8920.900
backache3214400.8890.9020.865
car_evaluation613,8200.8870.8910.884
prnn_synth220000.8800.8800.880
mfeat_zernike4716,0000.8600.8620.851
ecoli726100.8330.8110.791
lymphography1811800.8330.8510.831
tic_tac_toe976600.8230.8220.821
analcatdata_boxing2310500.8150.8130.813
confidence310260.8000.7780.775
biomed816700.7860.8350.735
lupus310350.7780.7810.769
phoneme543,2300.7650.7540.734
adult1478,1460.7600.7380.745
labor1610350.7500.7710.756
hepatitis1912400.7420.7420.742
led7725,6000.7340.7430.730
appendicitis710080.7270.5290.612
hayes_roth412800.7190.7530.710
sonar6016600.7140.7160.714
mfeat_fourier7616,0000.6700.7310.633
saheart936900.6560.6880.555
monk1644400.6520.6500.649
haberman324400.6450.5920.609
monk2648000.6200.3930.481
collins2338800.5980.6050.557
ring2059,2000.5850.7780.511
analcatdata_fraud1110230.5560.3090.397
movement_libras9028800.5560.6870.588
penguins726600.5370.3270.395
analcatdata_boxing1310560.5000.5000.500
flags4314200.4720.4870.462
mfeat_morphological616,0000.4450.3310.357
bupa527600.4350.1890.264
parity5+51089900.4180.4210.405
mfeat_factors21616,0000.3850.5210.346
parity5510000.2860.0820.127
schizo1427200.2210.3380.190
analcatdata_dmft463700.1810.1950.169
* number of features.
Table 8. Computational Efficiency of OIKAN Classifier on PMLB Classification Datasets.
Table 8. Computational Efficiency of OIKAN Classifier on PMLB Classification Datasets.
Dataset#Feat.RowsTrain (s)Train (MB)Pred (s)Pred (MB)
iris412001.1471.7560.0010.007
analcatdata_creditscore610400.8592.2760.0010.006
corral612801.1132.7880.0020.012
prnn_crabs716001.3874.0040.0020.014
mushroom2264,990100.766646.6070.0292.147
analcatdata_authorship706720131.617567.5420.0622.041
clean216852,7802944.12024,386.3061.165149.450
dermatology34292012.37070.3260.0260.182
optdigits6444,9601042.3423157.3500.07813.036
new_thyroid517201.6503.2650.0010.012
dis2930,17055.173488.6900.0211.070
mfeat_karhunen6416,000426.6941129.1900.0715.215
mfeat_pixel24016,0004278.02314,991.7240.59140.449
hypothyroid2525,30041.107317.9960.0180.837
monk3644304.1739.5450.0010.026
balance_scale450004.8927.2760.0020.029
clean11683800277.4641765.3480.4207.985
kr_vs_kp3625,56048.079611.3870.0231.051
page_blocks1043,78054.460140.8370.0160.684
ionosphere3428004.91767.6440.0120.092
analcatdata_lawsuit421102.0033.0360.0010.012
coil20008578570595.1989532.0180.09412.436
backache3214402.65534.2190.0100.029
car_evaluation61382014.80729.9100.0080.145
prnn_synth220001.7581.3520.0010.008
mfeat_zernike4716,000293.332630.6030.0493.516
ecoli726102.5636.5650.0020.028
lymphography1811801.77910.3940.0040.023
tic_tac_toe976607.66224.9420.0060.072
analcatdata_boxing2310501.1791.1100.0010.004
confidence310260.6491.1240.0010.004
biomed816701.8354.7960.0020.023
lupus310350.8441.0970.0010.003
phoneme543,23038.54574.1350.0040.217
adult1478,146410.054409.5130.0499.926
labor1610350.8277.5420.0040.008
hepatitis1912401.59211.8940.0050.029
led7725,60027.38361.5260.0060.318
appendicitis710080.8992.6400.0030.007
hayes_roth412801.1631.8660.0010.008
sonar6016603.661110.9480.0080.042
mfeat_fourier7616,000141.3731570.2640.0351.673
saheart936903.80612.0370.0020.043
monk1644404.0129.5730.0010.031
haberman324402.1692.5940.0010.010
monk2648004.39610.3460.0010.027
collins23388021.46949.6280.0170.153
ring2059,20077.061503.5070.0080.451
analcatdata_fraud1110230.5844.3400.0010.002
movement_libras902880256.894400.0990.1011.753
penguins726602.9326.6510.0020.026
analcatdata_boxing1310560.8601.1160.0010.004
flags4314206.84154.9170.0140.055
mfeat_morphological61600018.85335.4060.0040.113
bupa527602.6695.2740.0010.018
parity5+51089907.88933.2890.0040.032
mfeat_factors21616,0005980.85912,166.9620.89959.125
parity5510000.2871.9000.0010.002
schizo1427204.06316.1030.0040.071
analcatdata_dmft463706.1229.2730.0020.039
Table 9. Performance Summary of the OIKAN Regressor on PMLB Regression Benchmark Datasets.
Table 9. Performance Summary of the OIKAN Regressor on PMLB Regression Benchmark Datasets.
Dataset#Feat.RowsRMSER2MAPE
344_mv 10 97,842 0.950 0.992 2.772
523_analcatdata_neavote 2 1040 0.885 0.938 0.446
215_2dplanes 10 97,842 1.098 0.937 1.509
210_cloud 5 1032 0.353 0.920 0.407
624_fri_c0_100_5 5 1040 0.327 0.875 0.453
649_fri_c0_500_5 5 4000 0.382 0.862 2.156
229_pwLinear 10 1600 1.386 0.862 0.418
595_fri_c0_1000_10 10 8000 0.386 0.855 0.958
230_machine_cpu 6 1670 87.062 0.851 0.732
1027_ESL 4 3900 0.507 0.836 0.082
590_fri_c0_1000_50 50 8000 0.399 0.825 1.034
609_fri_c0_1000_5 5 8000 0.438 0.823 1.495
1096_FacultySalaries 4 1000 1.936 0.797 0.036
561_cpu 7 1670 104.397 0.797 0.278
529_pollen 4 30,780 1.477 0.786 3.445
294_satellite_image 36 51,480 1.073 0.763 0.360
579_fri_c0_250_5 5 2000 0.439 0.757 2.838
635_fri_c0_250_10 10 2000 0.478 0.757 2.842
201_pol 48 96,000 22.270 0.708 4.76 × 1016
564_fried 10 97,842 2.693 0.705 0.229
621_fri_c0_100_10 10 1040 0.589 0.687 0.810
599_fri_c2_1000_5 5 8000 0.605 0.670 1.108
192_vineyard 2 1025 2.503 0.665 0.138
1089_USCrime 13 1036 20.204 0.663 0.090
589_fri_c2_1000_25 25 8000 0.635 0.623 0.952
612_fri_c1_1000_5 5 8000 0.600 0.615 1.584
623_fri_c4_1000_10 10 8000 0.618 0.609 1.610
4544_GeographicalOriginalofMusic 117 8470 0.580 0.597 4.051
620_fri_c1_1000_25 25 8000 0.642 0.592 1.125
592_fri_c4_1000_25 25 8000 0.613 0.570 1.322
586_fri_c3_1000_25 25 8000 0.691 0.569 1.448
588_fri_c4_1000_100 100 8000 0.678 0.555 1.224
583_fri_c1_1000_50 50 8000 0.657 0.530 2.462
616_fri_c4_500_50 50 4000 0.691 0.528 1.074
605_fri_c2_250_25 25 2000 0.686 0.498 1.127
581_fri_c3_500_25 25 4000 0.768 0.493 6.904
617_fri_c3_500_5 5 4000 0.808 0.417 1.925
225_puma8NH 8 65,530 4.265 0.413 2.799
1029_LEV 4 8000 0.718 0.396 2.96 × 1014
1030_ERA 4 8000 1.607 0.376 0.425
527_analcatdata_election2000 14 1007 96,238.090 0.344 0.708
1028_SWD 10 8000 0.672 0.307 0.145
547_no2 7 4000 0.627 0.301 0.142
218_house_8L 8 91,135 46,944.468 0.215 2.71 × 1017
591_fri_c1_100_10 10 1040 0.687 0.124 1.454
228_elusage 2 1012 26.929 - 0.589
557_analcatdata_apnea1 3 3800 3925.899 - 1.89 × 1018
556_analcatdata_apnea2 3 3800 4036.276 - 1.85 × 1018
485_analcatdata_vehicle 4 1026 373.732 - 0.821
522_pm10 7 4000 0.915 - 0.278
227_cpu_small 12 65,530 189.186 - 4.24 × 1015
562_cpu_small 12 65,530 48.787 - 2.24 × 1015
542_pollution 15 1008 267.912 - 0.262
574_house_16H 16 91,135 53,044.147 - 4.88 × 1017
197_cpu_act 21 65,530 22.030 - 1.19 × 1015
573_cpu_act 21 65,530 164.100 - 2.62 × 1015
Table 10. Time and Memory Complexity of the OIKAN Regressor on PMLB Regression Benchmarks.
Table 10. Time and Memory Complexity of the OIKAN Regressor on PMLB Regression Benchmarks.
Dataset#Feat.RowsTrain (s)Train (MB)Pred (s)Pred (MB)
344_mv 10 97,842 282.833 301.584 0.013 3.197
523_analcatdata_neavote 2 1040 0.806 0.715 0.000 0.002
215_2dplanes 10 97,842 266.519 301.576 0.008 1.827
210_cloud 5 1032 0.766 1.950 0.000 0.004
624_fri_c0_100_5 5 1040 0.750 1.963 0.001 0.004
649_fri_c0_500_5 5 4000 3.216 7.437 0.001 0.015
229_pwLinear 10 1600 1.341 5.944 0.002 0.018
595_fri_c0_1000_10 10 8000 6.397 29.555 0.001 0.028
230_machine_cpu 6 1670 1.547 3.613 0.001 0.015
1027_ESL 4 3900 3.340 5.539 0.001 0.018
590_fri_c0_1000_50 50 8000 7.390 357.336 0.001 0.035
609_fri_c0_1000_5 5 8000 6.389 14.828 0.001 0.031
1096_FacultySalaries 4 1000 0.539 1.446 0.001 0.003
561_cpu 7 1670 1.575 4.159 0.001 0.018
529_pollen 4 30,780 26.735 43.379 0.004 0.143
294_satellite_image 36 51,480 103.953 1222.647 0.063 6.982
579_fri_c0_250_5 5 2000 1.681 3.741 0.001 0.008
635_fri_c0_250_10 10 2000 1.749 7.422 0.001 0.010
201_pol 48 96,000 235.943 3885.826 0.084 16.249
564_fried 10 97,842 267.517 301.579 0.010 2.806
621_fri_c0_100_10 10 1040 0.748 3.885 0.001 0.006
599_fri_c2_1000_5 5 8000 6.481 14.826 0.001 0.030
192_vineyard 2 1025 0.504 0.699 0.001 0.002
1089_USCrime 13 1036 0.608 5.518 0.004 0.011
589_fri_c2_1000_25 25 8000 6.805 106.004 0.002 0.047
612_fri_c1_1000_5 5 8000 6.678 14.826 0.001 0.032
623_fri_c4_1000_10 10 8000 6.391 29.553 0.001 0.030
4544_GeographicalOriginalofMusic 117 8470 121.030 1926.686 0.022 0.279
620_fri_c1_1000_25 25 8000 6.669 106.002 0.001 0.035
592_fri_c4_1000_25 25 8000 7.189 106.000 0.002 0.043
586_fri_c3_1000_25 25 8000 6.689 106.006 0.001 0.033
588_fri_c4_1000_100 100 8000 11.997 1340.217 0.008 0.123
583_fri_c1_1000_50 50 8000 7.678 357.328 0.008 0.088
616_fri_c4_500_50 50 4000 4.011 182.768 0.009 0.081
605_fri_c2_250_25 25 2000 1.791 30.500 0.005 0.033
581_fri_c3_500_25 25 4000 3.504 57.044 0.003 0.031
617_fri_c3_500_5 5 4000 3.323 7.436 0.001 0.013
225_puma8NH 8 65,530 55.508 159.550 0.006 0.394
1029_LEV 4 8000 6.511 11.306 0.001 0.029
1030_ERA 4 8000 6.586 11.302 0.001 0.040
527_analcatdata_election2000 14 1007 0.711 5.975 0.004 0.015
1028_SWD 10 8000 7.321 29.554 0.002 0.049
547_no2 7 4000 3.689 9.905 0.001 0.032
218_house_8L 8 91,135 164.753 218.753 0.009 2.444
591_fri_c1_100_10 10 1040 0.741 3.883 0.001 0.006
228_elusage 2 1012 0.498 0.692 0.000 0.002
557_analcatdata_apnea1 3 3800 3.182 3.872 0.001 0.021
556_analcatdata_apnea2 3 3800 3.160 3.868 0.001 0.021
485_analcatdata_vehicle 4 1026 0.510 1.489 0.001 0.004
522_pm10 7 4000 3.650 9.903 0.001 0.028
227_cpu_small 12 65,530 63.710 258.117 0.018 0.998
562_cpu_small 12 65,530 62.503 258.115 0.005 0.867
542_pollution 15 1008 0.675 6.639 0.005 0.016
574_house_16H 16 91,135 176.405 539.557 0.021 6.381
197_cpu_act 21 65,530 77.496 602.564 0.029 2.597
573_cpu_act 21 65,530 78.804 602.562 0.020 2.781
Table 11. Performance Distribution of OIKAN Classifier Across PMLB Datasets.
Table 11. Performance Distribution of OIKAN Classifier Across PMLB Datasets.
Performance TierOIKAN CountNotable Datasets
Perfect (1.000)6mushroom, iris, analcatdata_authorship,
analcatdata_creditscore, corral, prnn_crabs
Excellent (0.95+)17monk3 (0.955), balance_scale (0.952),
dermatology (0.986)
Good (0.80–0.94)18tic_tac_toe (0.823), car_evaluation (0.887),
ionosphere (0.930)
Fair (0.60–0.79)12adult (0.760), penguins (0.537), phoneme (0.765)
Poor (<0.60)8analcatdata_dmft (0.181), parity5 (0.286)
Table 12. Comparative Performance of OIKAN and Baseline Models on Classification Datasets.
Table 12. Comparative Performance of OIKAN and Baseline Models on Classification Datasets.
ModelDatasetAccuracyPrecisionF1
OIKANiris (150 × 4)111
ElasticNet111
XGBoost0.9666670.9694440.966514
DecisionTree111
OIKANmonk3 (432 × 6)0.9550.9560.995
ElasticNet0.7027030.7166420.702703
XGBoost0.9819820.9819820.981982
DecisionTree0.9729730.973080.972946
OIKANmushroom (8124 × 22)111
ElasticNet0.9532310.9532350.95322
XGBoost111
DecisionTree111
OIKANcar_evaluation (1728 × 6)0.8870.8910.884
ElasticNet0.8121390.8016170.803303
XGBoost0.9797690.9857810.981074
DecisionTree0.9653180.9739080.967508
OIKANkr_vs_kp (3196 × 36)0.9470.9470.947
ElasticNet0.950.9519530.950049
XGBoost0.9890630.9890990.989059
DecisionTree0.9859380.9861520.985926
OIKANcoil2000 (9822 × 85)0.9090.8920.900
ElasticNet0.9374050.8929460.910041
XGBoost0.9262090.8967490.909317
DecisionTree0.8870230.892940.889946
Table 13. Comparative Performance of OIKAN and Baseline Models on Regression Datasets.
Table 13. Comparative Performance of OIKAN and Baseline Models on Regression Datasets.
ModelDatasetRMSER2MAPE
OIKAN1027_ESL (488 × 4)1.1418720.1705390.192851
ElasticNet0.7819920.6109850.140495
XGBoost0.6398340.7395670.092441
DecisionTree0.7079470.6811670.092292
OIKAN192_vineyard (632 × 4)2.6943280.6112470.125973
ElasticNet2.4549290.6772620.148458
XGBoost3.0133960.5137210.169473
DecisionTree4.0120560.1380010.217817
OIKAN225_puma8NH (8192 × 8)5.878939-0.114661.53969
ElasticNet4.665710.297932.446583
XGBoost3.5038050.6040643.433912
DecisionTree4.6422250.304983.676655
OIKAN197_cpu_act (8192 × 21)1.19 × 1011−4.8 × 10191.67 × 1024
ElasticNet8.9179870.7345984.3 × 1015
XGBoost2.4822310.9794388.13 × 1013
DecisionTree3.5187980.958681.37 × 1013
OIKAN215_2dplanes (40k × 10)2.169760.7533292.666862
ElasticNet3.2118940.4594742.692004
XGBoost1.0189630.9455981.315541
DecisionTree1.3706010.9015731.936285
OIKAN229_pwLinear (10k × 10)1.6577090.80180.639679
ElasticNet2.4344350.5725520.766036
XGBoost1.8460410.7542070.644108
DecisionTree2.4138890.5797360.829767
Table 14. Quantitative Comparison of OIKAN and XGBoost on Credit Score.
Table 14. Quantitative Comparison of OIKAN and XGBoost on Credit Score.
Model & ParametersTraining Time (Total)Weighted
F1-Score
AccuracyFormula
OIKANClassifier (with 5× Data Augmentation)30m 47s0.470.56Yes
OIKANClassifier (without Data Augmentation) 2.8 s0.630.64Yes
XGBClassifier4.5 s0.750.75No
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yedilkhan, D.; Zhalgasbayev, A.; Saleshova, S.; Khaimuldin, N. OIKAN: A Hybrid AI Framework Combining Symbolic Inference and Deep Learning for Interpretable Information Retrieval Models. Algorithms 2025, 18, 639. https://doi.org/10.3390/a18100639

AMA Style

Yedilkhan D, Zhalgasbayev A, Saleshova S, Khaimuldin N. OIKAN: A Hybrid AI Framework Combining Symbolic Inference and Deep Learning for Interpretable Information Retrieval Models. Algorithms. 2025; 18(10):639. https://doi.org/10.3390/a18100639

Chicago/Turabian Style

Yedilkhan, Didar, Arman Zhalgasbayev, Sabina Saleshova, and Nursultan Khaimuldin. 2025. "OIKAN: A Hybrid AI Framework Combining Symbolic Inference and Deep Learning for Interpretable Information Retrieval Models" Algorithms 18, no. 10: 639. https://doi.org/10.3390/a18100639

APA Style

Yedilkhan, D., Zhalgasbayev, A., Saleshova, S., & Khaimuldin, N. (2025). OIKAN: A Hybrid AI Framework Combining Symbolic Inference and Deep Learning for Interpretable Information Retrieval Models. Algorithms, 18(10), 639. https://doi.org/10.3390/a18100639

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop