AI Reasoning in Deep Learning Era: From Symbolic AI to Neural–Symbolic AI
Abstract
:1. Introduction
2. Historical View: From Symbolic Systems to Hybrid Intelligence
2.1. A Brief History of AI Reasoning
2.1.1. Symbolic Reasoning Era (1950s–1980s)
2.1.2. Statistical and Data-Driven Era (1990s–2010s)
2.1.3. Neural–Symbolic Integration Era (2016–Present)
2.2. Definition and Formalization of AI Reasoning
2.2.1. Formal Characterizations Across Paradigms
- is background knowledge (e.g., axioms, ontologies);
- are current inputs or observations;
- are conclusions derived by logical entailment;
- ⊧ is syntactic or semantic entailment (e.g., modus ponens, resolution).
- X are observed data (e.g., features, facts);
- Y is the target variable to be inferred;
- K is prior knowledge encoded in a probabilistic graphical model;
- are model parameters (e.g., conditional probabilities);
- is the most probable or expected outcome.
- X is the high-dimensional input (e.g., image pixels, word tokens, graph embeddings);
- is the task-dependent output (e.g., answer, class label, entity).
- X is the user input or natural context (e.g., question, image caption);
- P is the prompt scaffold or few-shot template (e.g., chain-of-thought);
- T is a large pretrained Transformer model;
- is the model-generated output (e.g., answer, plan, proof explanation).
- X are input observations (e.g., a question, an image, a scene graph);
- is a neural encoder or predictor parameterized by ;
- is symbolic knowledge, such as rules, ontologies, or graphs;
- are structured outputs inferred jointly from symbolic and neural components.
2.2.2. Categorization of AI Reasoning Across Dimensions
- Symbolic Reasoning: Relies on discrete, human-interpretable representations such as logic rules, graphs, and ontologies [94,95]. Inference is typically performed using deductive or rule-based systems, enabling traceability and formal verification [88]. Such systems dominate early expert systems and theorem provers.
- Statistical Reasoning: Model uncertainty using probability distributions over structured data. Reasoning tasks involve belief updating, probabilistic inference, and marginalization, often leveraging tools like Bayesian Networks [47], HMMs [96], or MLNs [49]. Logic-based probabilistic systems such as ProbLog [50] and PSL [89] also fall under this category.
- Neural Reasoning: Employs continuous, learned representations within neural networks. Inference emerges from multi-layer transformations and pattern abstraction, without explicit rule structures. Despite its black-box nature, neural reasoning has demonstrated success in perception-rich and language-heavy tasks [14,90].
- Hybrid (Neural–Symbolic) Reasoning: Attempts to unify the interpretability of symbolic models with the flexibility of neural networks. This includes architectures that inject symbolic priors into differentiable computation (e.g., ∂ILP [19], Logical Tensor Networks [69]), or use neural controllers to invoke symbolic tools (e.g., NS-CL [22], NeurASP [81], K-BERT [20]).
3. Technical View: Architectures, Mechanisms, and Trends of AI Reasoning in Deep Learning Era
- Knowledge-Representation Integration: How can background knowledge, logical rules, or ontologies be injected into gradient-based models while preserving learnability and robustness?
- Inference Control and Modularity: How can neural architectures learn and execute structured reasoning steps—such as multi-hop deduction or conditional branching—without explicit supervision or logic templates?
- Consistency, Verifiability, and Hallucination Mitigation: Neural models, especially LLMs, often generate plausible but factually incorrect inferences. How can we constrain these systems to produce faithful and logically sound conclusions?
3.1. Differentiable Logic Programming
- is observable or given knowledge, such as symbolic facts extracted from structured data (e.g., knowledge graphs), outputs of perception modules (e.g., scene relations), or labeled logical atoms;
- is a parameterized set of logic rule templates. In some settings, the structure of rules (e.g., Horn clauses) is predefined, and only rule weights are learned. In more general settings, both the structure and weights of rules can be induced from data—commonly referred to as Neural Inductive Logic Programming (Neural ILP). This variant enables the system to discover, select, or generate effective symbolic patterns during training;
- is a differentiable reasoning engine that approximates logical inference, such as matrix computation, graph neural networks, tensor composition, and probabilistic logic semantics, which may differ among different methods;
- is a loss function combining predictive error and rule regularization.
- is standard task loss (e.g., cross-entropy);
- is a rule-level regularization term that enforces sparsity, logical consistency, or syntactic constraints over the soft rule set;
- is a hyperparameter controlling the strength of the symbolic regularizer.
- Facts: ;
- Soft Rule Set: , where the rule template is and is a learnable rule confidence weight;
- Reasoning Engine: aggregates soft predictions from each rule application. The operator simulates differentiable forward chaining for rule over facts ;
- Supervision Target: ;
- Training Objective: , where is cross-entropy loss between predicted and ground truth facts, and regularizes rule sparsity or logical coherence.
3.2. Abductive Learning
3.3. Program Induction and Neural–Symbolic Concept Learners
3.4. LLM-Based Reasoning
- Chain-of-Thought (CoT) Prompting: Introduced to enable intermediate reasoning steps in LLMs [92], CoT encourages models to produce step-by-step symbolic or numeric inferences, improving accuracy on arithmetic, commonsense, and logic tasks.
- Program-Aided Reasoning (PAL): LLMs generate intermediate code (e.g., in Python) that is executed externally to perform symbolic or procedural computation [127]. This extends program induction to tool-based execution.
- : “Each pencil costs $2, and you buy 3 pencils.”
- : “Total cost of pencils: $2 × 3 = $6.” (CoT)
- : “Each pen costs $5, and you buy 2 pens.”
- : “Total cost of pens: $5 × 2 = $10.” (CoT)
- : “Total cost = $6 + $10 = $16.” (Final answer)
- Tool orchestration: Determining when and how to invoke external tools remains non-trivial, especially under uncertainty;
- Reasoning faithfulness: LLM-generated reasoning paths may be logically inconsistent or hallucinated, even if answers are correct;
- Symbol grounding: Aligning abstract reasoning steps with executable semantics requires fine-grained interface design;
- Reproducibility and verifiability: Unlike formal methods, LLM reasoning traces can vary stochastically, complicating reliability.
3.5. Logic-Aware Transformers
- are input data, typically structured data, text, or symbolic representations;
- is a Transformer model parameterized by , augmented with logic-aware mechanisms such as constrained attention or logical embedding;
- are symbolic logic constraints or logical structure templates integrated within the Transformer, guiding attention mechanisms or loss functions;
- is a loss function combining task-specific predictive losses and logic constraint losses.
- is task-specific loss (e.g., cross-entropy);
- is a logic-based regularization term enforcing adherence to symbolic constraints;
- is a hyperparameter controlling constraint regularization strength.
- Input Data (): A natural language question, e.g., “Which movies directed by Christopher Nolan have a rating above 8?”;
- Logical Constraints (): Domain-specific logical form constraints, e.g., a logical template ensuring valid SQL-like outputs: SELECT Movie WHERE Director = X AND Rating > Y;
- Logic-aware Transformer (): Transformer-based encoder–decoder architecture with attention mechanisms guided by logical form templates and constraints, ensuring output validity;
- Supervision Target: Ground-truth logical form SELECT Movie WHERE Director =“Christopher Nolan” AND Rating > 8;
- Training Objective: , balancing predictive accuracy and logical consistency.
3.6. Knowledge-Augmented Reasoning
- are input data (e.g., natural language queries, perceptual inputs);
- is a structured external knowledge base (e.g., knowledge graphs, logical rules);
- is a neural encoder–decoder model parameterized by ;
- is a knowledge integration mechanism, e.g., embedding lookups, GNNs, or differentiable reasoning layers;
- is the loss function combining prediction accuracy and knowledge grounding losses.
- Input Data (): A natural language query, e.g., “Who is the author of ’The Lord of the Rings’?”;
- Knowledge Base (): a structured knowledge graph with entities (authors, books) and relationships (authoredBy);
- Neural Model (): Transformer-based encoder–decoder that encodes the question and decodes the answer;
- Integration Mechanism (): GNN that retrieves relevant entities and relations from the knowledge graph, producing embeddings integrated into the Transformer’s hidden states;
- Supervision Target: ground-truth answer J.R.R. Tolkien;
- Training Objective: , optimizing accuracy and knowledge grounding consistency.
3.7. Multimodal Neuro-Symbolic Reasoning
- are input data across multiple modalities (e.g., images, text, audio, video);
- is symbolic representation derived from multimodal inputs, such as scene graphs, logical forms, or semantic frames;
- is a neural multimodal encoder–decoder parameterized by ;
- is a symbolic reasoning module, often implemented through differentiable symbolic reasoning engines or logic-based inference mechanisms;
- is the loss function balancing prediction accuracy, modality alignment, and symbolic consistency.
- Multimodal Inputs (): A video clip and a natural language question, e.g., “Why did the person enter the room?”;
- Symbolic Representation (): A structured event graph representing entities, actions, and temporal–causal relationships extracted from the video;
- Neural Multimodal Model (): A Transformer-based multimodal encoder that encodes video frames and linguistic queries into unified multimodal embeddings;
- Symbolic Reasoning Module (): A differentiable reasoning engine that performs logical inference on the event graph to infer the causal reason;
- Supervision Target: Natural language answer, e.g.,“to attend a meeting”;
- Training Objective: , optimizing answer correctness and symbolic reasoning coherence.
3.8. Conclusion and Emerging Trends
4. Application View: Benchmarks, Datasets, and Reasoning-Oriented AI Systems
4.1. Reasoning-Centric Tasks and Benchmarks
- Deductive QA. Tasks such as ProofWriter [189] and FOLIO [190] involve formal logic reasoning, requiring models to infer conclusions from natural language premises using entailment, conjunction, and implication rules. These benchmarks emphasize the need for systematic generalization over formal logical forms.
- Abductive and Causal QA. Datasets such as AbductiveNLI [194], ART [195], CausalQA [196], and SituatedQA [197] evaluate models on inferring plausible causes or explanations behind observed scenarios. Such tasks reflect the importance of abductive reasoning and counterfactual analysis in explainable AI.
- Symbolic Planning. Symbolic planning focuses on solving goal-directed tasks by generating action sequences under symbolic representations of states, transitions, and constraints. In contrast to purely reactive control policies, symbolic planning requires agents to explicitly reason about future states, preconditions, and causality. This often entails the construction of intermediate symbolic structures such as logic programs, action graphs, or scene representations, enabling long-horizon planning, interpretability, and task compositionality. PUZZLES [200] present structured environments composed of discrete logic-based tasks (e.g., grid games, puzzles, combinatorial path planning), designed to test whether agents can generalize across symbolic domains and solve algorithmic reasoning problems under limited feedback. RSBench [201] introduces a suite of neuro-symbolic reasoning environments targeting concept-level evaluation. SCERL [202] adapts textual reinforcement learning environments for safe and interpretable planning, covering sub-domains like side-effect minimization and reward uncertainty. In the domain of physical and robotic environments, RLBench [203] serves as a high-dimensional benchmark featuring over 100 task variants ranging from simple object manipulation to multi-step tool use.
- Multi-Agent and Interactive Planning. Frameworks like CAMEL [171] leverage symbolic role assignments and structured dialogue policies to coordinate among collaborative agents. They enable decentralized planning, intention modeling, and joint task execution in social or multi-agent settings.
- Video and Event Reasoning. Benchmarks including CLEVRER [206], NExT-QA [207], and VideoCoT [82] evaluate temporal and causal reasoning in video contexts, such as predicting future states, identifying event chains, and explaining dynamic processes. Symbolic modules or causal priors are often critical in modeling these temporal dependencies.
- Embodied and Situated Reasoning. In robotics and embodied environments, agents must perform goal-oriented reasoning from partial observations. Systems increasingly utilize symbolic representations (e.g., scene graphs or task logic) derived from perceptual inputs to support grounding, action abstraction, and generalizable planning. Recent approaches such as Embodied Chain-of-Thought Reasoning (ECoT) [208] and Inner Monologue [209] demonstrate how integrating structured reasoning steps and leveraging language model feedback can enhance robotic control and planning capabilities in complex environments.
- Program Synthesis. Tasks like CODET [213], NL2Bash [214], and MathQA [215] involve generating symbolic code from language descriptions or problems. Success in these tasks depends on precise logical translation, error handling, and explanation capabilities, particularly in mathematical or command-line environments.
4.2. Reasoning Frameworks and Toolkits
- Question Answering and Explainable Search. Neuro-symbolic systems such as DeepProbLog [80] and DSPy [77] have been integrated into QA pipelines to provide interpretable reasoning traces alongside factual answers. Logic-enhanced retrieval and reasoning has also shown promise in scientific QA and legal document analysis.
- Tool-Augmented Agent Systems. Toolchain frameworks such as LangChain [169], DSPy [77], and AgentBench [166] allow LLMs to invoke APIs, retrieve documents, and invoke solvers in complex reasoning workflows. These systems have been deployed in domains like software engineering, autonomous planning, and complex report generation.
5. Closing Remarks and Future Directions
- Unified Architectures. Existing systems are often task-specific or handcrafted. Achieving general-purpose, reusable reasoning modules remains an unsolved problem.
- Symbol-Vector Bridging. Seamlessly combining discrete symbolic structures with continuous neural representations requires more principled modeling and training strategies.
- Reasoning under Uncertainty. While probabilistic and fuzzy logic frameworks exist, efficient integration with deep perception remains limited in practice.
- Explainability and Trust. As reasoning systems are increasingly deployed in sensitive applications such as healthcare and law, their transparency, robustness, and ethical alignment become essential.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Turing, A.M. Computing machinery and intelligence. Mind 1950, 59, 433–460. [Google Scholar] [CrossRef]
- Russell, S.; Norvig, P. Artificial Intelligence: A Modern Approach, 4th ed.; Pearson: London, UK, 2021. [Google Scholar]
- Goertzel, B.; Pennachin, C. Artificial General Intelligence; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Marcus, G. Deep learning: A critical appraisal. arXiv 2018, arXiv:1801.00631. [Google Scholar]
- Lake, B.M.; Ullman, T.D.; Tenenbaum, J.B.; Gershman, S.J. Building Machines That Learn and Think Like People. Behav. Brain Sci. 2017, 40, e253. [Google Scholar] [CrossRef] [PubMed]
- Pearl, J. Causality: Models, Reasoning, and Inference, 2nd ed.; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
- Wang, X.; Zhang, S.; Cen, J.; Gao, C.; Zhang, Y.; Zhao, D.; Sang, N. CLIP-guided Prototype Modulating for Few-shot Action Recognition. Int. J. Comput. Vis. 2023, 132, 1899–1912. [Google Scholar] [CrossRef]
- Brachman, R.J.; Levesque, H.J. Knowledge Representation and Reasoning; The Morgan Kaufmann Series in Artificial Intelligence; Morgan Kaufmann: San Francisco, CA, USA, 2004. [Google Scholar]
- Newell, A.; Simon, H.A. Computer Science as Empirical Inquiry: Symbols and Search. Commun. ACM 1976, 19, 113–126. [Google Scholar] [CrossRef]
- Shortliffe, E.H.; Buchanan, B.G. A model of inexact reasoning in medicine. Math. Biosci. 1975, 23, 351–379. [Google Scholar] [CrossRef]
- Campbell, M.; Hoane, A.J.; Hsu, F.h. Deep Blue. Artif. Intell. 2002, 134, 57–83. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 4–9 December 2017; NIPS’17. pp. 6000–6010. [Google Scholar]
- Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; van den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the Game of Go with Deep Neural Networks and Tree Search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef]
- Garcez, A.d.; Lamb, L.C. Neurosymbolic AI: The 3rd Wave. Artif. Intell. Rev. 2022, 55, 2245–2274. [Google Scholar] [CrossRef]
- Bengio, Y. From System 1 Deep Learning to System 2 Deep Learning. Available online: https://neurips.cc/virtual/2019/invited-talk/15488 (accessed on 11 December 2019).
- Zhang, B.; Zhu, J.; Hang, S. Toward the third generation of artificial intelligence. Sci. Sin. (Informationis) 2020, 9, 1281–1302. [Google Scholar] [CrossRef]
- Evans, R.; Grefenstette, E. Learning explanatory rules from noisy data. J. Artif. Intell. Res. 2018, 61, 1–64. [Google Scholar] [CrossRef]
- Liu, W.; Zhou, P.; Zhao, Z.; Wang, Z.; Ju, Q.; Deng, H.; Wang, P. K-BERT: Enabling Language Representation with Knowledge Graph. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 2901–2908. [Google Scholar]
- Bordes, A.; Usunier, N.; Garcia-Durán, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. In Proceedings of the 27th International Conference on Neural Information Processing Systems–Volume 2, Red Hook, NY, USA, 5–10 December 2013; NIPS’13. pp. 2787–2795. [Google Scholar]
- Mao, J.; Gan, C.; Kohli, P.; Tenenbaum, J.B.; Wu, J. The neuro-symbolic concept learner: Interpreting scenes, words, and sentences from natural supervision. arXiv 2019, arXiv:1904.12584. [Google Scholar]
- Besold, T.R.; d’Avila Garcez, A.; Bader, S.; Bowman, H.; Domingos, P.; Hitzler, P.; Kuehnberger, K.U.; Lamb, L.C.; Lowd, D.; Lima, P.M.V.; et al. Neural-Symbolic Learning and Reasoning: A Survey and Interpretation. In Neuro-Symbolic Artificial Intelligence: The State of the Art; IOS Press: Amsterdam, The Netherlands, 2017. [Google Scholar]
- Wan, Z.; Liu, C.K.; Yang, H.; Li, C.; You, H.; Fu, Y.; Wan, C.; Krishna, T.; Lin, Y.; Raychowdhury, A. Towards Cognitive AI Systems: A Survey and Prospective on Neuro-Symbolic AI. arXiv 2024, arXiv:2401.01040. [Google Scholar]
- Li, Z.Z.; Zhang, D.; Zhang, M.L.; Zhang, J.; Liu, Z.; Yao, Y.; Xu, H.; Zheng, J.; Wang, P.J.; Chen, X.; et al. From System 1 to System 2: A Survey of Reasoning Large Language Models. arXiv 2025, arXiv:2502.17419. [Google Scholar]
- Bock, H.H.; Diday, E. Analysis of Symbolic Data: Exploratory Methods for Extracting Statistical Information from Complex Data; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
- Colmerauer, A.; Roussel, P. The birth of Prolog. In History of Programming Languages—II; Association for Computing Machinery: New York, NY, USA, 1996; pp. 331–367. [Google Scholar]
- Newell, A.; Simon, H.A. GPS, a program that simulates human thought. In Computation & Intelligence: Collected Readings; American Association for Artificial Intelligence: Washington, DC, USA, 1995; pp. 415–428. [Google Scholar]
- Michalski, R.S. Variable-valued logic and its applications to pattern recognition and machine learning. In Computer Science and Multiple-Valued Logic; Rine, D.C., Ed.; Elsevier: Amsterdam, The Netherlands, 1977; pp. 506–534. [Google Scholar] [CrossRef]
- Buchanan, B.G.; Feigenbaum, E.A. Dendral and meta-dendral: Their applications dimension. Artif. Intell. 1978, 11, 5–24. [Google Scholar] [CrossRef]
- Giarratano, J.C.; Riley, G. Expert Systems: Principles and Programming; Brooks/Cole Publishing Co.: Pacific Grove, CA, USA, 1989. [Google Scholar]
- Forgy, C.L. Rete: A fast algorithm for the many pattern/many object pattern match problem. In Readings in Artificial Intelligence and Databases; Elsevier: Amsterdam, The Netherlands, 1989; pp. 547–559. [Google Scholar]
- Reiter, R. A logic for default reasoning. Artif. Intell. 1980, 13, 81–132. [Google Scholar] [CrossRef]
- McCarthy, J. Circumscription—A form of non-monotonic reasoning. Artif. Intell. 1980, 13, 27–39. [Google Scholar] [CrossRef]
- Gelfond, M.; Lifschitz, V. Classical negation in logic programs and disjunctive databases. New Gener. Comput. 1991, 9, 365–385. [Google Scholar] [CrossRef]
- Doyle, J. A truth maintenance system. Artif. Intell. 1979, 12, 231–272. [Google Scholar] [CrossRef]
- Fikes, R.E.; Nilsson, N.J. STRIPS: A new approach to the application of theorem proving to problem solving. Artif. Intell. 1971, 2, 189–208. [Google Scholar] [CrossRef]
- Nilsson, N.J. Shakey the Robot. In SRI International AI Center Technical Note; SRI International: Menlo Park, CA, USA, 1984. [Google Scholar]
- Dung, P.M. On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games. Artif. Intell. 1995, 77, 321–357. [Google Scholar] [CrossRef]
- Modgil, S.; Prakken, H. The ASPIC+ framework for structured argumentation: A tutorial. Argum. Comput. 2014, 5, 31–62. [Google Scholar] [CrossRef]
- Brachman, R.J.; Schmolze, J.G. An overview of the KL-ONE knowledge representation system. Cogn. Sci. 1985, 9, 171–216. [Google Scholar] [CrossRef]
- Baader, F. The Description Logic Handbook: Theory, Implementation and Applications; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
- McGuinness, D.L.; Van Harmelen, F. OWL web ontology language overview. W3C Recomm. 2004, 10, 2004. [Google Scholar]
- Hintikka, J. Knowledge and Belief: An Introduction to the Logic of the Two Notions; Texts in Philosophy; King’s College Publications: London, UK, 2005. [Google Scholar]
- Pnueli, A. The temporal logic of programs. In Proceedings of the 18th Annual Symposium on Foundations of a Computer Science (sfcs 1977), Providence, RI, USA, 31 October–2 November 1977; pp. 46–57. [Google Scholar]
- Emerson, E.A. Temporal and modal logic. In Formal Models and Semantics; Elsevier: Amsterdam, The Netherlands, 1990; pp. 995–1072. [Google Scholar]
- Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1988. [Google Scholar]
- Koller, D.; Friedman, N. Probabilistic Graphical Models: Principles and Techniques; MIT Press: Cambridge, MA, USA, 2009. [Google Scholar]
- Richardson, M.; Domingos, P. Markov logic networks. Mach. Learn. 2006, 62, 107–136. [Google Scholar] [CrossRef]
- De Raedt, L.; Kimmig, A.; Toivonen, H. ProbLog: A probabilistic Prolog and its application in link discovery. In Proceedings of the IJCAI 2007, Proceedings of the 20th International Joint Conference on Artificial Intelligence, Hyderabad, India, 6–12 January 2007; pp. 2462–2467. [Google Scholar]
- Sato, T.; Kameya, Y. PRISM: A language for symbolic-statistical modeling. In Proceedings of the Fifteenth International Joint Conference on Artificial Intelligence, Nagoya, Japan, 23–29 August 1997; Volume 97, pp. 1330–1339. [Google Scholar]
- Riguzzi, F. Extended semantics and inference for the Independent Choice Logic. Log. J. IGPL 2009, 17, 589–629. [Google Scholar] [CrossRef]
- Jaeger, M. Relational Bayesian networks. In Proceedings of the Thirteenth Conference on Uncertainty in Artificial Intelligence (UAI), Providence, RI, USA, 1–3 August 1997; pp. 266–273. [Google Scholar]
- Getoor, L.; Taskar, B. Introduction to Statistical Relational Learning; MIT Press: Cambridge, MA, USA, 2007. [Google Scholar]
- De Raedt, L.; Frasconi, P.; Kersting, K.; Muggleton, S. (Eds.) Probabilistic Inductive Logic Programming: Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
- Michalski, R.S. Knowledge Acquisition Through Conceptual Clustering: A Theoretical Framework and an Algorithm for Partitioning Data into Conjunctive Concepts. J. Policy Anal. Inf. Syst. 1980, 4, 219–244. [Google Scholar]
- Michalski, R.S.; Stepp, R.E. Automated construction of classifications: Conceptual clustering versus numerical taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 1983, PAMI-5, 396–410. [Google Scholar] [CrossRef]
- Fisher, D.H. Knowledge Acquisition via Incremental Conceptual Clustering. Mach. Learn. 1987, 2, 139–172. [Google Scholar] [CrossRef]
- Pearl, J.; Mackenzie, D. The Book of Why: The New Science of Cause and Effect; BasicBooks: New York, NY, USA, 2018. [Google Scholar]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. arXiv 2018, arXiv:1812.08434. [Google Scholar] [CrossRef]
- Kok, S.; Sumner, M.; Richardson, M.; Singla, P.; Poon, H.; Lowd, D.; Domingos, P. The Alchemy System for Statistical Relational AI; Technical Report; Department of Computer Science and Engineering, University of Washington: Seattle, WA, USA, 2007; Available online: http://alchemy.cs.washington.edu (accessed on 20 April 2025).
- Niu, F.; Zhang, C.; Ré, C.; Shavlik, J. Tuffy: Scaling up statistical inference in Markov Logic Networks using an RDBMS. Proc. Vldb Endow. 2011, 4, 373–384. [Google Scholar] [CrossRef]
- Dong, H.; Mao, J.; Lin, C.G. Neural Logic Machines. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Serafini, L.; d’Avila Garcez, A. Learning and Reasoning with Logic Tensor Networks. In Proceedings of the Conference of the Italian Association for Artificial Intelligence, Genova, Italy, 29 November 2016. [Google Scholar]
- Zhou, Z.H. Abductive learning: Towards bridging machine learning and logical reasoning. Sci. China Inf. Sci. 2019, 62, 76101. [Google Scholar] [CrossRef]
- Andreas, J.; Rohrbach, M.; Darrell, T.; Klein, D. Neural Module Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 39–48. [Google Scholar]
- Johnson, J.; Hariharan, B.; van der Maaten, L.; Hoffman, J.; Fei-Fei, L.; Zitnick, C.L.; Girshick, R. CLEVR: A diagnostic dataset for compositional language and elementary visual reasoning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2901–2910. [Google Scholar]
- OpenAI. GPT-4 Technical Report. 2023. Available online: https://openai.com/research/gpt-4 (accessed on 24 April 2025).
- Anthropic. The Claude 3 Model Family: Opus, Sonnet, Haiku. 2024. Available online: https://www.anthropic.com/news/claude-3-family (accessed on 17 April 2025).
- Yao, S.; Zhao, J.; Yu, D.; Du, N.; Shafran, I.; Narasimhan, K.; Cao, Y. React: Synergizing reasoning and acting in language models. In Proceedings of the International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
- Schick, T.; Dwivedi-Yu, J.; Dessì, R.; Raileanu, R.; Lomeli, M.; Hambro, E.; Zettlemoyer, L.; Cancedda, N.; Scialom, T. Toolformer: Language models can teach themselves to use tools. Adv. Neural Inf. Process. Syst. 2023, 36, 68539–68551. [Google Scholar]
- Khattab, O.; Singhvi, A.; Maheshwari, P.; Zhang, Z.; Santhanam, K.; Vardhamanan, S.; Haq, S.; Sharma, A.; Joshi, T.T.; Moazam, H.; et al. Dspy: Compiling declarative language model calls into self-improving pipelines. arXiv 2023, arXiv:2310.03714. [Google Scholar]
- Wang, B.R.; Huang, Q.Y.; Deb, B.; Halfaker, A.; Shao, L.Q.; McDuff, D.; Awadallah, A.H.; Radev, D.; Gao, J.F. Logical Transformers: Infusing Logical Structures into Pre-Trained Language Models. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, Toronto, ON, Canada, 9–14 July 2023. [Google Scholar]
- Parmar, M.; Patel, N.; Varshney, N.; Nakamura, M.; Luo, M.; Mashetty, S.; Mitra, A.; Baral, C. LogicBench: Towards Systematic Evaluation of Logical Reasoning Ability of Large Language Models. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; pp. 13679–13707. [Google Scholar] [CrossRef]
- Manhaeve, R.; Dumancic, S.; Kimmig, A.; Demeester, T.; De Raedt, L. DeepProbLog: Neural Probabilistic Logic Programming. In Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
- Yang, Z.; Ishay, A.; Lee, J. NeurASP: Embracing Neural Networks into Answer Set Programming. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, Yokohama, Japan, 7–15 January 2020; Bessiere, C., Ed.; pp. 1755–1762. [Google Scholar] [CrossRef]
- Wang, Y.; Zeng, Y.; Zheng, J.; Xing, X.; Xu, J.; Xu, X. VideoCoT: A Video Chain-of-Thought Dataset with Active Annotation Tool. In Proceedings of the 3rd Workshop on Advances in Language and Vision Research (ALVR), Bangkok, Thailand, 16 August 2024; Gu, J., Fu, T.J.R., Hudson, D., Celikyilmaz, A., Wang, W., Eds.; pp. 92–101. [Google Scholar] [CrossRef]
- Surís, D.; Menon, S.; Vondrick, C. ViperGPT: Visual Inference via Python Execution for Reasoning. arXiv 2023, arXiv:2303.08128. [Google Scholar]
- Huang, Y.X.; Sun, Z.Q.; Li, G.Y.; Tian, X.B.; Dai, W.Z.; Hu, W.; Jiang, Y.; Zhou, Z.H. Enabling Abductive Learning to Exploit Knowledge Graph. In Proceedings of the 32th International Joint Conference on Artificial Intelligence (IJCAI), Macao SAR, China, 19–25 August 2023; pp. 2730–2736. [Google Scholar]
- Saparov, A.; He, H. Language modeling via stochastic processes and data augmentation. arXiv 2022, arXiv:2205.09310. [Google Scholar]
- Marcus, G.; Davis, E. Rebooting AI: Building Artificial Intelligence We Can Trust; Pantheon: London, UK, 2019. [Google Scholar]
- Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach, 3rd ed.; Prentice Hall: Hoboken, NJ, USA, 2010. [Google Scholar]
- Robinson, J.A. A machine-oriented logic based on the resolution principle. J. ACM (JACM) 1965, 12, 23–41. [Google Scholar] [CrossRef]
- Bach, S.H.; Broecheler, M.; Huang, B.; Getoor, L. Hinge-loss markov random fields and probabilistic soft logic. J. Mach. Learn. Res. 2017, 18, 1–67. [Google Scholar]
- Battaglia, P.; Hamrick, J.B.; Bapst, V.; Sanchez-Gonzalez, A.; Zambaldi, V.; Malinowski, M.; Tacchetti, A.; Raposo, D.; Santoro, A.; Faulkner, R.; et al. Relational inductive biases, deep learning, and graph networks. arXiv 2018, arXiv:1806.01261. [Google Scholar]
- Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
- Wei, J.; Wang, X.; Schuurmans, D.; Bosma, M.; Xia, F.; Chi, E.; Le, Q.V.; Zhou, D.; Ichter, B. Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural Inf. Process. Syst. 2022, 35, 24824–24837. [Google Scholar]
- Ganguli, D.; Hernandez, D.; Lovitt, L.; DasSarma, N.; Henighan, T.; Jones, A.; Joseph, N.; Kernion, J.; Mann, B.; Askell, A.; et al. Predictability and surprise in large generative models. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, 21–24 June 2022; pp. 1747–1764. [Google Scholar]
- Abiteboul, S.; Hull, R.; Vianu, V. Foundations of Databases; Addison-Wesley: Boston, MA, USA, 1995. [Google Scholar]
- Baader, F.; Horrocks, I.; Sattler, U. The Description Logic Handbook; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
- Rabiner, L.R. A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition; IEEE: New York, NY, USA, 1989; Volume 77, pp. 257–286. [Google Scholar]
- De Moura, L.; Bjørner, N. Z3: An Efficient SMT Solver. In Tools and Algorithms for the Construction and Analysis of Systems (TACAS); Springer: Berlin/Heidelberg, Germany, 2008; pp. 337–340. [Google Scholar]
- Muggleton, S. Inverse entailment and Progol. New Gener. Comput. 1995, 13, 245–286. [Google Scholar] [CrossRef]
- Cropper, A.; Muggleton, S. Learning efficient logic programs. In Proceedings of the ILP Conference, London, UK, 4–6 September 2016. [Google Scholar]
- Falkenhainer, B.; Forbus, K.; Gentner, D. The structure-mapping engine: Algorithm and examples. Artif. Intell. 1989, 41, 1–63. [Google Scholar] [CrossRef]
- Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the 38th International Conference on Machine Learning (ICML), Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
- Kaelbling, L.P.; Littman, M.L.; Cassandra, A.R. Planning and acting in partially observable stochastic domains. Artif. Intell. 1998, 101, 99–134. [Google Scholar] [CrossRef]
- Palmirani, M.; Governatori, G.; Rotolo, A.; Sartor, G.; Tabet, S.; Boley, H. LegalRuleML: XML-based rules and norms. In Proceedings of the Rule-Based Modeling and Computing on the Semantic Web (RuleML), America, Ft. Lauderdale, FL, USA, 3–5 November 2011; pp. 298–312. [Google Scholar]
- Yang, Z.; Qi, P.; Zhang, S.; Bengio, Y.; Cohen, W.W.; Salakhutdinov, R.; Manning, C.D. HotpotQA: A dataset for diverse, explainable multi-hop question answering. arXiv 2018, arXiv:1809.09600. [Google Scholar]
- Reddy, S.; Chen, D.; Manning, C.D. CoQA: A conversational question answering challenge. Trans. Assoc. Comput. Linguist. 2019, 7, 249–266. [Google Scholar] [CrossRef]
- Taylor, R.; Kardas, M.; Cucurull, G.; Scialom, T.; Hartshorn, A.; Saravia, E.; Poulton, A.; Kerkez, V.; Stojnic, R. Galactica: A large language model for science. arXiv 2022, arXiv:2211.09085. [Google Scholar]
- Goodwin, T.; Demner-Fushman, D. Enhancing Question Answering by Injecting Ontological Knowledge through Regularization. In Proceedings of the Deep Learning Inside Out (DeeLIO): The First Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, Online, 19–20 November 2020; pp. 56–63. [Google Scholar] [CrossRef]
- Valmeekam, K.; Ellis, K.; Solar-Lezama, A.; Tenenbaum, J.B. DreamCoder: Growing generalizable, interpretable knowledge with wake-sleep Bayesian program learning. In Proceedings of the NeurIPS, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
- Morel, R.; Cropper, A.; Ong, C.-H.L. Typed meta-interpretive learning of logic programs. In Proceedings of the Logics in Artificial Intelligence (JELIA), Rende, Italy, 7–11 May 2019; pp. 198–213. [Google Scholar]
- Saad, F.A.; Cusumano-Towner, M.F.; Schaechtle, U.; Rinard, M.C.; Mansinghka, V.K. Bayesian synthesis of probabilistic programs for automatic data modeling. ACM Program. Lang. 2019, 3, 1–32. [Google Scholar] [CrossRef]
- Tai, K.S.; Socher, R.; Manning, C.D. Improved semantic representations from tree-structured long short-term memory networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics (ACL), Beijing, China, 26–31 July 2015; pp. 1556–1566. [Google Scholar]
- Adrita, B.; Cara, W.; Pascal, H. Concept Induction Using LLMs: A User Experiment for Assessment. In Proceedings of the International Conference on Neural-Symbolic Learning and Reasoning (NeSy), Barcelona, Spain, 9–12 September 2024. [Google Scholar]
- Yi, K.; Wu, J.; Gan, C.; Torralba, A.; Kohli, P.; Tenenbaum, J. Neural-Symbolic VQA: Disentangling Reasoning from Vision and Language Understanding. In Proceedings of the NeurIPS, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Reiter, R. A theory of diagnosis from first principles. Artif. Intell. 1987, 32, 57–95. [Google Scholar] [CrossRef]
- Ramírez, M.; Geffner, H. Plan Recognition as Planning. In Proceedings of the 21st International Joint Conference on Artificial Intelligence (IJCAI), Pasadena, CA, USA, 11–17 July 2009; pp. 1778–1783. [Google Scholar]
- Glória-Silva, D.; Ferreira, R.; Tavares, D.; Semedo, D.; Magalhaes, J. Plan-Grounded Large Language Models for Dual Goal Conversational Settings. In Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (EACL), Valletta, Malta, 21–22 March 2024; pp. 1271–1292. [Google Scholar]
- Liang, B.; Su, Q.; Zhu, S.; Liang, Y.; Tong, C. VidEvent: A Large Dataset for Understanding Dynamic Evolution of Events in Videos. Proc. AAAI Conf. Artif. Intell. 2025, 39, 5128–5136. [Google Scholar] [CrossRef]
- Shutova, E.; Sun, L.; Korhonen, A. Metaphor identification using verb and noun clustering. In Proceedings of the 23rd International Conference on Computational Linguistics, Beijing, China, 23–27 August 2010; pp. 1002–1010. [Google Scholar]
- Borgwardt, K.; Kriegel, H.P. Shortest-path kernels on graphs. In Proceedings of the Fifth IEEE International Conference on Data Mining (ICDM), Houston, TX, USA, 27–30 November 2005; pp. 74–81. [Google Scholar]
- Du, X.; Ji, H. Retrieval-augmented generative question answering for event argument extraction. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP), Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 4567–4578. [Google Scholar]
- Webb, T.; Holyoak, K.J.; Lu, H. Emergent analogical reasoning in large language models. Nat. Hum. Behav. 2023, 7, 1526–1541. [Google Scholar] [CrossRef]
- Bengio, Y. The consciousness prior. arXiv 2017, arXiv:1709.08568. [Google Scholar]
- LeCun, Y. A path towards autonomous machine intelligence. arXiv 2022, arXiv:2206.06927. [Google Scholar]
- Garcez, A.d.; Gori, M.; Lamb, L.C.; Serafini, L.; Spranger, M.; Tran, S.N. Neural-symbolic computing: An effective methodology for principled integration of machine learning and reasoning. arXiv 2019, arXiv:1905.06088. [Google Scholar]
- Yang, F.; Yang, Z.; Cohen, W.W. Differentiable learning of logical rules for knowledge base reasoning. Adv. Neural Inf. Process. Syst. 2017, 30, 2319–2328. [Google Scholar]
- Purgał, S.J.; Cerna, D.M.; Kaliszyk, C. Differentiable inductive logic programming in high-dimensional space. arXiv 2022, arXiv:2208.06652. [Google Scholar]
- Gao, K.; Inoue, K.; Cao, Y.; Wang, H. Learning first-order rules with differentiable logic program semantics. arXiv 2022, arXiv:2204.13570. [Google Scholar]
- Shindo, H.; Nishino, M.; Yamamoto, A. Differentiable inductive logic programming for structured examples. Proc. AAAI Conf. Artif. Intell. 2021, 35, 5034–5041. [Google Scholar] [CrossRef]
- Gao, K.; Inoue, K.; Cao, Y.; Wang, H. A Differentiable First-Order Rule Learner for Inductive Logic Programming. Artif. Intell. 2024, 331, 104108. [Google Scholar] [CrossRef]
- Rocktäschel, T.; Riedel, S. End-to-end differentiable proving. Adv. Neural Inf. Process. Syst. 2017, 30, 3788–3800. [Google Scholar]
- Cohen, W.W. Tensorlog: A differentiable deductive database. arXiv 2016, arXiv:1605.06523. [Google Scholar]
- Zimmer, M.; Feng, X.; Glanois, C.; Jiang, Z.; Zhang, J.; Weng, P.; Li, D.; Hao, J.; Liu, W. Differentiable logic machines. arXiv 2021, arXiv:2102.11529. [Google Scholar]
- Shindo, H.; Pfanschilling, V.; Dhami, D.S.; Kersting, K. Learning differentiable logic programs for abstract visual reasoning. Mach. Learn. 2024, 113, 8533–8584. [Google Scholar] [CrossRef]
- Takemura, A.; Inoue, K. Differentiable Logic Programming for Distant Supervision. In Proceedings of the European Conference on Artificial Intelligence, Santiago de Compostela, Spain, 19–24 October 2024. [Google Scholar]
- Serafini, L.; Garcez, A.d. Logic tensor networks: Deep learning and logical reasoning from data and knowledge. arXiv 2016, arXiv:1606.04422. [Google Scholar]
- Riegel, R.; Gray, A.; Luus, F.; Khan, N.; Makondo, N.; Akhalwaya, I.Y.; Qian, H.; Fagin, R.; Barahona, F.; Sharma, U.; et al. Logical Neural Networks. arXiv 2020, arXiv:2006.13155. [Google Scholar]
- Šourek, G.; Železný, F.; Kuželka, O. Beyond graph neural networks with lifted relational neural networks. Mach. Learn. 2021, 110, 1695–1738. [Google Scholar] [CrossRef]
- Geh, R.L.; Gonçalves, J.; Silveira, I.C.; Mauá, D.D.; Cozman, F.G. dPASP: A probabilistic logic programming environment for neurosymbolic learning and reasoning. In Proceedings of the International Conference on Principles of Knowledge Representation and Reasoning, Hanoi, Vietnam, 2–8 November 2024; Volume 21, pp. 731–742. [Google Scholar]
- Li, Z.; Huang, J.; Naik, M. Scallop: A Language for Neurosymbolic Programming. Proc. ACM Program. Lang. 2023, 7, 1463–1487. [Google Scholar] [CrossRef]
- Shindo, H.; Pfanschilling, V.; Dhami, D.S.; Kersting, K. α ilp: Thinking visual scenes as differentiable logic programs. Mach. Learn. 2023, 112, 1465–1497. [Google Scholar] [CrossRef]
- Peirce, C.S. Collected Papers of Charles Sanders Peirce; Harvard University Press: Cambridge, MA, USA, 1935; Volume 5. [Google Scholar]
- Poole, D. A logical framework for default reasoning. Artif. Intell. 1988, 36, 27–47. [Google Scholar] [CrossRef]
- Dai, W.Z.; Muggleton, S.H. Abductive Knowledge Induction from Raw Data. In Proceedings of the 30th International Joint Conference on Artificial Intelligence (IJCAI), Montreal, QC, Canada, 19–26 August 2021; pp. 2730–2736. [Google Scholar]
- Shao, J.J.; Hao, H.R.; Yang, X.W.; Li, Y.F. Abductive Learning for Neuro-Symbolic Grounded Imitation. In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.1, Toronto, ON, Canada, 3–7 August 2025; KDD ’25. pp. 1221–1232. [Google Scholar] [CrossRef]
- Hu, Y.Y.; Yu, Y. Enhancing Neural Mathematical Reasoning by Abductive Combination with Symbolic Library. arXiv 2023, arXiv:2203.14487. [Google Scholar]
- Hu, W.C.; Dai, W.Z.; Jiang, Y.; Zhou, Z.H. Efficient rectification of neuro-symbolic reasoning inconsistencies by abductive reflection. Proc. AAAI Conf. Artif. Intell. 2025, 39, 17333–17341. [Google Scholar] [CrossRef]
- Sun, Z.H.; Zhang, R.Y.; Zhen, Z.; Wang, D.H.; Li, Y.J.; Wan, X.; You, H. Systematic Abductive Reasoning via Diverse Relation Representations in Vector-symbolic Architecture. arXiv 2025, arXiv:2501.11896. [Google Scholar]
- Camposampiero, G.; Hersche, M.; Terzić, A.; Wattenhofer, R.; Sebastian, A.; Rahimi, A. Towards Learning Abductive Reasoning Using VSA Distributed Representations. In Neural-Symbolic Learning and Reasoning; Besold, T.R., d’Avila Garcez, A., Jimenez-Ruiz, E., Confalonieri, R., Madhyastha, P., Wagner, B., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 370–385. [Google Scholar]
- Hu, S.; Ma, Y.; Liu, X.; Wei, Y.; Bai, S. Stratified Rule-Aware Network for Abstract Visual Reasoning. Proc. AAAI Conf. Artif. Intell. 2021, 35, 1561–1569. [Google Scholar] [CrossRef]
- Jin, Y.; Liu, J.; Luo, Z.; Peng, Y.; Qin, Z.; Dai, W.Z.; Ding, Y.X.; Zhou, K. Pre-Training Meta-Rule Selection Policy for Visual Generative Abductive Learning. arXiv 2025, arXiv:2503.06427. [Google Scholar]
- Yang, X.W.; Wei, W.D.; Shao, J.J.; Li, Y.F.; Zhou, Z.H. Analysis for Abductive Learning and Neural-Symbolic Reasoning Shortcuts. In Proceedings of the 41st International Conference on Machine Learning, Vienna, Austria, 21–27 July 2024; Volume 235, pp. 56524–56541. [Google Scholar]
- Gupta, T.; Kembhavi, A. Visual Programming: Compositional visual reasoning without training. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
- Kamali, D.; Barezi, E.J.; Kordjamshidi, P. NeSyCoCo: A Neuro-Symbolic Concept Composer for Compositional Generalization. Proc. AAAI Conf. Artif. Intell. 2025, 39, 4184–4193. [Google Scholar] [CrossRef]
- Liang, J.; Huang, W.; Xia, F.; Xu, P.; Hausman, K.; Ichter, B.; Florence, P.; Zeng, A. Code as Policies: Language Model Programs for Embodied Control. In Proceedings of the 6th Conference on Robot Learning (CoRL), Auckland, New Zealand, 14–18 December 2022. [Google Scholar]
- Shin, R.; Kant, N.; Gupta, K.; Bender, C.; Trabucco, B.; Singh, R.; Song, D. Synthetic Datasets for Neural Program Synthesis. In Proceedings of the 7th International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Ellis, K.; Wong, L.; Nye, M.; Sable-Meyer, M.; Cary, L.; Anaya Pozo, L.; Hewitt, L.; Solar-Lezama, A.; Tenenbaum, J.B. DreamCoder: Growing generalizable, interpretable knowledge with wake–sleep Bayesian program learning. Philos. Trans. R. Soc. A 2023, 381, 20220050. [Google Scholar] [CrossRef] [PubMed]
- Khan, R.M.; Gulwani, S.; Le, V.; Radhakrishna, A.; Tiwari, A.; Verbruggen, G. LLM-Guided Compositional Program Synthesis. arXiv 2025, arXiv:2503.15540. [Google Scholar]
- Liang, C.; Berant, J.; Le, Q.; Forbus, K.D.; Lao, N. Neural symbolic machines: Learning semantic parsers on freebase with weak supervision. arXiv 2016, arXiv:1611.00020. [Google Scholar]
- Duan, X.; Wang, X.; Zhao, P.; Shen, G.; Zhu, W. DeepLogic: Joint Learning of Neural Perception and Logical Reasoning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 4321–4334. [Google Scholar] [CrossRef]
- Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; et al. Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 2023, 24, 1–113. [Google Scholar]
- Patil, S.G.; Zhang, T.; Wang, X.; Gonzalez, J.E. Gorilla: Large Language Model Connected with Massive APIs. arXiv 2023, arXiv:2305.15334. [Google Scholar]
- Gravitas, S. AutoGPT: An Autonomous GPT-4 Experiment. GitHub Repository. 2023. Available online: https://github.com/Torantulino/Auto-GPT (accessed on 20 May 2025).
- Wang, X.; Wei, J.; Schuurmans, D.; Le, Q.; Chi, E.; Narang, S.; Chowdhery, A.; Zhou, D. Self-consistency improves chain of thought reasoning in language models. arXiv 2022, arXiv:2203.11171. [Google Scholar]
- Zhou, D.; Schärli, N.; Hou, L.; Wei, J.; Scales, N.; Wang, X.; Schuurmans, D.; Cui, C.; Bousquet, O.; Le, Q.; et al. Least-to-most prompting enables complex reasoning in large language models. arXiv 2022, arXiv:2205.10625. [Google Scholar]
- Yao, S.; Yu, D.; Zhao, J.; Shafran, I.; Griffiths, T.; Cao, Y.; Narasimhan, K. Tree of thoughts: Deliberate problem solving with large language models. Adv. Neural Inf. Process. Syst. 2023, 36, 11809–11822. [Google Scholar]
- Liu, X.; Yu, H.; Zhang, H.; Xu, Y.; Lei, X.; Lai, H.; Gu, Y.; Ding, H.; Men, K.; Yang, K.; et al. Agentbench: Evaluating llms as agents. arXiv 2023, arXiv:2308.03688. [Google Scholar]
- Lewkowycz, A.; Andreassen, A.; Dohan, D.; Dyer, E.; Michalewski, H.; Ramasesh, V.; Slone, A.; Anil, C.; Schlag, I.; Gutman-Solo, T.; et al. Solving quantitative reasoning problems with language models. Adv. Neural Inf. Process. Syst. 2022, 35, 3843–3857. [Google Scholar]
- AtlasUnified. PyCoT: A Pythonic Chain-of-Thought Dataset Series. Hugging Face. 2025. Comprehensive CoT Expansions Across Multiple Domains. Available online: https://huggingface.co/datasets/AtlasUnified/PyCoT-Collection_Main (accessed on 1 May 2025).
- Chase, H. LangChain: Building Applications with LLMs Through Composability. 2023. Available online: https://www.langchain.com/ (accessed on 1 May 2025).
- Hong, S.; Zheng, X.; Chen, J.; Cheng, Y.; Wang, J.; Zhang, C.; Wang, Z.; Yau, S.K.S.; Lin, Z.; Zhou, L.; et al. Metagpt: Meta programming for multi-agent collaborative framework. arXiv 2023, arXiv:2308.00352. [Google Scholar]
- Li, G.; Hammoud, H.; Itani, H.; Khizbullin, D.; Ghanem, B. Camel: Communicative agents for “mind” exploration of large language model society. Adv. Neural Inf. Process. Syst. 2023, 36, 51991–52008. [Google Scholar]
- Miranda, B.; Shinnar, A.; Pestun, V.; Trager, B. Transformer Models for Type Inference in the Simply Typed Lambda Calculus: A Case Study in Deep Learning for Code. arXiv 2023, arXiv:2304.10500. [Google Scholar]
- Zhou, W.; Le Bras, R.; Choi, Y. Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, Toronto, ON, Canada, 9–14 July 2023; Rogers, A., Boyd-Graber, J., Okazaki, N., Eds.; pp. 10452–10465. [Google Scholar] [CrossRef]
- Herzig, J.; Berant, J. Span-based semantic parsing for compositional generalization. arXiv 2020, arXiv:2009.06040. [Google Scholar]
- Poulis, A.; Tsalapati, E.; Koubarakis, M. Transformers in the Service of Description Logic-Based Contexts. In Knowledge Engineering and Knowledge Management; Springer Nature: Cham, Switzerland, 2025; pp. 328–345. [Google Scholar]
- Brinkmann, B.J.; Smith, J.D.; Lee, C.M. How Transformers Solve Propositional Logic Problems: A Mechanistic Analysis. In Proceedings of the Workshop on Mechanistic Interpretability of Neural Networks at ICLR, Vienna, Austria, 27 July 2024. Workshop paper. [Google Scholar]
- Wang, X.; Gao, T.; Zhu, Z.; Zhang, Z.; Liu, Z.; Li, J.; Tang, J. KEPLER: A Unified Model for Knowledge Embedding and Pre-Trained Language Representation. Trans. Assoc. Comput. Linguist. 2021, 9, 176–194. [Google Scholar] [CrossRef]
- Winters, T.; Marra, G.; Manhaeve, R.; De Raedt, L. DeepStochLog: Neural Stochastic Logic Programming. Proc. AAAI Conf. Artif. Intell. 2022, 36, 10090–10100. [Google Scholar] [CrossRef]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Luo, L.; Zhao, Z.; Gong, C.; Haffari, G.; Pan, S. Graph-constrained Reasoning: Faithful Reasoning on Knowledge Graphs with Large Language Models. arXiv 2024, arXiv:2410.13080. [Google Scholar]
- Mondal, D.; Modi, S.; Panda, S.; Singh, R.; Rao, G.S. KAM-CoT: Knowledge Augmented Multimodal Chain-of-Thoughts Reasoning. Proc. AAAI Conf. Artif. Intell. 2024, 38, 18798–18806. [Google Scholar] [CrossRef]
- Khan, M.J.; Breslin, J.G.; Curry, E. NeuSyRE: Neuro-Symbolic Visual Understanding and Reasoning Framework based on Scene Graph Enrichment. Semant. Web 2023, 15, 1389–1413. [Google Scholar] [CrossRef]
- Verma, A.; Murali, V.; Singh, R.; Kohli, P.; Chaudhuri, S. Programmatically Interpretable Reinforcement Learning. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 5045–5054. [Google Scholar]
- Kimura, D.; Chaudhury, S.; Ono, M.; Tatsubori, M.; Agravante, D.J.; Munawar, A.; Wachi, A.; Kohita, R.; Gray, A. LOA: Logical Optimal Actions for Text-based Interaction Games. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations, Online, 1–6 August 2021; pp. 227–231. [Google Scholar]
- Delfosse, Q.; Shindo, H.; Dhami, D.S.; Kersting, K. Interpretable and Explainable Logical Policies via Neurally Guided Symbolic Abstraction. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
- Yang, M.; Liu, F.; Chen, Z.; Shen, X.; Hao, J.; Wang, J. CausalVAE: Disentangled Representation Learning via Neural Structural Causal Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 9593–9602. [Google Scholar]
- Wang, L.; He, Z.; Dang, R.; Shen, M.; Liu, C.; Chen, Q. Vision-and-Language Navigation via Causal Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024. [Google Scholar]
- Liang, W.; Kekić, A.; von Kügelgen, J.; Buchholz, S.; Besserve, M.; Gresele, L.; Schölkopf, B. Causal Component Analysis. In Proceedings of the Thirty-seventh Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
- Tafjord, O.; Dalvi, B.; Clark, P. ProofWriter: Generating Implications, Proofs, and Abductive Statements over Natural Language. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, 1–6 August 2021; Zong, C., Xia, F., Li, W., Navigli, R., Eds.; pp. 3621–3634. [Google Scholar] [CrossRef]
- Han, S.; Schoelkopf, H.; Zhao, Y.; Qi, Z.; Riddell, M.; Zhou, W.; Coady, J.; Peng, D.; Qiao, Y.; Benson, L.; et al. FOLIO: Natural Language Reasoning with First-Order Logic. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Miami, FL, USA, 12–16 November 2024; Al-Onaizan, Y., Bansal, M., Chen, Y.N., Eds.; pp. 22017–22031. [Google Scholar] [CrossRef]
- Talmor, A.; Herzig, J.; Lourie, N.; Berant, J. CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; Burstein, J., Doran, C., Solorio, T., Eds.; pp. 4149–4158. [Google Scholar] [CrossRef]
- Huang, L.; Le Bras, R.; Bhagavatula, C.; Choi, Y. Cosmos QA: Machine Reading Comprehension with Contextual Commonsense Reasoning. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Inui, K., Jiang, J., Ng, V., Wan, X., Eds.; pp. 2391–2401. [Google Scholar] [CrossRef]
- Mihaylov, T.; Clark, P.; Khot, T.; Sabharwal, A. Can a Suit of Armor Conduct Electricity? A New Dataset for Open Book Question Answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J., Eds.; pp. 2381–2391. [Google Scholar] [CrossRef]
- Bhagavatula, C.; Bras, R.L.; Malaviya, C.; Sakaguchi, K.; Holtzman, A.; Rashkin, H.; Downey, D.; Yih, S.W.t.; Choi, Y. Abductive commonsense reasoning. arXiv 2019, arXiv:1908.05739. [Google Scholar]
- Du, L.; Ding, X.; Liu, T.; Qin, B. Learning Event Graph Knowledge for Abductive Reasoning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021. [Google Scholar] [CrossRef]
- Bondarenko, A.; Wolska, M.; Heindorf, S.; Blübaum, L.; Ngonga Ngomo, A.C.; Stein, B.; Braslavski, P.; Hagen, M.; Potthast, M. CausalQA: A Benchmark for Causal Question Answering. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; Calzolari, N., Huang, C.R., Kim, H., Pustejovsky, J., Wanner, L., Choi, K.S., Ryu, P.M., Chen, H.H., Donatelli, L., Ji, H., et al., Eds.; pp. 3296–3308. [Google Scholar]
- Zhang, M.; Choi, E. SituatedQA: Incorporating Extra-Linguistic Contexts into QA. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; Moens, M.F., Huang, X., Specia, L., Yih, S.W.t., Eds.; pp. 7371–7387. [Google Scholar] [CrossRef]
- Camburu, O.M.; Rocktäschel, T.; Lukasiewicz, T.; Blunsom, P. e-SNLI: Natural Language Inference with Natural Language Explanations. In Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
- Dalvi, B.; Jansen, P.; Tafjord, O.; Xie, Z.; Smith, H.; Pipatanangkura, L.; Clark, P. Explaining Answers with Entailment Trees. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; Moens, M.F., Huang, X., Specia, L., Yih, S.W.t., Eds.; pp. 7358–7370. [Google Scholar] [CrossRef]
- Estermann, B.; Lanzendörfer, L.A.; Niedermayr, Y.; Wattenhofer, R. PUZZLES: A Benchmark for Neural Algorithmic Reasoning. In Advances in Neural Information Processing Systems; Globerson, A., Mackey, L., Belgrave, D., Fan, A., Paquet, U., Tomczak, J., Zhang, C., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2024; Volume 37, pp. 127059–127098. [Google Scholar]
- Bortolotti, S.; Marconato, E.; Carraro, T.; Morettin, P.; van Krieken, E.; Vergari, A.; Teso, S.; Passerini, A. A Neuro-Symbolic Benchmark Suite for Concept Quality and Reasoning Shortcuts. In Proceedings of the Thirty-Eighth Annual Conference on Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 9–15 December 2024; pp. 1–44. [Google Scholar]
- Mirchandani, R.; Sundar, K.; Mohamed, S.; Krueger, D. SCERL: A Text-based Safety Benchmark for Reinforcement Learning Problems. In Proceedings of the NeurIPS Datasets and Benchmarks Track, Virtual, 28 November 2022. [Google Scholar]
- James, S.; Ma, Z.; Rovick Arrojo, D.; Davison, A.J. RLBench: The Robot Learning Benchmark & Learning Environment. IEEE Robot. Autom. Lett. 2020, 5, 3019–3026. [Google Scholar]
- Hudson, D.A.; Manning, C.D. Gqa: A new dataset for real-world visual reasoning and compositional question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6700–6709. [Google Scholar]
- Park, D.H.; Hendricks, L.A.; Akata, Z.; Rohrbach, A.; Schiele, B.; Darrell, T.; Rohrbach, M. Multimodal Explanations: Justifying Decisions and Pointing to the Evidence. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Yi, K.; Gan, C.; Li, Y.; Kohli, P.; Wu, J.; Torralba, A.; Tenenbaum, J.B. Clevrer: Collision events for video representation and reasoning. arXiv 2019, arXiv:1910.01442. [Google Scholar]
- Xiao, J.; Shang, X.; Yao, A.; Chua, T.S. Next-qa: Next phase of question-answering to explaining temporal actions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9777–9786. [Google Scholar]
- Zawalski, M.; Chen, W.; Pertsch, K.; Mees, O.; Finn, C.; Levine, S. Robotic Control via Embodied Chain-of-Thought Reasoning. arXiv 2024, arXiv:2407.08693. [Google Scholar]
- Huang, W.; Xia, F.; Xiao, T.; Chan, H.; Liang, J.; Florence, P.; Zeng, A.; Tompson, J.J.R.; Mordatch, I.; Chebotar, Y.; et al. InnerMonologue: Embodied Reasoning through Planning with Language Models. In Proceedings of the Conference on Robot Learning (CoRL), Auckland, New Zealand, 14–18 December 2022. [Google Scholar]
- Yu, T.; Zhang, R.; Yang, K.; Yasunaga, M.; Wang, D.; Li, Z.; Ma, J.; Li, I.; Yao, Q.; Roman, S.; et al. Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task. arXiv 2018, arXiv:1809.08887. [Google Scholar]
- Price, P.; Fisher, W.M.; Bernstein, J.; Pallett, D.S. The DARPA 1000-word resource management database for continuous speech recognition. In Proceedings of the ICASSP-88, International Conference on Acoustics, Speech, and Signal Processing, New York, NY, USA, 11–14 April 1988; IEEE Computer Society: New York, NY, USA, 1988; pp. 651–652. [Google Scholar]
- Zhang, Y.; Deriu, J.; Katsogiannis-Meimarakis, G.; Kosten, C.; Koutrika, G.; Stockinger, K. ScienceBenchmark: A Complex Real-World Benchmark for Evaluating Natural Language to SQL Systems. In Proceedings of the VLDB Endowment, Vancouver, BC, Canada, 28 August–1 September 2023. [Google Scholar]
- Chen, B.; Zhang, F.; Nguyen, A.; Zan, D.; Lin, Z.; Lou, J.G.; Chen, W. CodeT: Code Generation with Generated Tests. In Proceedings of the Eleventh International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
- Lin, X.V.; Wang, C.; Zettlemoyer, L.; Ernst, M.D. NL2Bash: A Corpus and Semantic Parser for Natural Language Interface to the Linux Operating System. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018), Miyazaki, Japan, 7–12 May 2018; Calzolari, N., Choukri, K., Cieri, C., Declerck, T., Goggi, S., Hasida, K., Isahara, H., Maegaard, B., Mariani, J., Mazo, H., et al., Eds.; [Google Scholar]
- Amini, A.; Gabriel, S.; Lin, P.; Chaturvedi, S.; Farhadi, A.; Hajishirzi, H. MathQA: Towards Interpretable Math Word Problem Solving with Operation-Based Formalisms. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar]
- Zhou, S.; Xu, F.F.; Zhu, H.; Zhou, X.; Lo, R.; Sridhar, A.; Cheng, X.; Ou, T.; Bisk, Y.; Fried, D.; et al. Webarena: A realistic web environment for building autonomous agents. arXiv 2023, arXiv:2307.13854. [Google Scholar]
- Oh, J.H.; Kadowaki, K.; Kloetzer, J.; Iida, R.; Torisawa, K. Open-Domain Why-Question Answering with Adversarial Learning to Encode Answer Texts. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; Korhonen, A., Traum, D., Màrquez, L., Eds.; pp. 4227–4237. [Google Scholar] [CrossRef]
- Xu, Q.; Hong, F.; Li, B.; Hu, C.; Chen, Z.; Zhang, J. On the Tool Manipulation Capability of Open-source Large Language Models. arXiv 2023, arXiv:2305.16504. [Google Scholar]
- Nakano, R.; Hilton, J.; Balaji, S.; Wu, J.; Ouyang, L.; Kim, C.; Hesse, C.; Jain, S.; Kosaraju, V.; Saunders, W.; et al. Webgpt: Browser-assisted question-answering with human feedback. arXiv 2021, arXiv:2112.09332. [Google Scholar]
- Suhr, A.; Zhou, S.; Zhang, A.; Zhang, I.; Bai, H.; Artzi, Y. A corpus for reasoning about natural language grounded in photographs. arXiv 2018, arXiv:1811.00491. [Google Scholar]
- Thrush, T.; Jiang, R.; Bartolo, M.; Singh, A.; Williams, A.; Kiela, D.; Ross, C. Winoground: Probing vision and language models for visio-linguistic compositionality. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5238–5248. [Google Scholar]
- Zhong, V.; Xiong, C.; Socher, R. Seq2SQL: Generating structured queries from natural language using reinforcement learning. arXiv 2017, arXiv:1709.00103. [Google Scholar]
- Yu, T.; Zhang, R.; Er, H.Y.; Li, S.; Xue, E.; Pang, B.; Lin, X.V.; Tan, Y.C.; Shi, T.; Li, Z.; et al. Cosql: A conversational text-to-sql challenge towards cross-domain natural language interfaces to databases. arXiv 2019, arXiv:1909.05378. [Google Scholar]
- Dries, A.; Kimmig, A.; Meert, W.; Renkens, J.; Van den Broeck, G.; Vlasselaer, J.; De Raedt, L. ProbLog2: Probabilistic Logic Programming. In Machine Learning and Knowledge Discovery in Databases; Bifet, A., May, M., Zadrozny, B., Gavalda, R., Pedreschi, D., Bonchi, F., Cardoso, J., Spiliopoulou, M., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 312–315. [Google Scholar]
- Gebser, M.; Kaminski, R.; Kaufmann, B.; Schaub, T. clingo = ASP + Control: Preliminary Report. arXiv 2014, arXiv:1405.3694. [Google Scholar]
- Leone, N.; Pfeifer, G.; Faber, W.; Eiter, T.; Gottlob, G.; Perri, S.; Scarcello, F. The DLV system for knowledge representation and reasoning. ACM Trans. Comput. Log. (TOCL) 2006, 7, 499–562. [Google Scholar] [CrossRef]
- Byrd, W.E.; Holk, E.; Friedman, D.P. miniKanren, live and untagged: Quine generation via relational interpreters (programming pearl). In Proceedings of the 2012 Annual Workshop on Scheme and Functional Programming, Copenhagen, Denmark, 9–15 September 2012; pp. 8–29. [Google Scholar]
- Manandhar, S.; Džeroski, S.; Erjavec, T. Learning multilingual morphology with CLOG. In International Conference on Inductive Logic Programming; Springer: Berlin/Heidelberg, Germany, 1998; pp. 135–144. [Google Scholar]
- Gebser, M.; Kaminski, R.; Kaufmann, B.; Schaub, T. Answer Set Solving in Practice; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
Methodological Lineage | Time of Emergence | Core Idea | Key Challenges | Representative Works/ Systems |
---|---|---|---|---|
Logicism (Formal Logic) | 1950s–1960s | Deductive reasoning via formal logic | Noisy data, scalability | GPS [28], Resolution Theorem Proving, Prolog [27], VL1 [29] |
Expert Systems | 1970s–1980s | Rule-based knowledge representation | Brittleness, knowledge acquisition bottleneck | MYCIN [11], DENDRAL [30], CLIPS [31], OPS5 [32] |
Non-monotonic Logic | 1980s | Reasoning with defaults and retractable beliefs | Computational complexity | Default Logic [33], Circumscription [34], ASP [35], TMS [36] |
Planning Systems | 1970s | Action sequences via logical state transitions | State-space explosion | STRIPS [37], Shakey [38] |
Argumentation | Late 1980s–1990s | Reasoning via structured debates and conflicts | Semantics ambiguity, expressivity limits | Dung’s AF [39], ASPIC+ [40] |
Semantic Networks and DL | 1970s–1990s | Hierarchical concept representation and ontology | Limited expressivity, integration with logic | KL-ONE [41], Description Logic [42], OWL [43] |
Modal and Temporal Logic | 1960s–1980s | Reasoning with modalities and temporal operators | Integration with learning, decidability | Epistemic Logic [44], LTL, and CTL [45,46] |
Methodological Lineage | Time of Emergence | Core Idea | Key Challenges | Representative Works/Systems |
---|---|---|---|---|
Probabilistic Graphical Models | 1990s | Inference over uncertain variables using graphs | Structure design, scalability | Bayesian Networks [47], Markov Networks [48] |
Markov Logic Networks | Mid-2000s | Softening logical rules with probabilistic weights | Inference complexity, large-scale training | MLNs [49], Alchemy [66] |
Probabilistic Logic Programming | Early 2000s | Extending logic programming with probabilistic facts | Semantic complexity, grounding issues | ProbLog [50], PRISM [51], LPAD [52] |
Statistical Relational Learning | 2000s | Learning statistical patterns in relational domains | Overfitting, relational sparsity | Relational Bayes Nets [53], Relational Markov Networks [54], Tuffy [67], ILP [55] |
Causal Inference and SCMs | 1990s–2000s | Modeling interventions and counterfactuals | Assumptions, identifiability | SCMs [7], do-calculus [59] |
Deep Neural Networks (DNNs) | 2006 | End-to-end learning for hierarchical abstraction | Lack of logic structure, poor interpretability | CNN [63], RNN [64], GNN [65], Transformer [14], AlphaGo [15] |
Kernel and Ensemble Methods | 1990s–2010s | Structured prediction using statistical learning | Feature engineering, limited abstraction | SVM [60], Decision Trees [61], Random Forests [62] |
Methodological Lineage | Time of Emergence | Core Idea | Key Challenges | Representative Works/Systems |
---|---|---|---|---|
Differentiable Logic Programming | 2016 | Learnable logical rules via gradients | Optimization stability, scalability | ∂ILP [19], Logical Tensor Networks [69], Neural Logic Machines [68] |
Abductive Learning | 2019– | Combining symbolic abduction with neural perception | Symbol grounding, search complexity | Abductive Learning Framework [70], ABL-KG [84] |
Program Induction and NS-Concept Learners | 2016– | Learning symbolic programs and structures | Compositionality, sample efficiency | NS-CL [22], Neural Module Networks [71], CLEVR-CoGenT [72] |
LLM-based Reasoning and Tool Use | 2022– | Prompt-driven symbolic planning via LLMs | Verifiability, hallucination, tool coverage | ReAct [75], Toolformer [76], DSPy [77] |
Logic-aware Transformers | 2021– | Structural priors in Transformer architectures | Complexity, generalization | LogicT5 [85], Logical Transformers [78] |
Knowledge-augmented Reasoning | 2019– | Injecting KG and symbolic structure into models | Alignment, representation mismatch | DeepProbLog [80], NeurASP [81], K-BERT [20] |
Multimodal Neuro-symbolic Reasoning | 2020– | Reasoning over visual and multimodal inputs | Visual grounding, temporal logic | NS-CL [22], CLEVR-CoGenT [72], VideoCoT [82], ViperGPT [83] |
Task Structure | Symbolic Reasoning | Statistical Reasoning | Neural Reasoning | Neural–Symbolic Reasoning |
---|---|---|---|---|
Deductive Reasoning | C: Theorem Provers [88], Datalog [94] O: Ontology-based QA [107], Legal Inference [95] | C: Probabilistic Rules (MLN) [49] O: MLN for OIE QA [49] | C: Logic-aware Transformers [85] O: Chain-of-thought LLMs [92] | C: NeurASP [81], Neuro-symbolic Planners [108] O: Toolformer [76], DSP-Logic [77] |
Inductive Reasoning | C: ILP [98], FOIL [98] O: Meta-ILP [99], Meta-Interpretive Learning [109] | C: Bayesian Nets [47], ProbLog [50] O: Probabilistic Program Synthesis [110] | C: NLM [68], TreeLSTM Learners [111] O: GPT Concept Learners [112] | C: NS-CL [22], Neural Logic Machines [68] O: SceneGraph Reasoning [113] |
Abductive Reasoning | C: Logic Diagnosis [114], ABox Completion [95] O: Plan Recognition [115], Commonsense Inference [70] | C: Probabilistic Causal Models [7] O: PSL for Temporal Explanations [89] | C: Causal LLMs, Plan Explanation [116] O: Emergent causal LLMs [93] | C: Abductive Learning [70], Rule Induction [80] O: DSPy [77], Script Induction + Symbol Decoders [117] |
Analogical Reasoning | C: Structure Mapping Engines [100] O: Metaphor Resolution Systems [118] | C: Similarity Search over Graphs [119] O: Retrieval-aug. Analogical QA [120] | C: CLIP-based Visual Analogy [101] O: GPT-4 Analogical QA [121], Visual CoT [82] | C: Neural Module Networks (NMN) [71] O: Tool-enhanced Analogy QA [76] |
Domain | Dataset / Benchmark | Focus / Highlights |
---|---|---|
Question Answering | ProofWriter [189], FOLIO [190] | Deductive reasoning with formal logic entailment |
CSQA [191], CosmosQA [192], OBQA [193] | Commonsense reasoning with background knowledge | |
AbductiveNLI [194], ART [195] | Hypothesis selection based on plausible explanation | |
WhyQA [217], CausalQA [196] | Cause–effect inference and causal trace evaluation | |
Symbolic Reasoning | ToolBench [218], WebArena [216], AgentBench [166] | LLM-based reasoning with API tools and task orchestration |
HotPotQA [104], WebGPT [219] | Multi-hop tool-guided question answering | |
Multimodal Reasoning and Perception | CLEVR [72], GQA [204], VQA-X [205] | Visual relational and compositional reasoning |
CLEVRER [206], NExT-QA [207] | Temporal reasoning and event-based causal inference | |
NLVR2 [220], Winoground [221] | Visual-linguistic grounding and referential ambiguity | |
Program Induction and Semantic Parsing | Spider [210], ATIS [211], NL2SQL [222] | Mapping questions to executable SQL/logical forms |
NL2Bash [214], MathQA [215], CODET [213] | Program synthesis and symbolic reasoning from examples | |
CoSQL [223] | Conversational semantic parsing with symbolic schema linking |
Toolkit / Library | Paradigm | Modality | Key Capabilities | Example Use Cases |
---|---|---|---|---|
ProbLog [50], ProbLog2 [224] | Probabilistic Logic Programming | Symbolic | Probabilistic inference over logic programs | Knowledge base reasoning, uncertain facts |
DeepProbLog [80], NeurASP [81] | Neural–Symbolic Integration | Symbolic + Perception | Combine logic rules with neural outputs | VQA, handwritten digit inference |
LTN [69], LNN [136] | Differentiable Logic | Symbolic | Real-valued logic via fuzzy operators | Logic-constraint learning, semantic consistency |
AlphaILP [140], NEUMANN [133] | Neural ILP | Symbolic + Visual | Learning rules over graph/scene structure | Visual scene reasoning, object relation induction |
DSPy [77], LangChain [169], AgentBench [166] | Tool-Augmented LLMs | Language + API Tools | Modular agent control with tools + LLMs | Autonomous planning, reasoning via tool calls |
Alchemy [66], PSL [89] | Statistical Relational Learning | Symbolic | Soft logic + scalable inference | Entity resolution, joint inference tasks |
ASP Tools (clingo [225], DLV [226]) | Answer Set Programming | Symbolic | Non-monotonic rule-based reasoning | Diagnosis, planning, combinatorial search |
MiniKanren [227], CLOG [228] | Functional Logic Programming | Symbolic | Symbolic search with functional constructs | Program synthesis, theorem proving |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liang, B.; Wang, Y.; Tong, C. AI Reasoning in Deep Learning Era: From Symbolic AI to Neural–Symbolic AI. Mathematics 2025, 13, 1707. https://doi.org/10.3390/math13111707
Liang B, Wang Y, Tong C. AI Reasoning in Deep Learning Era: From Symbolic AI to Neural–Symbolic AI. Mathematics. 2025; 13(11):1707. https://doi.org/10.3390/math13111707
Chicago/Turabian StyleLiang, Baoyu, Yuchen Wang, and Chao Tong. 2025. "AI Reasoning in Deep Learning Era: From Symbolic AI to Neural–Symbolic AI" Mathematics 13, no. 11: 1707. https://doi.org/10.3390/math13111707
APA StyleLiang, B., Wang, Y., & Tong, C. (2025). AI Reasoning in Deep Learning Era: From Symbolic AI to Neural–Symbolic AI. Mathematics, 13(11), 1707. https://doi.org/10.3390/math13111707