1. Introduction
Why still work on ontologies while generative artificial intelligence (AI), in particular large language models (LLM) today allow solving problems for which ontologies have been used traditionally? The reason is that there are still numerous use cases, also in industry, in which the explicit specification of concepts and their relationships are most useful. Ontologies [
1] serve exactly this purpose. Many ontologies have been developed for the life sciences domain, e.g., Gene Ontology (GO) (
https://geneontology.org), SNOMED CT (
https://www.snomed.org) and MeSH (Medical Subject Headings) (
https://meshb.nlm.nih.gov) (all weblinks accessed on 6 January 2026). However, in the AI domain and the field of ML in particular, there are only a handful of ontologies available (e.g., [
2,
3,
4,
5]). In this article, an ontology for the ML field is presented which is named
ML Ontology. ML Ontology is based on existing work and expands on the following aspects:
- 1.
Volume: ML Ontology, together with OntoDM, ranks among the most comprehensive ontologies for the domain of ML with respect to the number of ML concepts and relationships modelled.
- 2.
Performance: ML Ontology allows high-performance query access suitable for industry applications.
- 3.
Balance between simplicity and expressiveness: ML Ontology is based on simple technology particularly suited for industry use, while at the same time offers expressive power, including sophisticated reasoning where needed.
- 4.
Extensibility and adaptability: ML Ontology is modularised and can be easily extended and adapted to differing use cases.
- 5.
Built-in quality management: ML Ontology comprises built-in quality checks that can be adapted use-case-specifically in order to ensure quality requirements from industry.
- 6.
- 7.
Many of those properties are crucial for using ontologies in industry applications, including sufficient volume, performance, expressiveness, extensibility and adaptability, quality and conformance to standards. In this article, the term industry-ready is used for ontologies that exhibit those properties.
ML Ontology has been developed over a period of more than five years by more than 15 different knowledge engineers with knowledge in the ML field. In the current version it consists of ca. 5000 RDF triples. The following two use cases use ML Ontology, and demonstrate its industry readiness:
- 1.
Configuration wizard: ML Ontology serves as the information backbone for a data science platform. Its usage in the configuration wizard for expert configurations of ML training runs is demonstrated.
- 2.
Interactive help: ML Ontology is used as the information source for the interactive help system of the data science platform, helping users understand complex terminology in machine learning.
The main contribution of this article is an industry-ready ontology for the ML field, which expands on previous work with built-in quality management and a balance between simplicity and expressiveness, leading to extensibility, adaptability and high-performance access.
The remainder of this article is structured as follows.
Section 2 references related work.
Section 3 explains the methodology used for developing ML Ontology.
Section 4 is the core of this article and outlines the architecture of ML Ontology. In
Section 5, two use cases are introduced which demonstrate successful application of ML ontology.
Section 6 discusses findings and
Section 7 concludes this article.
3. Methodology
The construction of ML Ontology was carried out following the METHONTOLOGY framework [
12,
13], which is a well-established methodology for ontology development that provides a structured approach to the specification, conceptualization, formalization, and implementation of ontologies. The process ensures methodological rigor and traceability across all stages of development.
3.1. Specification
The initial step involves defining the purpose, scope, and intended use of the ontology. ML Ontology was designed to provide common vocabulary for ML concepts and their relationships to support applications that integrate different ML tools and frameworks. Competence questions were formulated to guide the ontology’s coverage and ensure alignment with user requirements. Examples include:
What are the main categories of machine learning algorithms (e.g., supervised, unsupervised, reinforcement)?
Which ML approaches are suitable for classification tasks (e.g., decision trees)?
What are the common evaluation metrics used for regression models (e.g., mean squared error)?
Which ML libraries support regression (e.g., scikit-learn)?
Domain experts from academia and industry were consulted in form of workshops to validate the scope and relevance.
3.2. Conceptualization and Formalization
In this phase, the knowledge gathered during specification was structured into a conceptual model. Key concepts, relationships, and attributes relevant to machine learning were identified. This included classes such as ML area, ML task, ML approach, Metric, ML library etc., along with their properties and interrelations. A glossary of terms was created to ensure consistency and avoid ambiguity. Hierarchical taxonomies and semantic relations were drafted to represent the domain knowledge. A shared spreadsheet was used to collaboratively define and refine the conceptual model.
As a step towards formalization, the modeling approach was defined. To ensure industry readiness, ML ontology uses a lightweight modelling approach and is implemented as a knowledge graph. The modelling language used is RDF/RDFS. ML Ontology includes limited number of RDFS classes (ca. 10) and a large number of instances (ca. 700) representing ML concepts. Some SKOS properties like skos:prefLabel, skos:altLabel or skos:broader are used for attributes and relationships. SPARQL is used as query language and SPARQL Update for reasoning; SHACL is used for quality assurance.
3.3. Implementation and Evaluation
ML Ontology was implemented manually in a series of iterative cycles by ML experts and knowledge engineers. Naming conventions on IRIs were defined to ensure consistency. They are based on preferred labels of instances, e.g.,
support_vector_machine for the instance with preferred label “support vector machine”. A toolchain including shared spreadsheets, an RDF generator (
https://github.com/hochschule-darmstadt/RDFGenerator) and a GitHub repository for version control was used.
Evaluation and quality assurance were performed in various ways. As part of a committer workflow, a single reviewer with domain expertise reviewed all changes before merging them into the main branch. In addition, automated quality assurance checks were implemented as part of the ontology and executed regularly. Finally, competency questions defined during specification were used to validate the ontology’s coverage and correctness.
4. Results
4.1. ML Ontology Schema
Classes: Definition of all classes and properties to be used in the ontology;
Instances: Instances of the specified classes with all their properties that model ML concepts and their relationships;
Inference rules: Forward-chaining inference rules may be applied in a use-case-specific manner to enhance the expressiveness of the ontology;
QA: Quality assurance (QA) checks can be used in a use-case-specific manner to ensure integrity of the ontology;
Queries: Sample queries that may be used in applications utilising the ontology.
The main classes of ML Ontology are as follows.
ML_area with instances for supervised learning, unsupervised learning and reinforcement learning;
ML_task with ca. 40 instances, e.g., for classification, regression, and clustering;
ML_approach with ca. 120 instances, e.g., for artificial neural network, random forest, and support vector machine;
Preprocessing_approach with ca. 10 instances, e.g., for principal component analysis and kernel approximation;
Metric with ca. 120 instances, e.g., for accuracy, f-measure and mean squared error;
ML_library with ca. 30 instances, e.g., for Scikit-learn, TensorFlow, and PyTorch;
AutoML_solution with ca. 30 instances, e.g., for AutoSklearn, TPOT, and Auto-PyTorch;
Configuration_item with ca. 240 instances, e.g., for ensemble size, use approach, and metric.
In addition, there are ca. 15 predefined sample queries (e.g., get all dataset types), 5 quality assurance checks (e.g., check unique class) and one inference rule (broader inheritance). Classes and instances are explained in more detail in the following sections.
4.2. Ontology Classes
Figure 2 shows the main classes of ML Ontology as a UML diagram. The classes are organised in two modules:
ML concepts formalise terminology in the ML domain: concepts and their interdependencies;
ML implementations specify ML libraries, AutoML solutions and their concrete configuration options.
Instances of all classes are specified by a set of common properties:
skos:prefLabel: the preferred label of the individual, e.g., “support vector machine”. This label is used for the IRI of the individual, e.g., support_vector_machine;
skos:altLabel: alternative labels or acronyms, e.g., “SVM”;
rdfs:comment: short description of the individual, e.g., “Set of methods for supervised statistical learning”;
In addition, there are class-specific properties and properties that link instances of difference classes:
used_for links ML approaches, preprocessing approaches and metrics to ML tasks they can be used for, e.g., the ML approach support vector machine can be used for ML tasks classification and regression.
belongs_to links ML tasks to ML areas, e.g., the ML task classification belongs to the ML area supervised learning.
can_perform links instances of ML library or AutoML solution to ML tasks, e.g., ML library scikit-learn can perform ML task classification.
ml_task links configuration items to ML tasks.
skos:broader is used for modelling taxonomies between instances of the same class, e.g., the broader term of tabular regression is regression.
4.3. Ontology Alignment
Ontology alignment allows identifying correspondences between concepts in different ontologies to enable semantic interoperability across heterogeneous data sources. It reduces ambiguity, enhances reusability of knowledge, and allows applications to make use of linked data.
In ML Ontology, classes have been aligned with concepts from the ML Schema Core Specification (Namespace <https://www.w3.org/ns/mls#>) and instances with Wikidata entries via
rdfs:seeAlso. See
Table 1 and
Table 2.
Beyond name-level correspondences, we evaluate alignment with ML Schema along class intent, hierarchy organization, and data-level constraints:
ML_task vs. mls:Task: Both denote problem descriptions in ML. ML Schema constrains tasks with mls:definedOn. ML Ontology models dataset-specific variants via has_dataset_type and taxonomies using skos:broader.
ML_approach/Preprocessing_approach vs. mls:Algorithm: Both capture algorithmic families independent of implementation. ML Ontology distinguishes general ML approaches and preprocessing approaches; both can be projected to mls:Algorithm.
ML_library/AutoML_solution vs. mls:Implementation: ML Schema requires implementations to mls:implements an mls:Algorithm and to expose hyperparameters and quality characteristics. In ML Ontology, implementations are linked to tasks and approaches via can_perform and configuration relations (e.g., use_approach). Hyperparameters are represented as Configuration_item with enumerations and broader alignment terms.
Metric vs. mls:EvaluationMeasure: Both denote model assessment measures; ML Ontology’s Metric instances linked via used_for to tasks are comparable with mls:EvaluationMeasure.
4.4. Ontology Instances
Figure 3 shows an example subset of ontology instances, their attributes and relationships as a UML diagram. RDF triples with their subject–predicate–object structure are oriented at natural languages. A triple in Turtle notation like
can be read as “The ML task regression belongs to the ML area supervised learning.” In the following, the various ML concepts depicted in
Figure 3 are explained in order to illustrate the kind of knowledge being represented in ML Ontology. The areas explained are indicated with numbers in red circles, starting in the centre of the diagram.
(1) ML task tabular regression is a specialisation of the concept regression, hence the broader relationship. The specified dataset type is tabular.
tabular is an instance of the auxiliary class
Enum (not depicted in
Figure 2), which allows formalising categorised named vocabularies. In this case, the category is dataset type and instances are tabular, text, etc. Where there are Wikidata entries for respective concepts, there are links to respective entries, e.g., for regression the entry Q208042 (
https://www.wikidata.org/entity/Q208042 (accessed on 6 January 2026)).
(2) The ML task regression belongs to ML area supervised learning (acronym “SL”, Wikidata ID Q334384). Another ML task belonging to supervised learning is classification. The metric accuracy can be used for classification and the metric mean squared error can be used for regression.
(3) An ML task belonging to ML area unsupervised learning is feature reduction, just like the preprocessing approach principal component analysis (PCA).
(4) The ML approaches random forest and artificial neural network (ANN) can both be used for ML tasks classification and regression. The ML approach multilayer perceptron (MLP) is a specialisation of artificial neural network, hence the broader relationship.
(5) Auto-Sklearn is an AutoML solution that can be used for generating ML models for ML library Scikit-learn. Both can perform ML tasks classification and regression.
(6) We show exemplary a configuration item tabular regression in autosklearn. It configures multiple parameter values, one of which is use approach for configuring the ML approaches that Auto-Sklearn supports (here random forest and artificial neural network). The broader term is use approach, necessary for aligning the parameter configurations between various AutoML solutions.
4.5. Ontology Queries
Traditionally, ontologies and queries are separated. Ontologies are published, e.g., as open data, whereas SPARQL queries are developed as part of applications using the data. However, we argue that ontologies and queries belong together because predicates and classes used in queries depend on their definitions in the ontology. Therefore, sample SPARQL queries are published as part of ML Ontology, providing application developers wishing to access ML Ontology a starting point for their implementation.
SHACL is employed for storing SPARQL sample queries within ML Ontology. Consider Listing 1 with statements for specifying a SPARQL query which gets all ML tasks and the respective ML areas they belong to, e.g., classification belongs to supervised learning.
| Listing 1: Sample ontology query. |
![Applsci 16 00843 i001 Applsci 16 00843 i001]() |
The class sh:SPARQLExecutable (Namespace <http://www.w3.org/ns/shacl#>) is instantiated and used with the property sh:select for specifying the SPARQL query.
4.6. Quality Assurance
Like many other ontologies, ML Ontology has been curated by a number of different knowledge engineers over a longer period of time. As in software development, architectural integrity is an important aspect in ontology development. In software development, regression testing is a standard approach for ensuring quality over development time. We consider the same to be applicable for ontology development. To operationalize built-in quality management, ML Ontology ships with a small but expressive base profile of declarative constraints that are evaluated as SPARQL checks embedded in the ontology. Each check is defined as a SPARQL SELECT query stored alongside the data (using SHACL/SPARQL executable forms) and interpreted with the conventional pass criterion “no results returned”; any bindings produced by a query represent concrete violations that can be reported or fixed. This design keeps the constraints versioned with the ontology and ensures that validation semantics co-evolve with the schema and instances.
From a software engineering perspective, embedding the checks in the ontology rather than in application code provides the following benefits: (i) single source of truth for invariants across heterogeneous consumers; (ii) portability and reproducibility of validation independent of specific stacks or runtimes; (iii) tight coupling of constraints with ontology versions enabling traceable evolution; and (iv) ease of automation in CI pipelines and data release processes without duplicating logic in multiple applications.
The current base profile groups checks into typing, vocabulary, and schema conformance.
Table 3 summarizes the shipped checks and their intent.
Coverage. The typing checks ensure that all IRIs used as objects correspond to declared classes and that no individual is assigned multiple ML Ontology classes. The vocabulary check verifies that every predicate from the ontology namespace appearing in the data is explicitly declared as an rdf:Property. The schema-conformance check reports the observed subject and object classes for each predicate, enabling inspection of domain–range alignment and the detection of potential type inconsistencies. All checks are implemented as SPARQL SELECT queries and follow a “no results means compliance” convention.
Execution. In practice, the checks are executed automatically as part of ontology CI and release workflows and can be rerun on ingestion by consuming systems. Since each check is a pure query artifact stored with the ontology, validation is independent of a particular application and can be reproduced across RDF engines.
Example. Listing 2 shows a simple sample SPARQL query for check unique class.
| Listing 2: Sample check. |
![Applsci 16 00843 i002 Applsci 16 00843 i002]() |
The expected behaviour is that this query returns no results. However, if there are multiple class assignments for a subject, then those are listed in the result set. An example could be XGBoost which is the name of an ML approach as well as the name of an ML library. The query will return a row, if erroneously the same IRI had been used for both concepts, ML approach and ML class.
4.7. Reasoning
A lightweight approach for modelling ML Ontology has been chosen deliberately. However, some use cases may require sophisticated reasoning [
14]. For this reason, a lightweight yet powerful reasoning mechanism is employed in ML Ontology, which has been proposed in own previous work [
10].
Consider the following example depicted in
Figure 4. For ML tasks, various hierarchies have been specified. For example, classification may be refined according to the number of classes (binary classification, multiclass classification) or according to the dataset types (tabular classification, text classification, etc.).
skos:broader has been chosen deliberately as a predicate with rather weak semantics to allow for various kinds of hierarchical groupings. The root of this hierarchy, the ML task classification, is specified as belonging to the ML area supervised learning. Logically speaking, one can state that if an ML task belongs to some ML area then all specific variants of this ML task belong to the same ML area, e.g., text classification belongs to supervised learning since classification belongs to supervised learning. In order to foster completeness of ML Ontology, one could add a quality check, ensuring that for all ML tasks, an ML area is assigned which it belongs to, forcing knowledge engineers to manually model all belongs to relationships. However, this is not only tedious but also introduces redundancy that should be avoided.
Another approach, which has been chosen, is to introduce use-case-specific reasoning to add correct belongs_to relationships even when they have not been modelled explicitly. Consider Listing 3 with a declaration of the property belongs_to.
| Listing 3: Declaration of the property belongs_to. |
![Applsci 16 00843 i003 Applsci 16 00843 i003]() |
Like all properties used as predicates in ML Ontology, belongs_to is declared as an RDF Property. In addition, the meta-property inheritance specifies its inheritance behaviour as broader inherited. Inheritance is a mechanism adopted in computer science, in particular in object-oriented programming, for propagating properties across hierarchical structures, enabling derived entities to access characteristics of their predecessors. Listing 4 shows a use-case-specific rule that implements such inheritance behaviour for properties that have been declared as broader inherited.
| Listing 4: Example rule. |
![Applsci 16 00843 i004 Applsci 16 00843 i004]() |
An INSERT statement from the SPARQL Update specification is used which allows inserting new triples into an ontology. This allows implementing production rules that can be used for forward chaining reasoning. In this rule, triples are selected of which predicate ?p is broader inherited. Then, triples ?b ?p ?o are selected, for which the subject ?b is in a direct or indirect broader relationship with an individual ?s. If those conditions are satisfied then a new triple ?s ?p ?o is inserted into the ontology.
For example, if the query variable ?p is bound to the predicate belongs to and ?b is bound to classification and ?s is bound to text classification, then the following RDF triple is added to the ontology.
This inference rule is stored as part of ML Ontology using the SHACL class sh:SPARQLRule. If reasoning is needed in some use case context then all inference rules need to be executed after loading ML Ontology into an RDF store and before executing queries. After this, a query for ML tasks that belongs to supervised learning will also return text classification even if this is not explicitly specified in the ontology.
Note that other predicates used in ML task (e.g., skos:prefLabel, has_dataset_type, etc.) do not have the meta-property broader inherited and therefore those values are not inherited along the broader hierarchy. So, the label for classification will not be inherited to text classification.
Another property which has been specified as broader inherited is used_for which is used in various classes (e.g., ML approach, preprocessing approach, metric) for linking those concepts to ML tasks they can be used for. Take the example of the ML approach multilayer perceptron (MLP) which is a specialisation of artificial neural network (ANN). ANN can be used for classification and regression. Therefore, by applying the above inference rule, MLP can also be used for classification and regression even if this has not been modelled explicitly. Another example is the preprocessing approach kernel principal component analysis which is a specialisation of principal component analysis (PCA). PCA can be used for feature reduction. Therefore, by applying the inference rule, kernel principal component analysis can also be used for feature reduction.
This mechanism allows use-case-specific reasoning to be implemented transparently and easily. Ontology modelling is still lightweight while the full expressive power of SPARQL may be used for specifying complex inference rules. From an operational standpoint, inferred triples are materialized using SPARQL Update at ontology release time—i.e., whenever ML Ontology is updated and re-deployed as an information backbone of an application—by computing the closure into a dedicated inferred graph that is dropped and recomputed on each release. This batch rematerialization avoids per-triple synchronization and data bloat, constrains growth via selective rules and forward chaining. Batch rematerialization has been chosen due to its simplicity and low cost at the present scale [
15,
16,
17,
18].
6. Discussion
In this section, ML Ontology is compared with requirements specified in
Section 1.
6.1. Volume
ML Ontology is notably comprehensive, encompassing ca. 700 individuals that define key machine learning concepts. With roughly 5000 RDF triples, it ranks among the largest domain-specific ontologies for ML that we are aware of. The only ontology for the ML domain comparable in volume is OntoDM with ca. 660 classes representing ML concepts, described by ca. 3700 annotations (AnnotationAssertions, subClassOf relationships, AnnotationProperties, DisjointClasses etc.). The ML Schema Core Specification as an upper-level ontology only specifies 25 classes. According to [
4], DMOP includes ca. 720 classes representing ML concepts, described by ca. 4300 annotations (data properties, logical axioms etc.). However, the ontology does not seem to be publicly available, just like the Exposé ontology for data mining experiments. Finally, the MEX vocabulary comprises ca. 250 classes.
To situate ML Ontology relative to prior work,
Table 4 offers a compact comparison along scope focus, public availability, size, reasoning/validation support, and intended use.
To formally evaluate correctness and completeness of ML Ontology, a gold standard for ML terminology would be needed. However, to the best of our knowledge, no such gold standard exists. Therefore, an experiment has been performed to estimate correctness and completeness of ML Ontology. The experiment was performed using a state-of-the-art generative AI chatbot (MS Copilot with GPT-5). In this experiment, instances of the main ML Ontology classes were presented and correctness of the instances (false positives for evaluating precision) and missing instances (false negatives for evaluating recall) was prompted for.
Table 5 shows the results of this experiment.
The experiment has shown that ML Ontology has a high precision of 1.00, indicating that all instances included in the ontology are correct. This is not surprising since the ontology has been curated by domain experts. The recall is 0.71, suggesting that while the ontology covers a substantial portion of relevant instances, there are still some missing concepts. Samples are shown in
Table 5. To include those concepts in ML Ontology is subject to ongoing and future work. The overall F1 Score is 0.83.
6.2. Performance
A benchmark has been performed to evaluate the performance of SPARQL queries on ML Ontology. For this, 9 SPARQL queries of varying complexity have been executed on ML Ontology. Simple queries retrieve all instances of a specific class, medium queries join classes, and high-complexity queries contain aggregations, filters, or transitive closures. Each query has been executed 100 times with execution time and memory usage recorded.
Setup: In-memory SPARQL engine: rdflib 7.1.4, CPU: Intel Core Ultra 7 165H, 3.80 GHz, 32 GB RAM.
Table 6 shows the results of this benchmark including the number of results, execution time, and memory usage.
The table shows three sample queries for each complexity category including the number of results. The time column shows the medium execution time from 100 runs in milliseconds. The memory column shows the memory usage observed at query execution in megabytes. The execution times are all below 100 ms, even for complex queries, indicating good performance. Memory usage is also low, with a maximum of 3.35 MB observed for medium-complexity queries. Overall, the benchmark demonstrates that SPARQL queries on ML Ontology can be executed efficiently with low resource consumption, making it suitable for real-time applications.
6.3. Balance Between Simplicity and Expressiveness
ML Ontology deliberately uses lightweight modelling approach (knowledge graph implemented in RDF/RDFS) which is particularly suited for industry use. Reasoning in RDF-based systems is commonly implemented using (i) OWL reasoning under Description Logic (DL) semantics or (ii) SPARQL-based rule execution with forward materialization, optionally organized through SHACL rules. Both approaches support automated inference, but they differ fundamentally in semantics, expressiveness, performance characteristics, and operational behavior.
In SPARQL-based inferencing, rules are expressed as
CONSTRUCT or
INSERT queries that derive new triples and materialize them into the RDF graph. SHACL Rules and SHACL-SPARQL provide a standardized organizational layer for such rule sets [
24], defining rule targets, conditions, and metadata, while delegating execution to SPARQL query evaluation. The resulting materialization pipeline repeatedly applies rules until a fixpoint is reached, similar to rule engines such as VLog [
25] or WebPIE [
26]. This approach is operational, closed-world, and highly flexible, enabling arbitrary graph transformations.
OWL 2 DL provides a model-theoretic semantics based on expressive Description Logics [
27]. OWL reasoners support classification, satisfiability checking, and query answering under open-world, monotonic semantics. The language includes constructs such as existential and universal quantification, cardinality restrictions, and complex role inclusion axioms. OWL reasoning provides strong global guarantees but is computationally more demanding: full OWL 2 DL reasoning is ExpTime to N2ExpTime-complete [
27], and empirical studies show that even optimized reasoners struggle with large or expressive ontologies [
28,
29]. Query answering under OWL Direct Semantics requires specialized optimizations and still incurs substantial overhead compared to simple SPARQL evaluation [
30,
31].
We compare both approaches with respect to expressiveness, performance, and operational complexity. For an overview see
Table 7.
SPARQL UPDATE supports arbitrary insert/delete operations on RDF graphs. Since iterative application of such rules constitutes a graph-rewriting system, and graph-rewriting systems are known to be Turing-complete [
32], iterative SPARQL UPDATE is computationally Turing-complete. Practical systems already exploit this expressiveness for iterative graph algorithms [
33,
34]. This allows SPARQL-based inferencing to express transformations and procedural behaviors that lie beyond the expressive power of OWL 2 DL. Conversely, OWL provides model-theoretic constructs such as existential and universal quantification, cardinality restrictions, and complex role inclusion axioms [
27] that are not natively available in SPARQL rules unless explicitly simulated operationally. Thus, SPARQL-based inferencing is more expressive in a computational sense, whereas OWL is more expressive in a logical and model-theoretic sense.
SPARQL-based materialization typically exhibits polynomial-time data complexity and scales well in practice, since inference reduces to repeated SPARQL query evaluation [
35]. Once materialization is complete, query answering is extremely fast because it operates over an explicitly closed-world graph. In contrast, OWL reasoning is significantly more computationally demanding: full OWL 2 DL reasoning is ExpTime to N2ExpTime-complete [
27], and empirical studies show that even optimized reasoners struggle with large or expressive ontologies [
28,
29]. Query answering under OWL Direct Semantics requires specialized optimizations and still incurs substantial overhead compared to simple SPARQL evaluation [
30,
31].
SPARQL-based materialization is conceptually simple and easy to integrate into existing RDF pipelines. Rules are transparent, debuggable, and can be organized and documented using SHACL [
24]. OWL reasoning, while offering strong formal guarantees, requires more expertise in ontology engineering, careful modeling to avoid unintended inferences, and specialized tooling. Debugging and maintaining OWL ontologies in production environments is often more complex than managing SPARQL rule sets.
In total, SPARQL-based materialization with SHACL organization is a suitable choice for ML Ontology, balancing expressiveness, performance, and operational simplicity. It enables flexible, efficient inferencing tailored to practical application needs without the complexity and overhead of full OWL reasoning.
6.4. Extensibility and Adaptability
ML Ontology is modularised and can be easily extended and adapted to various use cases. The ontology schema provides a clear separation between classes, instances, inference rules, queries and QA checks (see
Figure 1). ML Ontology classes are separated in ML concepts and ML implementations (see
Figure 2). ML Ontology has been extended continuously over a period of more than five years and the ontology schema has proven stable.
ML Ontology is designed to be platform-agnostic and reusable across technological environments. It is serialized in RDF/Turtle and relies exclusively on Semantic Web standards (RDF/RDFS, SKOS, SHACL, SPARQL), ensuring independence from any specific runtime or storage system. The two application scenarios presented in this manuscript serve as implementation exemplars rather than architectural constraints; none of the classes, properties, or inference/QA mechanisms depend on features of a particular proprietary platform.
The module ML concepts provides a stable, domain-level vocabulary for machine learning independent of implementation concerns. Its classes (e.g., ML_area, ML_task) and relations (e.g., used_for, belongs_to, skos:broader) formalise shared concepts and their interdependencies. This module is intentionally generic and directly reusable in external contexts without modification.
The module ML implementations (e.g., classes ML_library, AutoML_solution) is also generic in its schema. The level of instance-level detail is purposefully scoped to concrete application needs. In the present release, configuration items are modelled more deeply for AutoML solutions because the motivating use cases focus on AutoML orchestration. If a deployment prioritises direct configuration of ML libraries instead, the ontology can be extended incrementally by adding new Configuration_item instances and enumerations, and by linking them to ML_library via existing relations (e.g., can_perform, ml_task, used_for), without changing the core schema.
To reuse and tailor ML Ontology in other environments, practitioners typically:
Extend instances: add domain-specific individuals for ML_task, ML_approach, Metric, and, when needed, new Configuration_item vocabularies for targeted libraries or frameworks.
Wire semantics: reuse existing predicates (used_for, belongs_to, can_perform, ml_task) to express applicability, capabilities, and configuration scopes; optionally refine taxonomies via skos:broader.
Maintain alignments: keep or extend published mappings to ML Schema (class-level) and Wikidata (instance-level) to facilitate integrations.
Leverage QA and reasoning: embedded SHACL/SPARQL checks and the optional forward-chaining rule are orthogonal to any platform and can be reused unchanged; projects may add deployment-specific checks or rules as needed.
Adapting the ontology to additional tools or ecosystems requires only the addition of instances and lightweight enumerations, while the class/property schema remains stable. This design preserves generality in the conceptual core, keeps the implementation module schema generic, and localises application-specific effort to instance additions and optional profiles, thereby supporting reuse beyond the proprietary platform showcased in the examples.
6.5. Built-In Quality Management
ML Ontology comprises built-in quality checks that can be adapted use-case-specifically in order to ensure that quality requirements from the industry are met. As opposed to just documenting modelling guidelines, quality checks can be executed regularly, allowing guideline violations to be detected and fixed. In practice, when the quality checks were implemented first, dozens of modelling errors were detected and fixed. Due to the awareness created by the quality checks, modelling errors have become rare. ML Ontology has be co-edited by more than 15 knowledge engineers over more than 5 years and still has not deteriorated in quality.
6.6. Standards
ML Ontology is solely based on Semantic Web (SW) standards including RDF/RDFS and SPARQL, and uses state-of-the-art ontology schemas such as SKOS and SHACL. Experiments have shown that ML Ontology can easily be deployed as a Labeled Property Graph (LPG) if needed. ML Ontology could be imported without adaptation into Neo4j and the pre-defined SPARQL queries could be automatically converted to correct Cypher queries.
6.7. Public Availability
In total, the criteria sufficient volume, performance, expressiveness, extensibility and adaptability, quality and conformance to standards are fulfilled. Therefore, ML ontology can be regarded as industry-ready.
6.8. Limitations
While ML Ontology has been designed to be performant, extensible, and adaptable, certain limitations should be acknowledged. ML is a fast-evolving field, and the ontology requires periodic updates to incorporate emerging concepts and techniques. This is particularly relevant for generative AI approaches, which are currently only partially covered. A semi-automated process for identifying missing concepts using generative AI has been outlined in
Section 6 and may be used for future updates. In addition, automated alignment tools (e.g., AML [
36]) may be used. Additionally, while the ontology has been tested in specific use cases, broader validation across diverse industry applications would further establish its robustness and versatility.
7. Conclusions and Future Work
In this article, ML Ontology has been presented, an industry-ready ontology for the ML domain. ML Ontology is comprehensive, provides good performance and is extensible and adaptable. In comparison to existing work, ML Ontology provides novel features, including built-in queries and quality assurance. Notably, ML Ontology provides sophisticated reasoning while being based on lightweight modelling languages (RDF/RDFS). Its industry-readiness has been demonstrated by benchmarks and use case implementations within a data science platform.
ML Ontology can be considered ready for use in industry applications requiring the formalisation of ML concepts and their relationships. As future work, we envisage the extension of ML Ontology as an incremental, use-case-driven process. Further applications accessing ML Ontology may define requirements for extensions, e.g., the parametrisation of ML libraries. As ML Ontology is open source, we regard this as a community effort.