Next Article in Journal
Optimal Cutoff Value of the Tumor Mutation Burden for Immune Checkpoint Inhibitors: A Lesson from 175 Pembrolizumab-Treated Cases Among 6403 Breast Cancer Patients
Previous Article in Journal
Hallux Strength Matters: Its Relationship with Physical Performance and Predictive Role in Fall Risk
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

KAVAI: A Systematic Review of the Building Blocks for Knowledge-Assisted Visual Analytics in Industrial Manufacturing

1
Josef Ressel Centre for Knowledge-Assisted Visual Analytics for Industrial Manufacturing Data, St. Pölten University of Applied Sciences, 3100 St. Pölten, Austria
2
Pro2Future GmbH, 8010 Graz, Austria
3
Institute of Visual Computing, Graz University of Technology, 8010 Graz, Austria
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this research.
Appl. Sci. 2025, 15(18), 10172; https://doi.org/10.3390/app151810172
Submission received: 12 August 2025 / Revised: 5 September 2025 / Accepted: 10 September 2025 / Published: 18 September 2025
(This article belongs to the Section Applied Industrial Technologies)

Abstract

Industry 4.0 produces large volumes of sensor and machine data, offering new possibilities for manufacturing analytics but also creating challenges in combining domain knowledge with visual analysis. We present a systematic review of 13 peer-reviewed knowledge-assisted visual analytics (KAVA) systems published between 2014 and 2024, following PRISMA guidelines for the identification, screening, and inclusion processes. The survey is organized around six predefined building blocks, namely, user group, industrial domain, visualization, knowledge, data and machine learning, with a specific emphasis on the integration of knowledge and visualization in the reviewed studies. We find that ontologies, taxonomies, rule sets, and knowledge graphs provide explicit representations of expert understanding, sometimes enriched with annotations and threshold specifications. These structures are stored in RDF or graph databases, relational tables, or flat files, though interoperability is limited, and post-design contributions are not always persisted. Explicit knowledge is visualized through standard and specialized techniques, including thresholds in time-series plots, annotated dashboards, node–link diagrams, customized machine views from ontologies, and 3D digital twins with expert-defined rules. Line graphs, bar charts, and scatterplots are the most frequently used chart types, often augmented with thresholds and annotations derived from explicit knowledge. Recurring challenges include fragmented storage, heterogeneous data and knowledge types, limited automation, inconsistent validation of user input, and scarce long-term evaluations. Addressing these gaps will be essential for developing adaptable, reusable KAVA systems for industrial analytics.

1. Introduction

The ongoing digitalization of industrial manufacturing—driven by the adoption of cyber–physical systems, industrial IoT, and automated control infrastructure—has led to a massive increase in the availability of machine-generated data [1]. While this development offers new opportunities for monitoring, diagnosing, and optimizing production systems, it also introduces significant challenges for analysis: industrial data are often high volume, heterogeneous, noisy, and context-dependent [2,3]. This paradigm, commonly termed smart manufacturing, leverages data-driven analytics to orchestrate production and supply chains in real time [4].
Visual analytics (VA), “the science of analytical reasoning facilitated by interactive visual interfaces” [5], recognizes that analysis is not a routine process that can be fully automated [6]. It relies on the expertise, initiative, and domain knowledge of analysts, supported by visual representations and computational methods. Interaction is central to this process, enabling flexible, iterative exploration [7]. VA helps uncover trends, patterns, and anomalies. However, for VA to be effective, it must deliver precise, relevant, and interpretable information [8] that is tailored to the user’s mental model and task context.
In many industrial settings, analysts are domain experts (e.g., engineers, operators, or managers) who may not be specialists in data analysis or visualization. This creates a gap: while VA provides powerful interactive methods, it often struggles to capture and integrate the contextual knowledge and reasoning processes of these experts. Stoiber et al. [9] highlight that domain experts need systems that go beyond displaying data. They must also provide guidance and incorporate knowledge to support decision-making in complex environments.
In response to this gap, the concept of knowledge-assisted visual analytics (KAVA) was introduced [10,11,12,13]. KAVA extends classical VA by systematically integrating explicit knowledge (extracted tacit knowledge in an structured form)—such as rules, ontologies, provenance information, or domain constraints—into interactive visualization and computational analysis. Explicit knowledge in this context refers to knowledge that can be formalized, stored, and reused, which distinguishes it from tacit knowledge that remains tied to human expertise. By externalizing tacit knowledge and embedding explicit knowledge into the analytical workflow, KAVA enables context-aware, explainable, and more reusable VA solutions [9,11,12].
Several studies have demonstrated that KAVA can improve interpretability, support the onboarding of new users, and guide analysis workflows across different domains, ranging from biomedical data analysis (e.g., Ordino and KAVAGait) to general-purpose visualization environments such as Tableau and Advizor [9]. However, despite these promising advances, the systematic study of KAVA in industrial manufacturing remains limited. Industrial contexts are characterized by heterogeneous data sources, domain-specific tasks, and high demands for reliability and interoperability—making them an ideal yet challenging testbed for KAVA approaches.
This motivated our survey, which aimed to outline the state of the art of KAVA in industrial manufacturing and explore how knowledge, data, visualization, and interaction are addressed within existing systems and where gaps or best practices can be identified. We aimed to obtain a comprehensive understanding of the conceptual and technical building blocks relevant to the domain.
Thus, we conducted a systematic literature review following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) methodology [14,15] to ensure a transparent, comprehensive, and structured review process. The following research questions served as the foundation for this review:
RQ1: 
How is explicit knowledge modeled? Which knowledge modeling approaches are commonly used for knowledge representation in industrial manufacturing VA?
RQ2: 
Which visualization and interaction methods are used for externalizing explicit knowledge in industrial manufacturing?
RQ3: 
How are ML methods integrated into KAVA in industrial manufacturing?
To systematically address these questions, we grounded our analysis in the conceptual foundation provided by the “knowledge-assisted visual analytics (KAVA) model”, as introduced by Federico et al. [11] and Wagner [12]. This model opens the possibility to frame KAVA as an iterative process of knowledge generation, representation, application, and refinement. It emphasizes that knowledge in VA systems is not static—it evolves through interaction, feedback, and usage. Based on this model, as displayed in Figure 1, we identified a set of functional building blocks, illustrated in Figure 2, that structure our literature review and allow us to examine KAVA systems along consistent and conceptually meaningful dimensions (see Section 4):
The industrial domain within industrial manufacturing (e.g., automotive and packaging production) shapes every component of the KAVA [11] model. It plays a cross-cutting role in, for example, guiding data selection, using visualization techniques and interaction concepts, formalizing explicit knowledge, automating analysis, and determining the way that users apply and externalize their tacit knowledge. The building blocks, summarized in Figure 2, include the following:
  • Data: The types of data (e.g., time-dependent data and tabular data) that are described in the examined studies.
  • Knowledge: The knowledge modeling storage mechanisms (e.g., text, ontologies, and taxonomies) and the process, type, and origin of knowledge gathered in this survey.
  • Visualization: Visualization and interaction techniques (e.g., line plots and node–link diagrams).
  • Machine learning: Descriptions of the used machine learning (ML) methods, e.g., supervised, unsupervised, and semi-supervised ML.
  • User group: The main user group (e.g., engineers and managers) interacting with the KAVA system, contributing tacit knowledge, and steering exploration.
Contributions: By analyzing the literature across these dimensions, we systematically provide a review that discusses and helps to understand the field by investigating existing KAVA methods, specifically focusing on how explicit human domain knowledge is modeled and visualized in other applications, uncovering gaps and offering insights for future research and applications in industrial manufacturing. To summarize, our contributions are as follows:
  • We provide a structured overview of research on industrial manufacturing and KAVA (see Section 2);
  • We present our categorizations, which are the main building blocks of KAVA (see Section 4);
  • We contextualize the major ongoing challenges in the domain of KAVA in industrial manufacturing (see Section 5).

2. Background and Related Work

This article combines three thematic areas of data analytics research: interactive visual interfaces through visual analytics (VA) [5], knowledge externalization and knowledge-based analysis [16], and data-driven approaches for industrial manufacturing [2].
Before Section 4 systematically investigates the overlap of all three areas, the related work in each pair-wise combination of thematically adjacent fields is explored below.
There is a growing body of literature applying VA to industrial manufacturing data. Four surveys describe this field. Zhou et al. [17] carried out a systematic survey in 2019 and provided a thorough review of the visualization technologies used to analyze data throughout the manufacturing lifecycle by categorizing them into two primary areas: “replacement,” which employs immersive tools like virtual reality (VR), augmented reality (AR), and mixed reality (MR) to replicate and digitally enhance physical processes, and “creation,” which focuses on using visualization to drive innovation in design, production, testing, and maintenance. The survey covers various sectors, including the automotive, energy, and food processing sectors, and it discusses challenges such as integrating diverse data types, ensuring data security, achieving real-time processing, and designing user-centered interfaces. Cibulski et al. [18] retrospectively analyzed and compared 13 visualization research projects in the Austrian and German manufacturing industry. Moreover, Schreck et al. [3] highlighted the role of VA in optimizing industrial processes through computational analysis and interactive visualization. Their paper provides an overview of goals, concepts, and existing solutions for industrial applications by discussing the main data types and challenges in data acquisition. Furthermore, challenges such as heterogeneity, data ownership, and governance are presented.
In contrast, Wang et al. [19] focused on visualization and VA methods for multimedia data in industrial settings. They introduced a taxonomy based on data types, categorizing research based on visualization techniques, interaction methods, and specific application areas. They highlighted the need for integrated frameworks to manage diverse data formats.
As the above surveys did not address the roles of knowledge and KAVA methods, we plan to extend their contributions by focusing on how experts’ domain knowledge can be integrated into VA to enhance effective decision-making in industrial manufacturing environments.
While VA approaches integrate industrial domain knowledge through user interaction, previous research has also directly integrated explicit knowledge into industrial analytics. In particular, knowledge graphs have established themselves in recent years as a central method for externalizing domain expertise in a structured, machine-readable format. In a recent review study, Xiao et al. [20] discussed the role of process knowledge graphs in manufacturing and their significance for transforming process knowledge and the knowledge of domain experts and engineers into formalized graph structures that enable, for example, computer-aided process planning. The authors emphasized that such graphs allow for the storage and reuse of knowledge and can also visualize implicit relationships between processes, which supports scalable knowledge externalization in industrial scenarios. Furthermore, Buchgeher et al. [21] systematically investigated the application of knowledge graphs in the production and manufacturing domain. They showed that knowledge fusion and semantic integration are common use cases. Nevertheless, the broader focus is often on externalizing heterogeneous domain knowledge into unified, structured representations. Although numerous technical contributions already exist, the study indicated that empirical validation and industrial implementation have so far been limited. This shows that knowledge externalization continues to tackle challenges in practice, particularly concerning usability and adoption in industrial applications.
The central role of knowledge in VA has been highlighted in many theoretical works [22]. Models of the VA process, such as the knowledge generation model by Sacha et al. [23], typically place knowledge as the input and final output of interactive visual analysis. These process models anchor knowledge solely in the human space, describing it as tacit knowledge that is only available to humans through their cognitive processes. Yet, based on earlier calls to make knowledge directly available for VA [24,25], Federico et al. [11] introduced the conceptual model of KAVA, which pursues explicit knowledge that is available in the machine space in visualization and analytical methods. Some works have dealt with explicit knowledge in the context of their VA topic: Battle and Ottley [26,27] reviewed 41 papers to categorize various definitions of “insight”, a term often used for the output of VA. They observed that “incorporating user domain knowledge appears to be a consistent challenge within visualization research”; discussed how insights can be externalized in VA approaches; and concluded that knowledge links between findings, data, and domain knowledge, like a knowledge graph, are the most promising conceptualization. Li et al. [28] interviewed 19 knowledge graph practitioners regarding visualization aspects of their current work and derived VA challenges and opportunities for knowledge graphs. Ceneda et al. [29] analyzed 53 VA approaches and how they provide guidance to users, e.g., recommending possibly interesting parts of large scatterplot matrices or networks. The use of explicit knowledge for guidance is conceptualized in an extension of the KAVA model [9]. Explicit knowledge can also be externalized while building ML models using VA approaches such as visual interactive labeling [30]. In this regard, a survey of 52 approaches [31] compared the importance of different knowledge types but did not discuss how knowledge would be reintegrated as explicit knowledge. Thus, the only survey covering KAVA approaches directly was conducted in 2017 by Federico et al. [11].
It reviewed 32 KAVA approaches, their support for the knowledge process and type, and the origin of knowledge. This work will provide a follow-up review with a thematic focus on industrial manufacturing.

3. Scope and Methodology

Our survey focused on KAVA in industrial manufacturing, covering research published from 2014 to 2024. We followed the standardized search and research guidelines of the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) [14,15] to ensure a transparent and structured approach (Figure 3). We first conducted an exploratory pre-selection phase to refine the search criteria and scope. Based on this, a search string was developed and adapted according to the syntax and indexing of each selected database to download the studies. Each stage of the screening process, title screening, abstract reading, and full-paper review, was independently conducted by at least two reviewers. Any disagreements were resolved by consulting a third reviewer who served as a tie-breaker. Two reviewers carried out paper categorization, discussed their results, and agreed upon the final categorization.

3.1. Search Strategy

During the ideation phase, a structured keyword search was performed across six key academic databases: ACM Digital Library, IEEE Xplore, Springer-Link, Google Scholar, SAGE, and Eurographics. At first, keywords matching our research questions and criteria were collected, such as “Data Visualization”, “Industry”, “Knowledge”, “Smart Factory”, “Smart Manufacturing”, and “Visual Analytics”, along with their variants and synonyms. After agreeing on the keywords, initial experiments with variants of search terms were carried out as described in the following section.

3.1.1. Search Term Pre-Selection Phase

Initially, experimental searches were performed on the databases to estimate the expected number of results. This pre-selection was performed on papers published between 2014 and 2024. The search process used Boolean operators (AND, OR, NOT) and wildcards (e.g., “Visuali?ation” to capture spelling variations) to improve precision. We aimed for a query that would yield 1000–1300 results on Google Scholar. This approach was considered to balance the need for a comprehensive resource pool with a manageable scope.
The chosen search string from the pre-selection phase was as follows:
(“Data Visualization” OR “Data Visualisation”) AND “Dashboard” AND (“Visual Analytics” OR “Visual Analysis”) AND Knowledge AND (Smart Manufacturing OR Smart Factory OR Smart Production OR Industry*)
We manually screened the titles of the retrieved resources to obtain insights into which domains we should specifically exclude from our search, which were not suitable for our scope of industrial manufacturing (e.g., urban planning and smart cities, education, business, healthcare, and energy sectors). Following these insights, we explicitly defined domains outside of industrial settings as an exclusion criterion and adapted our final search strategy accordingly.

3.1.2. Finalized Search Strategy and Screening Criteria

After the pre-selection phase, we decided on the following search string, which was further adjusted to match the syntax of each specific database:
(“Data Visualization” OR “Data Visualisation”) AND “Dashboard” AND (“Visual Analytics” OR “Visual Analysis”) AND Knowledge AND (Smart Manufacturing OR Smart Factory OR Smart Production OR Industry) -cit -busin -energ -health -educ -student
A total of 1352 papers were retrieved after applying our search strategy to the databases. To ensure quality, we considered only peer-reviewed venues. Established publishers such as Springer, ACM, and IEEE enforce peer review as a publication standard; for other outlets, we validated peer-reviewed status by consulting the journal’s editorial or preface.
After removing duplicated entries, 1318 papers remained for the screening procedure, which included a title, abstract, and full-paper reading step. To ensure the relevance and quality of the included studies, progressively stricter inclusion and exclusion criteria were applied at each screening stage, as detailed in Table 1. While multiple authors contributed to the overall screening process, each stage was carried out independently by two authors who reviewed the same set of papers. If their decisions did not align, a third reviewer acted as a tie-breaker. To minimize bias and maintain transparency, we adhered to the principles of the PRISMA checklist, which we considered more suitable for our review than a formal bias scoring system.

4. Categorization and Results

After completing the screening process, 13 papers were selected for fine-grained categorization. The analysis was structured around six main entity groups, as illustrated in Figure 2 and Table 2, namely data [D] (as defined by Schreck et al. [3]), knowledge [ K ϵ ] (as in Federico and Wagner et al. [11]), and visualization [V, S] (according to the Data Visualization Catalogue [32]). In addition, we reported on visualization techniques that are not covered by the Data Visualization Catalogue because they are described in the source, which are labeled “as reported” in Table 2.
Furthermore, industrial domain, user group, and machine learning [A] were identified during the categorization process.
For ML, we defined supervised, unsupervised, and semi-supervised methods as the main categories.

4.1. Industrial Domain and User Group

KAVA systems can only succeed if they match the needs of their intended users. This requires understanding who the users are, what expertise they have, and the roles that they play in analysis and decision-making. Identifying these groups also means considering the industrial domains that they work in, since domain processes and data environments strongly shape system requirements. The next sections, therefore, examine the user group addressed in the reviewed publications and the industrial domain in which they operate.
Understanding the intended user groups of KAVA systems is essential for evaluating how well these systems support real-world manufacturing workflows. Across the 13 reviewed publications, operators and managers are the most frequently addressed groups, each appearing in 6 of the 13 works (see Table 3). Operators typically execute shop-floor tasks and interact with monitoring, control, or annotation interfaces, with systems emphasizing usability, real-time feedback, and contextual visualizations [33,34,35]. Managers focus on high-level oversight, decision-making, and performance monitoring, where visualizations aggregate data and highlight KPIs or process summaries [36,37]. Process engineers are identified in three papers. They are involved in designing, optimizing, and troubleshooting production processes. The systems for this group often support detailed process modeling and knowledge integration [35,38,39]. Machine engineers are also mentioned in three works; their responsibilities include configuring and maintaining equipment. The visual tools for this group often focus on sensor data, machine states, and fault diagnosis [35,39,40]. Domain experts appear in three papers. These users contribute specialized knowledge, often in the form of annotations, rules, or ontologies. Their input is crucial for knowledge modeling and system interpretability [41,42,43]. Analysts are also present in three papers. They are typically data scientists or engineers responsible for performing data-driven analyses. The systems supporting analysts often include advanced filtering, modeling, and visualization capabilities [38,39,44].
While user groups define the roles and responsibilities that KAVA systems must support, their needs cannot be fully understood without considering the industrial domain in which they operate. The reviewed publications address four recurring domains, which are the manufacturing industry (discrete production with items that can be counted, touched, and seen individually), process industry (the production of bulk materials through chemical, physical, or biological processes), heavy industry and machinery, and the automotive sector, summarized in Table 4. The following paragraphs summarize how these domains are represented in the reviewed publications.
The manufacturing industry and process industry are the most frequently represented domains, each appearing in 5 of the 13 reviewed works (see Table 4). Works in the manufacturing industry typically focus on discrete production environments, such as assembly lines, component fabrication, and smart factory setups. Examples include ontology-based systems for monitoring and planning [34,39]. The process industry, on the other hand, includes continuous production environments such as cement, chemical, or energy sectors. These systems often emphasize sensor integration, process modeling, and fault detection [35,37,41]. The automotive sector is represented in two papers [36,38], with a focus on vehicle manufacturing and battery module assembly. These works emphasize multi-stakeholder collaboration and knowledge contextualization. The heavy industry is addressed in one paper [40], focusing on large-scale equipment and infrastructure.

4.2. Data

The data essentially drive the analytical processes of KAVA systems. We adopt the data-type taxonomy as defined in Schreck et al. [3], which distinguishes between tabular, time-dependent, spatial, textual, graph/network, and multimedia data (e.g., image, audio, and video). Based on this classification, we categorize the 13 reviewed publications.
The analysis reveals a strong emphasis on structured formats: tabular, time-dependent, spatial, and textual data dominate the literature (see Table 5). Tabular data underpin most approaches, supporting performance diagnostics and anomaly detection in sensor logs and production metrics [36,39]. Time-dependent data are used for temporal analyses such as multivariate series for crane anomaly detection and trend analysis in production quality [37,40]. Spatial data inform contextualized visualizations via 3D factory models and shop-floor layouts [34,38]. Textual data enable the extraction of domain knowledge from manuals and thesaurus-guided annotation [35,45]. Graph and network data are used in only three papers, visualizing process dependencies and modeling relationships between annotated cases [36,44]. Multimedia data appear in two publications, supporting the CAD-based annotation of battery modules and metadata-driven visualizations based on image inputs [38,43].
After identifying the data types in the reviewed publications, we now consider how these data are represented visually.

4.3. Visualization

Visualization techniques are a core component of KAVA systems, serving as the interface between data, knowledge, and user interpretation. To understand how visual representations are employed in industrial manufacturing contexts, we categorized the reviewed publications based on the types of visualizations that they use. These were grouped into six main subcategories, namely, graphs/plots, tables, diagrams, maps, others, and as reported (in source)—the latter capturing visualization types not listed in standard catalogs but explicitly mentioned in the papers. For clarity and emphasis, the order of the subcategories in Table 6 reflects their frequency of occurrence across the reviewed works.
Across the reviewed literature, line graphs emerged as the most frequently used visualization technique, appearing in 7 of the 13 papers. For instance, Alm et al. [34] and Kaupp et al. [39] used line graphs to display sensor trends and performance metrics over time.
Bar charts were the second most common, featured in six papers, examples of which are Xu et al. [36] and Behringer et al. [42] (see Figure 4), where they were used to compare categorical values or summarize aggregated data. Scatterplots were employed in five studies, namely, [33,40,43,44,45], to explore variable relationships and detect anomalies.
Other common visualization types included histograms, radial bar charts, box-and-whisker plots, and adar charts, with each appearing in two publications [36,38,40,41,42]. These supported tasks such as distribution analysis (e.g., Eirich et al. [38]) and multivariate comparisons (e.g., Zhang et al. [41]). Less commonly used were tacked bar charts, area graphs, stream graphs, stacked area graphs, and parallel coordinate plots, each found in only one paper [33,34,40].
Table-based visualizations also played a role in supporting temporal and density-based analyses. Calendars were used in two papers [33,36]. Heat maps appeared in one paper [33]. For example, Wu et al. [33] used calendar views to visualize maintenance schedules and heat maps to indicate when faults occur.
Regarding the diagram category, network diagrams appeared in seven papers, which are listed in detail in Table 6. These were used to represent relationships between entities, such as in [35,36], where they were used to visualize process flows or semantic structures. Some of them could also be considered flowcharts, which were used to depict system logic or task sequences, such as in [36,39,42,44].
Among the map visualizations, connection maps were used in one paper [38] to illustrate the spatial layouts of manufacturing stations.
Regarding the other category, donut charts appeared in three papers [33,38,41], often used to represent categorical distributions or knowledge maturity levels. Pie charts were used in two papers [33,43], and circle packing was mentioned once [33] to visualize hierarchical data compactly. One paper used a Nightingale rose chart to show stall distribution [40].
The as reported category includes domain-specific or custom visualizations not listed in the catalogs. Three-dimensional models were used in three papers [34,36,41] to support spatial reasoning and digital twin representations. Other unique visualizations, such as 2D models of products, were used in two papers [38,43]. Other unique visualizations, such as radial heat maps, icicle plots, and Marey’s graphs, were each found in only one publication [36,37,38].
Visualization alone does not suffice for complex analysis in industrial manufacturing. KAVA builds on visual exploration by embedding explicit domain knowledge such as rules, ontologies, and structured annotations directly into the analytic workflow. This allows visual patterns to be interpreted within their industrial context and transformed into actionable insights. The following section investigates how explicit knowledge is modeled and applied in KAVA systems.

4.4. Knowledge

Explicit knowledge is understood as formalized and structured domain knowledge that can be shared, reused, and reasoned with [11]. It plays a central role in embedding domain expertise and guiding analytic processes. To examine how explicit knowledge is integrated into VA for industrial manufacturing, we adopt the categorization framework proposed by Federico et al. [11], which conceptualizes knowledge as flowing through different transformation pathways, which they call KAVA processes. An overview of how these pathways, knowledge types, origins, and storage mechanisms are represented across the reviewed systems is provided in Table 7.
These pathways describe how knowledge is created, transformed, and applied in KAVA systems by linking tacit knowledge ( K τ ), explicit knowledge ( K ϵ ), data (D), analysis (A), and visualization (V). They aim to show the dynamic interplay between human and machine in knowledge-assisted analytics. Following Federico et al. [11], we distinguish six pathways in the surveyed literature; the “Simulation” pathway is excluded, as it was not reported in our papers.

4.4.1. Data Analysis ( D A K ϵ ):

Explicit knowledge is automatically obtained from data through analytic methods such as clustering, anomaly detection, or classification. For instance, Xu et al. [36] apply outlier detection algorithms (the greedy algorithm/unsupervised and k-nearest neighbor graphs/semi-supervised algorithms), while Liu et al. [40] use unsupervised ML models such as Isolation Forest (iForest), k-nearest neighbors (KNNs), Local Outlier Factor (LOF), and One-Class Support Vector (OCSVM) to flag anomalies. Similarly, Wu et al. [33] use Pearson correlation coefficients for feature extraction and Gaussian Mixture Models for real-time equipment condition monitoring. Kaupp et al. [39,44] use the composition of multiple outlier detection algorithms such as ARIMA or autoencoders.

4.4.2. Knowledge Visualization ( K ϵ V P K τ ):

Explicit knowledge is directly visualized, enabling users to internalize structured knowledge. Xu et al. [36] update their knowledge representations carried out with Marey’s graphs in real time via brushing. Some papers use node–link diagrams for displaying semantic networks [35], for displaying production halls [38], and for displaying entity graphs [45]. Behringer et al. [42] display numeric data quality scores as color-coded bars and pre-defined rules as labels within their dashboard (see Figure 4). Kaupp et al. [39,44] display human-annotated abnormal cases with contexts and rules as glyphs on their dashboard. Clicking on them reveals more details. They also introduce an interactive user interface for creating such human-annotated abnormal cases. The simplest way to visualize knowledge is to display thresholds, as in Cruz et al. [37]. Other papers such as Liu et al. [40] do not explicitly mention how the human explicit knowledge is visualized. However, in their work, the knowledge changes the specification of the visualization, resulting in different visualizations. Berges et al. [43] embed expert knowledge into visualizations by generating customized views of each machine that display only the sensors actually installed on it, as defined in the ontology. User-defined constraints (e.g., time ranges, limits, and aggregations) are then reflected in dynamically adapted query forms and tailored charts. Zhang et al. [41] visualize explicit human knowledge by displaying expert-defined rules and thresholds in 3D digital twin views, using color codes and progress bars.

4.4.3. Direct Externalization ( K τ X K ϵ ):

Tacit knowledge is made explicit through direct user input. A frequent approach is annotation and labeling. For example, Wu et al. [33] and Xu et al. [36] allow users to highlight or label data samples via brushing, thereby explicitly marking normal or abnormal cases. Alm et al. [34] describe annotating documents, photos, and structured metadata, while Eirich et al. [38] support the annotation of CAD images by drawing rectangles on their interface. Another approach is rule and threshold specification. Cruz et al. [37] and Behringer et al. [42] report on explicitly setting thresholds for alerts. Liu et al. [40] allow experts to edit rule files and to manually exclude false anomalies. Berges et al. [43] employ ontology-driven query forms, where users select sensors and constraints (e.g., time ranges, limits, and aggregations), which are directly embedded into the visualization output. More advanced systems support interactive decision adjustments. Zhang et al. [41] provide a 3D digital twin interface where experts define targets and thresholds and can override algorithm outputs. These manual adjustments and expert opinions are recorded for later reuse, though the paper does not detail the integration mechanism. Similarly, Kaupp et al. [39,44] allow the creation of Context-Infused Cases (CICs) in which anomalies are annotated with structured rules and contextual descriptions that can be reapplied in subsequent analyses. Finally, some approaches rely on the manual entry of domain knowledge, such as those in Han et al. [35], where domain experts can adjust a knowledge graph. However, this is not directly carried out in the UI but in the database before the knowledge graph construction.

4.4.4. Interaction Mining ( K τ E S A K ϵ ):

Tacit knowledge is inferred when the system goes beyond applying explicit user inputs and instead derives more general rules or models from interaction patterns. For example, in Wu et al. [33], analysts brush over process data to mark normal samples. These brushed labels are not only stored as annotations but are also considered by the system to infer normal and abnormal ranges through correlation analysis and Gaussian Mixture Models, thus including expert input into predictive models. Similarly, Xu et al. [36] allow users to brush over time-series anomalies. The system interprets these selections to update its outlier detection results in real time. Kaupp et al. [39,44] manually mark unusual patterns on performance dashboards. Beyond simply storing these annotations, the system formalizes them into structured anomaly cases and trains predictive models of machine behavior. Behringer et al. [42] also employ a form of interaction mining in their data quality monitoring framework. Here, when domain experts repeatedly adjust thresholds for quality dimensions like completeness or accuracy, the system learns from these adjustments and updates its domain-specific rules automatically.

4.4.5. Intelligent Data Analysis ( ( D , K ϵ ) A K ϵ ):

Existing explicit knowledge is reused to improve automatic analysis methods. For example, Ameri et al. [45] employ a thesaurus-driven entity extractor, where domain-specific vocabularies guide the identification of manufacturing capabilities in unstructured data. Kaupp et al. [39,44] reuse previously defined Context-Infused Cases (CICs) to retrain their anomaly detection algorithms to achieve more accurate and context-aware fault identification. Similarly, Liu et al. [40] combine anomaly detection results from unsupervised learning models with expert-defined decision-tree rules.

4.4.6. Guidance ( K ϵ A S ):

Explicit knowledge is reused by the system to suggest or adapt visualization specifications. For example, Eirich et al. [38] encode ontological models of production processes, which the system uses to suggest suitable visualization types and provide dashboards tailored to different user groups. Similarly, Zhang et al. [41] integrate expert-defined rules and sensor-specific thresholds into their digital twin dashboards so that color-coded decision panels, progress bars, and alerts dynamically adapt based on these encoded constraints.
Beyond these transformation processes, the reviewed works also reflect differences in the types of knowledge addressed. Operational knowledge refers to how users interact with the system, such as brushing, filtering, or dashboard configuration [36]. Domain declarative knowledge encompasses facts, vocabularies, and semantic structures, often represented through ontologies or thesauri [34,35]. Domain procedural knowledge captures rules, workflows, or decision logic, such as maintenance schedules or quality thresholds [33,37].
Explicit knowledge also enters KAVA systems at different origins. Pre-design knowledge is formalized before system development, for instance, via expert interviews or domain standards [34]. Design-stage knowledge is embedded by developers, such as through predefined rules or ontology schemas [38]. Post-design knowledge is generated dynamically during system use. Sources include automated analysis [40]; single-user contributions such as the manual specification of error cases and diagnostic rules [44]; and multi-user collaboration, which produces shared knowledge bases [35].
Finally, the storage mechanisms determine how explicit knowledge persists and is reused. Ontologies, RDF knowledge graphs [35,40], and relational databases [39] are the most common storage mechanisms. Some systems also store structured diagnostic cases and rule sets for future sessions [44]. Others use cloud infrastructures [41] or modular repositories, allowing multiple users to contribute knowledge collaboratively and to expand and refine the stored content over time [42].
While explicit knowledge forms the conceptual backbone of KAVA systems, enabling users to contextualize, interpret, and structure information, many reviewed works further extend these capabilities with data-driven methods. In particular, data mining and machine learning (ML) techniques are central to automating the extraction of knowledge from large-scale industrial data. Since ML plays a recurring and critical role across several transformation pathways, we next provide a dedicated analysis of its use.

4.5. Machine Learning

The integration of data mining and machine learning (ML) techniques supports tasks such as anomaly detection [36], classification [45], clustering [40], and optimization [41]—often in combination with visual interfaces to enhance interpretability and user control. To understand how ML is applied in the reviewed literature, we categorized the papers based on the type of learning approach used. Table 8 presents this categorization.
Unsupervised learning is the most frequently used approach, appearing in 6 of the 13 papers (e.g., [39,40,44,45]). Semi-supervised learning follows with two mentions [33,36], while supervised learning is used in one paper [44]. Zhang et al. [41] describe a hybrid intelligent optimization system that combines evolutionary algorithms, reinforcement learning, and human-in-the-loop decision support. Unlike traditional ML approaches, such as supervised or unsupervised learning, this system relies on both data-driven models and expert knowledge. It also allows users to interact with the system, providing feedback to help make better decisions in complex, changing environments.
Notably, half of the reviewed papers (6 out of 13) either did not specify their ML approach or did not incorporate ML techniques at all.

5. Discussion

Having surveyed the existing literature on KAVA in industrial manufacturing, this section aims to discuss and synthesize our findings, highlighting recurrent patterns considered in the papers used for knowledge modeling and visualization. We also investigate the methodological challenges and open questions across our six core building blocks: user group, industrial domain, visualization, knowledge, data, and machine learning.

5.1. Main Findings

5.1.1. Wide Variations in User Group and Industrial Domain

Across the 13 surveyed papers, most approaches serve two primary audiences: operators, who require shop-floor visualization for monitoring and contextual views to react quickly to alerts, and managers, who require high-level summaries of key performance indicators (KPIs). However, the managers and operators are rarely specified in detail. Other stakeholders, such as process engineers, machine engineers, data analysts, and broadly defined domain experts, work with separate purpose-built interfaces, for example, workflow editors, ontology management tools, or custom filtering modules, rather than the shared dashboards used by operators and managers. The majority of systems are situated in the manufacturing and process industries, with fewer examples from automotive sectors or the heavy industry.

5.1.2. Heavy Use of Domain-Specific Visualization Techniques

Most reviewed KAVA systems rely on general-purpose visualization types. Line graphs, bar charts, and scatterplots are among the most used ones. In addition, simple network diagrams and flowcharts are used to show connections between elements or process steps. Several papers employ domain-specific visualizations, where standard charts are either inadequate or less commonly used than more specialized formats. For instance, Marey’s graphs excel in visualizing temporal dependencies [36]; others introduce radial heat maps [37], icicle plots [38], or custom 2D/3D models visualizing factory floors, products, or specialized diagrams when the spatial context is critical [34,36,38,41,43].
Two key insights emerge: First, the heavy reliance on standard visualization types points to a need for reusable visualization templates that can be parameterized for different manufacturing tasks rather than rebuilding similar views from scratch in each project. Second, although many systems store and apply explicit knowledge (rules, thresholds, ontologies, text documents, etc.), almost none of them include a declarative mapping from those knowledge models to visualization specifications. In practice, many visualization types and thresholds are still hard-coded rather than being automatically suggested or configured based on the underlying domain knowledge represented in the system.

5.1.3. Challenges in Storing, Visualizing, and Validating Knowledge

Our review shows that explicit knowledge in current KAVA systems falls mainly into two archetypes: (1) declarative knowledge, defined as vocabularies (e.g., taxonomies and ontologies), and (2) procedural knowledge, defined as rules (e.g., thresholds and if–then logic). In practice, most systems set and load a fixed set of terms or rules at design time, and then they allow individual users to add or adjust labels, thresholds, or configurations and/or let users annotate labels on a dashboard (e.g., by brushing or drawing rectangles) during a session. Only a few works feed those user edits back into their analytical pipeline for later reuse, for example, by retraining models or updating rules. However, such knowledge persistence might have been realized but not explicitly mentioned in the publication. Even if knowledge persistence is realized, whether evenly or unevenly among users, the question of input quality remains. When multiple users annotate the same data, how can we ensure that their labels are reliable? A possible solution for this could be the implementation of calculating inter-annotator agreement metrics, majority-voting schemes, or confidence scoring to validate annotations.
When it comes to storage, most platforms either keep ontologies and taxonomies in dedicated RDF or graph databases (e.g., Neo4j graph database as in Han et al. [35]) or store rule tables and configuration files in relational databases or flat files (e.g., Hadoop or MangoDB as in Wu et al. [33]). Some systems use cloud-based repositories to store and process industrial data streams, as in Zhang et al. [41], although aspects such as multi-user contributions or collaborative knowledge management are not addressed. Data silos rarely interoperate since ontology updates do not propagate to relational rule tables, and user-generated annotations often vanish at session end, a limitation also emphasized by Wagner’s KAVA model, which notes that current systems do not distinguish between single and multi-user settings and provide no clear specification of how explicit knowledge is stored, propagated, or validated—misleading knowledge can corrupt a system [47].
Knowledge visualization takes several forms. Thresholds and rule boundaries are most commonly drawn as lines or colored bands on time-series plots; categorical vocabularies appear as labels or color encodings on bar charts; and individual annotations show up as glyphs or icons on scatterplots and network diagrams. A number of systems embed richer structures—semantic networks and process flows rendered as node–link diagrams or expert rules and progress indicators overlaid on 3D digital twins.

5.1.4. Limited Integration of Machine Learning

ML techniques appear in roughly half of the reviewed systems, primarily using unsupervised models (e.g., for anomaly detection) and semi-supervised methods (e.g, for outlier detection), followed by supervised approaches or mixed methods. However, over half of the studies either omit ML entirely or do not report how the models were integrated with the knowledge and visualization layers. This limited integration is consistent with findings from the broader visual analytics community. Bernard et al.’s [30] visual-interactive labeling (VIAL) process explicitly examines the interplay between active learning and visual-interactive labeling, demonstrating both the potential and the challenges of aligning model-centered ML techniques with user-centered visualization. Their work highlights that, although ML can guide labeling and knowledge generation, integration with KAVA systems often remains ad hoc, hard-coded, or absent. This further supports our observation that systematic ML integration into KAVA remains an open challenge despite its wider adoption in Industry 4.0 contexts.

5.2. Open Challenges and Future Directions

While notable progress has been made in applying KAVA research to industrial contexts, our comprehensive review of the selected papers and related literature indicates that several challenges remain to be addressed.

5.2.1. Heterogeneity of Explicit Knowledge and Data

In the surveyed approaches, we encountered a variety of explicit knowledge, including basic rule tables and ontologies, as well as standards documents, process and architecture diagrams, machine-generated insights, and various user annotations.
Another challenge is posed by the heterogeneity of data: while tabular and time-dependent data are almost omnipresent, 9 out of 13 approaches incorporate an additional data category such as spatial, textual, graph, or multimedia data. This corroborates the results of an earlier survey on VA for industry [3,17,19] regarding data heterogeneity due to ongoing digitalization.
Future research should support a broad spectrum of knowledge and data types. A unified storage layer is needed to organize and version resources, propagate updates, and enable “what-if” simulations on updated rules or models.
Standardized representations would improve interoperability across KAVA systems, ease integration with industrial IT infrastructures, and support the reuse and exchange of knowledge across organizations and sectors.

5.2.2. Higher Level of Automation

A major challenge in KAVA research is achieving a higher level of automation across data processing, knowledge modeling, knowledge externalization, and knowledge generation. In this regard, Alm et al. [34] state that manual formalization is inconvenient and easily neglected by users due to competing priorities. Therefore, a higher level of automation is necessary for context detection, annotatation, and linking of data. Federico et al. [11] propose deriving explicit knowledge from user interactions as a route to automate knowledge externalization. Other authors also propose automation opportunities, such as AI-assisted pipeline generation based on user feedback [39], large language models (LLMs) to extract knowledge to automate the construction of process knowledge graphs (PKGs) [20], or the automatic alignment of entities between different knowledge graphs to fully utilize their capabilities [42].
This presents notable directions for future research: developing adaptive pipelines that combine explicit knowledge with automated inference to keep visualizations and dashboards continuously updated, validating and aligning knowledge models from streaming data and user interactions, and maintaining traceability at scale. Further opportunities include leveraging large language models and domain-specific reasoning tools to extract and formalize knowledge from heterogeneous sources, as well as developing automated methods to assess the quality of user-generated inputs. Beyond knowledge extraction, LLMs hold particular promise for semi-automated knowledge graph generation, which could reduce the manual effort of formalization, improve scalability, and enable cross-domain interoperability of industrial knowledge assets. Advancing these areas could lead to KAVA systems that not only externalize and visualize knowledge but also evolve dynamically with changing industrial contexts.

5.2.3. Effective User Interfaces

While automated data analysis approaches are attractive from the perspectives of time and cost, VA approaches provide opportunities beyond that. KAVA systems need to address, in particular, data analysis problems that cannot be solved fully automatically and effectively distribute tasks. Cibulski et al. [18] warn that overly complex dashboards can be overwhelming and thus ineffective. They recommend visualization onboarding [9] and guidance [48] due to their great potential in helping to lower the barrier to adoption and reduce manual effort for users. Thus, we encourage further research on such approaches, as well as automatically recommending suitable visualizations.

5.2.4. Validation of Human Annotations

In KAVA, human input is essential but prone to error, which makes its validation critical in industrial and manufacturing contexts, where incorrect data can lead to security risks or unplanned maintenance. Therefore, we motivate the issues of how to ensure that user-provided input is correct, how to manage conflicting input from multiple users, and how to incorporate confidence measures into the labeling process [9,11,25]. Also, the interview study by Li et al. [28] describes data quality as the most common challenge (reported by 15 out of 19 interviewees). In this regard, crowd-sourcing research has investigated how to manage distributed human input with techniques for task design, quality control, and the validation of user-generated data [49]. Recent work by Liu et al. addresses this challenge with a VA approach that involves experts in the labeling process within crowd-sourced settings. Their learning-from-crowds model supports the identification and correction of uncertain labels, e.g., through majority voting and interactive visualizations, for the progressive improvement of annotation quality. This work demonstrates how combining ML with human validation can enhance data quality while reducing manual effort [50]. Future research could explore leveraging crowd-sourcing, which generally involves large groups of non-experts; however, we motivate KAVA research that learns from those approaches for annotation validation in industrial settings.

5.2.5. Long-Term Studies and Real-World Scenarios

An open challenge is the lack of real-world evaluations of the proposed models and systems. Many tools show promising results in controlled and limited settings; however, their effectiveness in long-term industrial environments is largely unknown. Several authors highlight this issue and motivate studies to address this gap. For example, Xu et al. note that, with the planned deployment of their system in real production lines, its long-term usage can be studied [36]. Similarly, Wu et al. intend to conduct long-term studies with quantitative measurements and end-user feedback, which can further improve their system [33]. Behringer et al. [42] and Kaupp et al. [44] motivate the validation of their prototypes and models through extended, real-world scenarios, while Buchgeher et al. [21] point out the limited validation of existing solutions across many industrial sectors and applications. Also, Cibulski et al. [18] observed in their survey that none of the solutions with the highest technology readiness level had been evaluated. Yet such studies are essential to further advance KAVA by assessing its practical impact adaptability and scalability across dynamic manufacturing domains. As suggested by Federico et al. [11], novel evaluation methods can be conceptualized that instrumentalize the changes and flows of explicit knowledge as a unit of measurement for studies on the effectiveness of VA environments. Advancing this conceptual framework represents an important avenue for future research. In addition to individual case studies, there is a pressing need for longitudinal evaluation frameworks that assess the adoption of KAVA in industrial settings over extended periods. Such frameworks could capture the evolving interplay of human, organizational, and technological factors and provide structured evidence on adaptability, scalability, and sustained value creation in real-world manufacturing environments.

5.2.6. Generalizability

A recurring open challenge in VA is the limited generalizability of the proposed systems across domains and datasets (e.g., [19]), since tools are mostly developed and tailored to specific use cases, which makes it difficult to transfer findings to other scenarios. For example, Eirich et al. [38] acknowledge this limitation and state that their generalizability remains to be investigated in broader applications, while Wu et al. point out that their interactive techniques for user-steerable outlier detection can be potentially applied to more general usage scenarios [33]. However, the majority of interviewed knowledge graph practitioners from different application domains stated that domain-specific challenges stand in the way of generalized solutions [28]. Research needs to be conducted to determine how—at least in manufacturing—commonalities can provide a common basis for generalization. Recently, Subramonyam et al. published a review that discusses the existing literature and provides recommendations on how to address this challenge [31]. Among other points, the authors discuss how dependencies, such as dataset-specific assumptions or configuration constraints, are often insufficiently documented in research papers, which reduces the reproducibility and, consequently, the generalizability of research findings. They argue that more transparent reporting and more explicit reasoning about design choices are necessary to support the generalization and reuse of research outcomes. Therefore, we motivate generalizability and documentation standards to support the broader applicability of KAVA.

5.2.7. Methodological Limitations

The exclusion of gray literature, such as industrial reports and white papers, was a deliberate choice to ensure the reproducibility and quality of the included studies. However, this decision might have resulted in missing relevant insights from industry practice and could have increased the number of papers included in the systematic literature review. A further limitation concerns the vocabulary of our search strategy: by focusing on terms such as “data visualization,” “dashboard,” “visual analytics,” “visual analysis,” “knowledge,” and “smart manufacturing,” relevant work using alternative terminology (e.g., guidance, provenance, rule-based, ontology-driven, knowledge graph, explicit knowledge, or human-in-the-loop) may have been excluded. While we retained this string for consistency, we explicitly acknowledge this risk and provide the full database-specific queries, including all limits and language restrictions, in Supplementary Material S1 to ensure transparency and reproducibility.
Including such sources may be beneficial for future work provided that alternative mechanisms are applied to ensure their quality. In addition, while dual independent screening was conducted at each stage, no formal inter-reviewer agreement metrics (e.g., Cohen’s kappa) were calculated. Disagreements were resolved through a tie-breaker process to maintain consistency. Incorporating formal agreement metrics in future work could further strengthen methodological rigor.

6. Conclusions

This review of 13 KAVA systems in industrial manufacturing investigated how explicit knowledge is modeled and stored, how it is externalized through visualization and interaction, and how ML is integrated into these processes.
In answering RQ1, the analysis shows that explicit knowledge is most frequently represented as ontologies, taxonomies, knowledge graphs, and rule sets, often enriched by user-provided annotations and threshold specifications. These resources are stored in RDF or graph databases, relational tables, or flat files. However, interoperability between formats is limited. While several systems retain post-design contributions, such as annotated anomaly cases or updated ontologies, others keep them only temporarily, leading to inconsistent long-term availability. Concerns about the reliability of user-generated input are discussed in the literature, though none of the reviewed works systematically evaluate annotation quality.
Addressing RQ2, the surveyed systems predominantly employ line graphs, bar charts, and scatterplots, often enhanced with thresholds, annotations, and filters derived from explicit knowledge. Explicit knowledge is directly visualized in various ways. Some systems visualize knowledge directly. For example, Marey’s graphs update in real time through brushing, and node–link diagrams display semantic networks, production halls, or entity graphs. Dashboards use numeric quality scores as color-coded bars and pre-defined rules as labels. Human-annotated abnormal cases are displayed with contextual rules as glyphs, often as clickable elements for more details. Interactive interfaces allow for the creation of such annotated cases. The simplest approach is displaying thresholds directly. In other systems, explicit knowledge alters visualization specification, leading to different outputs without direct display. Expert knowledge can also define customized views. For instance, ontologies specify which sensors belong to a machine, and user-defined rules and constraints adapt queries and charts dynamically. In other cases, expert-defined rules and thresholds are displayed in 3D digital twin views using color codes and progress bars. Network diagrams and flowcharts show relationships and process logic. More specialized approaches employ radial heat maps, icicle plots, and custom 2D/3D product models when the spatial or hierarchical context is critical. Despite these integrations of knowledge, most configurations remain hard-coded. No reviewed system offers a fully declarative mapping from knowledge models to visualization specifications, which restricts automation and reuse.
With respect to RQ3, 6 of the 13 works incorporate ML, most often using unsupervised methods, followed by semi-supervised, supervised, and hybrid approaches. In some cases, outputs from these models are used to generate rules or are refined through human-in-the-loop processes. The remaining systems either do not integrate ML or do not specify their approach, and only a few detail how such models interact with knowledge representation and visualization, limiting reproducibility and transferability.
Across these three areas, recurring challenges, drawn from both the reviewed studies and related work, include the heterogeneity of knowledge and data types and the absence of unified storage that supports interoperability; limited automation in linking knowledge to visualizations and in externalizing knowledge; the need for interfaces that support varied user roles; the validation and reuse of human-generated annotations; a lack of long-term, real-world evaluations; and limited applicability beyond the specific industrial contexts studied. Progress in these areas will require integrated knowledge repositories that propagate updates, mechanisms for validating and reusing user contributions, tighter coupling of knowledge models with visualization specifications, and systematic assessments in operational manufacturing environments. These steps would improve the adaptability, reusability, and long-term impact of KAVA systems in industrial analytics.

Supplementary Materials

The following supporting information can be downloaded at: https://phaidra.fhstp.ac.at/o:7387. Table S1: Search terms used in Pre-Selection Phase.

Author Contributions

Conceptualization: A.J.B., S.G., J.V., C.S., and A.R.; Investigation: A.J.B., S.G., J.V., C.S., A.R., and M.W.; Methodology: A.J.B., S.G., J.V., C.S., A.R., and M.W.; Writing—original draft: A.J.B., and S.G.; Writing—review & editing: A.J.B., S.G., J.V., C.S., A.R., M.W., W.A., T.S., and J.S.; Supervision: W.A., and T.S.; Funding acquisition: M.W. All authors unanimously agree that both Adrian Böck and Stefanie Größbacher are first authors of the publication and may therefore be listed as first authors (jointly or individually). All authors have read and agreed to the published version of the manuscript.

Funding

The financial support provided by the Austrian Federal Ministry of Labour and Economy, the National Foundation for Research, Technology and Development, and the Christian Doppler Research Association is gratefully acknowledged.

Conflicts of Interest

Author Josef Suschnigg was employed by the company Pro2Future. Jan Vrablicz was also employed by Georg Fischer. All authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IoTInternet of Things
KAVAKnowledge-Assisted Visual Analytics
VAVisual Analytics
MLMachine Learning

References

  1. Posada, J.; Toro, C.; Barandiaran, I.; Oyarzun, D.; Stricker, D.; De Amicis, R.; Pinto, E.B.; Eisert, P.; Dollner, J.; Vallarino, I. Visual Computing as a Key Enabling Technology for Industrie 4.0 and Industrial Internet. IEEE Comput. Graph. Appl. 2015, 35, 26–40. [Google Scholar] [CrossRef]
  2. Tambare, P.; Meshram, C.; Lee, C.C.; Ramteke, R.J.; Imoize, A.L. Performance Measurement System and Quality Management in Data-Driven Industry 4.0: A Review. Sensors 2021, 22, 224. [Google Scholar] [CrossRef]
  3. Schreck, T.; Mutlu, B.; Streit, M. Visual Data Science for Industrial Applications. In Digital Transformation: Core Technologies and Emerging Topics from a Computer Science Perspective; Springer: Berlin/Heidelberg, Germany, 2023; pp. 447–471. [Google Scholar] [CrossRef]
  4. Wang, B.; Tao, F.; Fang, X.; Liu, C.; Liu, Y.; Freiheit, T. Smart Manufacturing and Intelligent Manufacturing: A Comparative Review. Engineering 2021, 7, 738–757. [Google Scholar] [CrossRef]
  5. Thomas, J.J.; Cook, K.A. Illuminating the Path: The Research and Development Agenda for Visual Analytics; National Visualization and Analytics Ctr: Richland, WA, USA, 2005. [Google Scholar]
  6. Wegner, P. Why Interaction is More Powerful Than Algorithms. Commun. ACM 1997, 40, 80–91. [Google Scholar] [CrossRef]
  7. Spence, R. Information visualization: Design for interaction, 2nd ed.; Pearson/Prentice Hall: Munich, Germany, 2007. [Google Scholar]
  8. Kielman, J.; Thomas, J.; May, R. Foundations and Frontiers in Visual Analytics. Inf. Vis. 2009, 8, 239–246. [Google Scholar] [CrossRef]
  9. Stoiber, C.; Ceneda, D.; Wagner, M.; Schetinger, V.; Gschwandtner, T.; Streit, M.; Miksch, S.; Aigner, W. Perspectives of visualization onboarding and guidance in VA. Vis. Inform. 2022, 6, 68–83. [Google Scholar] [CrossRef]
  10. Wagner, M.; Rind, A.; Thür, N.; Aigner, W. A knowledge-assisted visual malware analysis system: Design, validation, and reflection of KAMAS. Comput. Secur. 2017, 67, 1–15. [Google Scholar] [CrossRef]
  11. Federico, P.; Wagner, M.; Rind, A.; Amor-Amoros, A.; Miksch, S.; Aigner, W. The Role of Explicit Knowledge: A Conceptual Model of Knowledge-Assisted Visual Analytics. In Proceedings of the 2017 IEEE Conference on Visual Analytics Science and Technology (VAST), Phoenix, AZ, USA, 3–6 October 2017; pp. 92–103. [Google Scholar] [CrossRef]
  12. Wagner, M. Integrating Explicit Knowledge in the Visual Analytics Process. Ph.D. Thesis, TU Wien, Vienna, Austria, 2017. [Google Scholar] [CrossRef]
  13. Wagner, M.; Slijepcevic, D.; Horsak, B.; Rind, A.; Zeppelzauer, M.; Aigner, W. KAVAGait: Knowledge-Assisted Visual Analytics for Clinical Gait Analysis. IEEE Trans. Vis. Comput. Graph. (TVCG) 2018, 25, 1528–1542. [Google Scholar] [CrossRef] [PubMed]
  14. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  15. Page, M.J.; Moher, D.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. PRISMA 2020 explanation and elaboration: Updated guidance and exemplars for reporting systematic reviews. BMJ 2021, 372, n160. [Google Scholar] [CrossRef] [PubMed]
  16. Zupan, B.; Holmes, J.H.; Bellazzi, R. Knowledge-Based Data Analysis and Interpretation. Artif. Intell. Med. 2006, 37, 163–165. [Google Scholar] [CrossRef] [PubMed]
  17. Zhou, F.; Lin, X.; Liu, C.; Zhao, Y.; Xu, P.; Ren, L.; Xue, T.; Ren, L. A Survey of Visualization for Smart Manufacturing. J. Vis. 2019, 22, 419–435. [Google Scholar] [CrossRef]
  18. Cibulski, L.; Schmidt, J.; Aigner, W. Reflections on Visualization Research Projects in the Manufacturing Industry. IEEE Comput. Graph. Appl. 2022, 42, 21–32. [Google Scholar] [CrossRef]
  19. Wang, Y.; Zhu, Z.; Wang, L.; Sun, G.; Liang, R. Visualization and Visual Analysis of Multimedia Data in Manufacturing: A Survey. Vis. Inform. 2022, 6, 12–21. [Google Scholar] [CrossRef]
  20. Xiao, Y.; Zheng, S.; Shi, J.; Du, X.; Hong, J. Knowledge graph-based manufacturing process planning: A state-of-the-art review. J. Manuf. Syst. 2023, 70, 417–435. [Google Scholar] [CrossRef]
  21. Buchgeher, G.; Gabauer, D.; Martinez-Gil, J.; Ehrlinger, L. Knowledge Graphs in Manufacturing and Production: A Systematic Literature Review. IEEE Access 2021, 9, 55537–55554. [Google Scholar] [CrossRef]
  22. Bernard, J. The Human-Data-Model Interaction Canvas for Visual Analytics. arXiv 2025, arXiv:2505.07534. [Google Scholar] [CrossRef]
  23. Sacha, D.; Stoffel, A.; Stoffel, F.; Kwon, B.C.; Ellis, G.; Keim, D.A. Knowledge Generation Model for Visual Analytics. IEEE Trans. Vis. Comput. Graph. 2014, 20, 1604–1613. [Google Scholar] [CrossRef]
  24. Chen, M.; Ebert, D.; Hagen, H.; Laramee, R.S.; van Liere, R.; Ma, K.L.; Ribarsky, W.; Scheuermann, G.; Silver, D. Data, Information, and Knowledge in Visualization. IEEE Comput. Graph. Appl. 2009, 29, 12–19. [Google Scholar] [CrossRef]
  25. Wang, X.; Jeong, D.H.; Dou, W.; Lee, S.W.; Ribarsky, W.; Chang, R. Defining and Applying Knowledge Conversion Processes to a Visual Analytics System. Comput. Graph. 2009, 33, 616–623. [Google Scholar] [CrossRef]
  26. Battle, L.; Ottley, A. What Exactly Is an Insight? A Literature Review. In Proceedings of the IEEE Visualization and Visual Analytics (VIS), Melbourne, Australia, 21–27 October 2023; pp. 91–95. [Google Scholar] [CrossRef]
  27. Battle, L.; Ottley, A. What Do We Mean When We Say “Insight”? A Formal Synthesis of Existing Theory. IEEE Trans. Vis. Comput. Graph. 2024, 30, 6075–6088. [Google Scholar] [CrossRef] [PubMed]
  28. Li, H.; Appleby, G.; Brumar, C.D.; Chang, R.; Suh, A. Knowledge Graphs in Practice: Characterizing Their Users, Challenges, and Visualization Opportunities. IEEE Trans. Vis. Comput. Graph. 2024, 30, 584–594. [Google Scholar] [CrossRef]
  29. Ceneda, D.; Gschwandtner, T.; Miksch, S. A Review of Guidance Approaches in Visual Data Analysis: A Multifocal Perspective. Comput. Graph. Forum 2019, 38, 861–879. [Google Scholar] [CrossRef]
  30. Bernard, J.; Zeppelzauer, M.; Sedlmair, M.; Aigner, W. VIAL: A Unified Process for Visual Interactive Labeling. Vis. Comput. 2018, 34, 1189–1207. [Google Scholar] [CrossRef]
  31. Subramonyam, H.; Hullman, J. Are We Closing the Loop Yet? Gaps in the Generalizability of VIS4ML Research. IEEE Trans. Vis. Comput. Graph. 2024, 30, 672–682. [Google Scholar] [CrossRef]
  32. Ribecca, S. The Data Visualisation Catalogue. 2020. Available online: https://datavizcatalogue.com/home_list.html (accessed on 12 January 2024).
  33. Wu, W.; Zheng, Y.; Chen, K.; Wang, X.; Cao, N. A Visual Analytics Approach for Equipment Condition Monitoring in Smart Factories of Process Industry. In Proceedings of the 2018 IEEE Pacific Visualization Symposium (PacificVis), Kobe, Japan, 10–13 April 2018; pp. 140–149. [Google Scholar] [CrossRef]
  34. Alm, R.; Aehnelt, M.; Hadlak, S.; Urban, B. Annotated Domain Ontologies for the Visualization of Heterogeneous Manufacturing Data. In Proceedings of the 17th International Conference on Human Interface and the Management of Information, Angeles, CA, USA, 2–7 August 2015. Part I: Information and Knowledge Design. [Google Scholar] [CrossRef]
  35. Han, H.; Fu, D.; Jia, H. Construction of Knowledge Graphs Related to Industrial Key Production Processes for Query and Visualization. In Proceedings of the 2023 Chinese Intelligent Systems Conference, Ningbo, Zhejiang, China, 14–15 October; Jia, Y., Zhang, W., Fu, Y., Wang, J., Eds.; Springer: Singapore, 2023; pp. 855–863. [Google Scholar] [CrossRef]
  36. Xu, P.; Mei, H.; Ren, L.; Chen, W. ViDX: Visual Diagnostics of Assembly Line Performance in Smart Factories. IEEE Trans. Vis. Comput. Graph. 2017, 23, 291–300. [Google Scholar] [CrossRef]
  37. Cruz, A.B.; Sousa, A.; Cardoso, Â.; Valente, B.; Reis, A. Smart Data Visualisation as a Stepping Stone for Industry 4.0—A Case Study in Investment Casting Industry. In Proceedings of the Robot 2019: Fourth Iberian Robotics Conference, Porto, Portugal, 20–22 November; Silva, M.F., Luís Lima, J., Reis, L.P., Sanfeliu, A., Tardioli, D., Eds.; Springer: Cham, Switzerland, 2020; pp. 657–668. [Google Scholar] [CrossRef]
  38. Eirich, J.; Jäckle, D.; Sedlmair, M.; Wehner, C.; Schmid, U.; Bernard, J.; Schreck, T. ManuKnowVis: How to Support Different User Groups in Contextualizing and Leveraging Knowledge Repositories. IEEE Trans. Vis. Comput. Graph. 2023, 29, 3441–3457. [Google Scholar] [CrossRef] [PubMed]
  39. Kaupp, L.; Nazemi, K.; Humm, B. An Industry 4.0-Ready Visual Analytics Model for Context-Aware Diagnosis in Smart Manufacturing. In Proceedings of the 2020 24th International Conference Information Visualisation (IV), Melbourne, Australia, 7–11 September 2020; pp. 350–359. [Google Scholar] [CrossRef]
  40. Liu, J.; Zheng, H.; Jiang, Y.; Liu, T.; Bao, J. Visual Analytics Approach for Crane Anomaly Detection Based on Digital Twin. In Proceedings of the Cooperative Design, Visualization, and Engineering: 19th International Conference, CDVE 2022, Virtual Event, 25–28 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–12. [Google Scholar] [CrossRef]
  41. Zhang, K.; Xu, Q.; Liu, C.; Chai, T. Intelligent decision-making system for mineral processing production indices based on digital twin interactive visualization. J. Vis. 2024, 27, 421–436. [Google Scholar] [CrossRef]
  42. Behringer, M.; Hirmer, P.; Villanueva, A.; Rapp, J.; Mitschang, B. Enhancing Data Trustworthiness in Explorative Analysis: An Interactive Approach for Data Quality Monitoring. SN Comput. Sci. 2024, 5, 465. [Google Scholar] [CrossRef]
  43. Berges, I.; Ramírez-Durán, V.J.; Illarramendi, A. Facilitating Data Exploration in Industry 4.0. In Proceedings of the Advances in Conceptual Modeling: ER 2019 Workshops FAIR, MREBA, EmpER, MoBiD, OntoCom, and ER Doctoral Symposium Papers, Salvador, Brazil, 4–7 November 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 125–134. [Google Scholar] [CrossRef]
  44. Kaupp, L.; Nazemi, K.; Humm, B. Context-Aware Diagnosis in Smart Manufacturing: TAOISM, An Industry 4.0-Ready Visual Analytics Model. In Integrating Artificial Intelligence and Visualization for Visual Knowledge Discovery; Springer International Publishing: Cham, Switzerland, 2022; pp. 403–436. [Google Scholar] [CrossRef]
  45. Ameri, F.; Bernstein, W. A Thesaurus-guided Framework for Visualization of Unstructured Manufacturing Capability Data. In Proceedings of the IFIP Advances in Information and Communication Technology. APMS 2017 International Conference Advances in Production Management Systems (APMS 2017), Hamburg, Germany, 3–7 September 2017. [Google Scholar] [CrossRef]
  46. Behringer, M.; Hirmer, P.; Villanueva, A.; Rapp, J.; Mitschang, B. Unobtrusive Integration of Data Quality in Interactive Explorative Data Analysis. In Proceedings of the 25th International Conference on Enterprise Information Systems—Volume 2: ICEIS, Prague, Czech Republic, 24–26 April 2023; INSTICC. SciTePress: Setúbal, Portugal, 2023; pp. 276–285. [Google Scholar] [CrossRef]
  47. Wagner, M. Knowledge-Assisted Visual Analytics for Time-Oriented Data. Ph.D. Thesis, TU Wien, Vienna, Austria, 2017. [Google Scholar]
  48. Ceneda, D.; Gschwandtner, T.; Miksch, S. You Get by with a Little Help: The Effects of Variable Guidance Degrees on Performance and Mental State. Vis. Inform. 2019, 3, 177–191. [Google Scholar] [CrossRef]
  49. Bhatti, S.S.; Gao, X.; Chen, G. General framework, opportunities and challenges for crowdsourcing techniques: A Comprehensive survey. J. Syst. Softw. 2020, 167, 110611. [Google Scholar] [CrossRef]
  50. Liu, S.; Chen, C.; Lu, Y.; Ouyang, F.; Wang, B. An Interactive Method to Improve Crowdsourced Annotations. IEEE Trans. Vis. Comput. Graph. 2019, 25, 235–245. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Conceptual model of knowledge-assisted VA. The diagram is split into machine and human spaces, showing how knowledge is created, converted, and applied throughout the VA workflow [11]. Symbols for processes: A… analysis, V… visualization, X… externalization, P… perception/cognition, E… exploration; for containers: K ϵ … explicit knowledge D… data, S… specification, and K τ … tacit knowledge; and a non-persistent artifact: I… image.
Figure 1. Conceptual model of knowledge-assisted VA. The diagram is split into machine and human spaces, showing how knowledge is created, converted, and applied throughout the VA workflow [11]. Symbols for processes: A… analysis, V… visualization, X… externalization, P… perception/cognition, E… exploration; for containers: K ϵ … explicit knowledge D… data, S… specification, and K τ … tacit knowledge; and a non-persistent artifact: I… image.
Applsci 15 10172 g001
Figure 2. Functional building blocks that reflect the categorization schema of our survey. The category industrial domain influences all elements of the KAVA model [11]. The building block data covers [D] in the KAVA model, and our building block knowledge covers explicit [ K ϵ ] and tacit knowledge [ K τ ] and externalization [E]. The next building block for the survey is visualization, which covers [V] and the specification [S]. The building block machine learning is represented as [A] in the KAVA model. The building block user group covers exploration [E], tacit knowledge [ K τ ], perception [P], and externalization [X].
Figure 2. Functional building blocks that reflect the categorization schema of our survey. The category industrial domain influences all elements of the KAVA model [11]. The building block data covers [D] in the KAVA model, and our building block knowledge covers explicit [ K ϵ ] and tacit knowledge [ K τ ] and externalization [E]. The next building block for the survey is visualization, which covers [V] and the specification [S]. The building block machine learning is represented as [A] in the KAVA model. The building block user group covers exploration [E], tacit knowledge [ K τ ], perception [P], and externalization [X].
Applsci 15 10172 g002
Figure 3. PRISMA flow diagram illustrating the study selection and process.
Figure 3. PRISMA flow diagram illustrating the study selection and process.
Applsci 15 10172 g003
Figure 4. Data quality dashboard of Behringer et al. [42,46]: (a) overall quality score with color coding; (b) breakdown by dimension (accuracy, completeness, consistency, timeliness); (c) sample of raw records for context; (d) feature-level quality overview; (e) per-feature dimension scores; (f) value distribution histograms; (g) key statistical indicators (e.g., null counts, min/max, averages); (h) recommended remediation actions for identified issues.
Figure 4. Data quality dashboard of Behringer et al. [42,46]: (a) overall quality score with color coding; (b) breakdown by dimension (accuracy, completeness, consistency, timeliness); (c) sample of raw records for context; (d) feature-level quality overview; (e) per-feature dimension scores; (f) value distribution histograms; (g) key statistical indicators (e.g., null counts, min/max, averages); (h) recommended remediation actions for identified issues.
Applsci 15 10172 g004
Table 1. Screening stages with inclusion and exclusion criteria.
Table 1. Screening stages with inclusion and exclusion criteria.
Screening StageInclusion CriteriaExclusion Criteria
Title ScreeningStudies broadly addressing data visualizations, interaction methods, visual analytics, knowledge modeling approaches (e.g., ontologies), or data mining within the context of industrial manufacturing.Studies focusing on non-industrial domains (e.g., healthcare, education, cybersecurity, or energy).
Abstract ScreeningStudies focusing on industrial contexts. However, domain-agnostic studies were allowed if no specific application area was stated.Studies explicitly focusing on non-industrial applications; purely AR/VR/MR technologies; or unrelated fields not addressing knowledge modeling, visualizations, or visual analytics.
Full-Text ReviewPeer-reviewed studies with a clear focus on industrial applications, specifically addressing knowledge-assisted visual analytics (KAVA). Only studies explicitly mentioning an industrial setting and KAVA elements were included. We define KAVA elements as elements that include human explicit knowledge that is added before or during interaction with the visualization (e.g., (pre-) defined ontologies and knowledge graphs, texts and/or documents, thresholds set by humans, data labels for training an ML model, flow diagrams of processes defined by humans).Non-peer-reviewed studies (excluding book chapters), workshop papers, and studies not explicitly situated within an industrial context were excluded. Papers lacking explicit human knowledge were not considered, nor were outputs generated solely by ML models. Papers that may have included explicit human knowledge but failed to state it explicitly were also excluded. ML models that were modified by humans on a session basis (without saving the human input in a database or training a model for future use) were likewise excluded.
Table 2. Categorization overview: the six main entity groups data, knowledge, visualization, machine learning, user group, and industry domain and their subcategories with colors matching building blocks in Figure 2.
Table 2. Categorization overview: the six main entity groups data, knowledge, visualization, machine learning, user group, and industry domain and their subcategories with colors matching building blocks in Figure 2.
User GroupIndustrial DomainDataKnowledgeVisualizationMachine Learning
As ReportedAs ReportedTabularTime-DependentSpatialTextualGraphs/NetworksImage, Video, AudioProcessTypeOriginStorage MechanismGraphs/PlotsTablesDiagramsMaps/GeographicalOtherAs ReportedSupervisedUnsupervisedSemi-Supervised
Note: Cell background colors indicate category groupings and correspond to those used in this and all following tables throughout the manuscript.
Table 3. User groups addressed in the 13 reviewed KAVA publications. An “x” indicates that the group is mentioned in that paper; the final column shows the total count.
Table 3. User groups addressed in the 13 reviewed KAVA publications. An “x” indicates that the group is mentioned in that paper; the final column shows the total count.
[33][36][45][34][37][38][35][41][42][39][44][40][43]Total
User GroupOperatorxx x xx x 6
Managerxxxxx x 6
Process Engineer xx x 3
Machine Engineer x x x 3
Domain Expert xx x3
Analyst x xx 3
Table 4. Overview of the industrial domains addressed in the 13 reviewed publications. Each domain reflects the primary application context of the respective KAVA system. An “x” indicates that the paper explicitly targets or is situated within the corresponding domain. The final column shows the total count.
Table 4. Overview of the industrial domains addressed in the 13 reviewed publications. Each domain reflects the primary application context of the respective KAVA system. An “x” indicates that the paper explicitly targets or is situated within the corresponding domain. The final column shows the total count.
[33][36][45][34][37][38][35][41][42][39][44][40][43]Total
Industrial DomainManufacturing Industry xx xxx 5
Process Industryx x xx x5
Automotive Sector x x 2
Heavy Industry x 1
Table 5. Data types used in the 13 reviewed publications. An “x” marks the presence of a given data type in each paper. The final column shows the total count.
Table 5. Data types used in the 13 reviewed publications. An “x” marks the presence of a given data type in each paper. The final column shows the total count.
[33][36][45][34][37][38][35][41][42][39][44][40][43]Total
DataTabularxx xxxxxxxxxx12
Time-Dependentxx xx xxxxxx10
Spatial x x x x x5
Textual x x xxx5
Graphs/Networks x x x 3
Image, Audio, Video x x2
Table 6. Visualization techniques used across the 13 reviewed KAVA systems. The table lists each technique under the categories graphs/plots, tables, diagrams, maps, others, and as reported; marks its presence in each paper with an “x”; and shows the total count of papers employing that technique in the final column.
Table 6. Visualization techniques used across the 13 reviewed KAVA systems. The table lists each technique under the categories graphs/plots, tables, diagrams, maps, others, and as reported; marks its presence in each paper with an “x”; and shows the total count of papers employing that technique in the final column.
[33][36][45][34][37][38][35][41][42][39][44][40][43]Total
Visualization TypesGraphs/PlotsLine Graphx xx x xx x7
Bar Chart xxx xx x6
Scatterplotx x xxx5
Histogram x x 2
Radial Bar Chart x x 2
Box-and-Whisker Plot x x 2
Radar Chart x x 2
Stacked Bar Chartx 1
Stream Graphx 1
Area Graph x 1
Stacked Area Graph x 1
Parallel Coordinate Plot x 1
TablesCalendarxx 2
Heat Mapx 1
DiagramsNetwork Diagramsx x xx xxx 7
MapsConnection Map x 1
OtherDonut Chartx x x 3
Pie Chartx x2
Circle Packingx 1
Nightingale Rose Chart x 1
As Reported3D Model x x x 3
2D Model of Product x x2
Marey’s Graph x 1
Radial Heat Map x 1
Icicle Plot x 1
Table 7. Overview of knowledge modeling across the 13 reviewed KAVA systems, indicating for each publication which transformation pathways (e.g., direct externalization, interaction mining, intelligent data analysis) are supported, which knowledge types (operational, declarative, procedural) are addressed, at what stage knowledge enters the system (pre-design, design, post-design by data or users), and which storage mechanisms (graph databases, relational tables, flat files) are employed. An “x” denotes the presence of the feature in that paper. The final column shows the total count.
Table 7. Overview of knowledge modeling across the 13 reviewed KAVA systems, indicating for each publication which transformation pathways (e.g., direct externalization, interaction mining, intelligent data analysis) are supported, which knowledge types (operational, declarative, procedural) are addressed, at what stage knowledge enters the system (pre-design, design, post-design by data or users), and which storage mechanisms (graph databases, relational tables, flat files) are employed. An “x” denotes the presence of the feature in that paper. The final column shows the total count.
[33][36][45][34][37][38][35][41][42][39][44][40][43]Total
KnowledgePathwayKnowledge visualizationxxx xxxxxxxxx12
Direct externalizationxx xxxxxxxxxx10
Intelligent data analysisxxxx xx xxxx10
Data analysisxx xxxx 6
Interaction miningxx xx 4
Guidance xx x 3
Simulation 0
TypeDomain, declarativexxxxxxx xxxxx12
Operationalxx xx xxxxxxx11
Domain, proceduralxx x xxxxxx 9
OriginDesignxx xxxxxxxxxx12
Pre-designxxxxxxx xxx 10
Post-design, single userxx x xxxxxxxx10
Post-design, dataxx xxxxxxx 9
Post-design, multiple user xxx x 4
Storage Mechanism x xx xxxxxx x10
Table 8. Machine learning types used in the 13 reviewed publications. An “x” marks the presence of a given data type in each paper. The final column shows the total count.
Table 8. Machine learning types used in the 13 reviewed publications. An “x” marks the presence of a given data type in each paper. The final column shows the total count.
[33][36][45][34][37][38][35][41][42][39][44][40][43]Total
Machine LearningUnsupervisedxxx xxx 6
Semi-Supervisedxx 2
Supervised x 1
Hybrid Intelligent Optimization System x 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Böck, A.J.; Größbacher, S.; Vrablicz, J.; Stoiber, C.; Rind, A.; Suschnigg, J.; Schreck, T.; Aigner, W.; Wagner, M. KAVAI: A Systematic Review of the Building Blocks for Knowledge-Assisted Visual Analytics in Industrial Manufacturing. Appl. Sci. 2025, 15, 10172. https://doi.org/10.3390/app151810172

AMA Style

Böck AJ, Größbacher S, Vrablicz J, Stoiber C, Rind A, Suschnigg J, Schreck T, Aigner W, Wagner M. KAVAI: A Systematic Review of the Building Blocks for Knowledge-Assisted Visual Analytics in Industrial Manufacturing. Applied Sciences. 2025; 15(18):10172. https://doi.org/10.3390/app151810172

Chicago/Turabian Style

Böck, Adrian J., Stefanie Größbacher, Jan Vrablicz, Christina Stoiber, Alexander Rind, Josef Suschnigg, Tobias Schreck, Wolfgang Aigner, and Markus Wagner. 2025. "KAVAI: A Systematic Review of the Building Blocks for Knowledge-Assisted Visual Analytics in Industrial Manufacturing" Applied Sciences 15, no. 18: 10172. https://doi.org/10.3390/app151810172

APA Style

Böck, A. J., Größbacher, S., Vrablicz, J., Stoiber, C., Rind, A., Suschnigg, J., Schreck, T., Aigner, W., & Wagner, M. (2025). KAVAI: A Systematic Review of the Building Blocks for Knowledge-Assisted Visual Analytics in Industrial Manufacturing. Applied Sciences, 15(18), 10172. https://doi.org/10.3390/app151810172

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop